Few words about project
We’re looking for a Senior Data Scientist to join our team in Warsaw or remotely.
As a Senior Data Scientist, you’ll support data analysis workflows and data engineering efforts at the US-based private investment firm. You’ll work directly with other data scientists and data analysts to enable data-driven decision business processes. You’ll also cooperate with the data engineering team to design and implement common code for the data science team.
The recruitment process is well organized - it’s one video meeting, that lasts about 1,5-2hours - 30 minutes of a talk with a Manager (algorithmic and data-science related technical questions), 30 minutes of talk with the Data Analytic (cultural fit interview), 30 minutes of talk with the Senior Software Developer (programming questions).
There is a plan to work at least 4 hours a day overlap with the USA East Coast timezone, so your work should be organized in hours like 10:00-18:00 or 11:00-19:00 PL timezone.
You will be responsible for...
Extracting actionable insights from broad, open-ended questions to influence investment decisions
Developing common data analysis workflows using NLP (sentiment analysis), Geospatial analysis and customer segmentation methods
Designing a common library for other data scientists that includes storage connectors, proxy authentication, secrets management, standard data visualization methods, and more
Designing and implementing ML/DL models for text and customer segmentation data
Designing and building AWS infrastructure together with the Data Engineering team
Developing data technology stack including API services and ETL pipelines,
Designing datasets and schemes for consistency and easy access
What's important for us?
At least 5 years of professional experience in data-related roles or software engineering
Undergraduate or graduate degree in Computer Science, Engineering, Mathematics, or similar
Excellent command of spoken and written English, at least C1
Expertise in Python and SQL
Proficiency in statistics and machine learning, as well as Python libraries like Pandas, NumPy, matplotlib, seaborn, scikit-learn, Keras, etc
Excellent understanding of natural language processing (NLP)
Good understanding of geospatial analysis and data formats
Experience with AWS EC2, S3 and Athena
Ability to use Docker and create Dockerfiles
Ability to take an ambiguous analysis question and run with it independently
Creative problem-solving skills
Great customer service and troubleshooting skills
Nice to have:
Experience in designing and implementing REST API services
Experience in building ETL processes and data pipelines with platforms like Airflow or AWS services
Familiarity with Azure DevOps stack
Knowledge of Google Maps, Open Street Map or TomTom APIs
Experience with ArcGis ecosystem
Experience in operating within a secure networking environment, like a corporate proxy
Perfectly, if you have one month of notice period, or you are available ASAP, but we can also wait for you longer.