Key Responsibilities:
You will design and implement pipelines for large scale data collection, analysis and persistence as well as training and serving Machine Learning models
You will design and develop production-grade components from data analysis prototypes
You will work with Data Scientists, Product Managers, and Business Partners to frame problems within the business and product context
You will provide engineering good-practices regarding design and coding in the applied research team
Who you are:
You are a team player, willing to collaborate and communicate with others to solve problems
You are an interdisciplinary programmer with a wide scope of technical understanding
You have a real passion for solving difficult engineering problems
You are tech-savvy and passionate about new technological challenges
You are self-motivated, and able to work both independently and as part of a collaborative team
Skills and Experience:
Bachelor’s degree in Computer Science or equivalent practical experience
5 years of professional experience as an engineer in the industry
Expertise in Python, Java or Scala
Experience with Spark
Experience with building data pipelines (Airflow, Luigi, etc.)
Familiarity with Unix/Linux operating systems and commands
Preferred qualifications:
Experience with PySpark
Experience with Athena/Big Query
Knowledge of ML methods and statistics
Experience with Kubernetes and Docker
Please note, this position is not eligible for visa sponsorship.