Responsibilities:
• Build highly scalable data pipelines to handle ingestion and processing of data.
• Builds data models to support Data Sciences. Integrates data from a variety of sources, assuring that they adhere to data quality and accessibility standards.
• Analyzes current information technology environments to identify and assess critical capabilities and recommend solutions.
• Develop and support process flow analysis and ETL process design.
• Improve machine learning algorithms performance.
• Lead the entire software lifecycle including development, code reviews, testing, deployment, for data, ETL’s and RESTful API’s.
• Team player.
Requirements:
• BSc.degree with strong analytic focus preferred, 5-7 years industry experience.
• Strong data modeling experience required in a mixed NoSQL and SQL environments.
• Strong understanding of SQL.
• Experience with NoSQL-based storage.
• Experience with cloud environment.
• Languages: Python, T-SQL, Java or Scala. R is a plus.
• Experience with Kafka – a plus.
• Passionate about leading edge systems, tools and processes; uncompromising on test driven development, continuous integration, agile methodologies and tight collaboration.