Big Data Engineer / ETL Engineer

Berlin

Germany

Vilnius

Lithuania

Prague 

Czech Republic

Cambridge, MA

United States

Newtown, PA

United States

    Odysseus Data Services (https://odysseusinc.com/) - is a fast-growing US-based company focused on the creation of a single platform that will combine medical history data from clinics all over the world.  Our goal is to provide a vast information base to be used by scientific researchers, medical centers, and institutes, as well as the pharmaceutical industry and insurance companies.

    Today Odysseus Data Services is an active member of the Observational Health Data Sciences and Informatics (or OHDSI, pronounced "Odyssey") consortium (
https://www.ohdsi.org/), whose main mission is to improve healthcare systems around the world by collecting and organizing data on the treatment of various diseases and then combining this data into a single, standardized, universally-accessible knowledge base to bring out the value of health data through large-scale analytics.

    Due to our rapid growth, we now have a vacancy for a Big Data Engineer in the city of Berlin, Prague, Vilnius, and Newtown, PA.

Responsibilities

  • Data Management and Big Data Engineering e.g. complex ETL, data analysis, data mappings, advanced transformations
  • Architect and utilize Big Data Platforms for Big Data projects
  • Develop solutions in Agile fashion following Scrum process and Test-Driven approach enabled with Atlassian tools (JIRA, Confluence)
  • Develop solutions by applying industry best practices and coding standards
  • Document architecture design by creating necessary architecture artifacts, including UML domain models, component and deployment diagrams. Document business requirements as user stories
  • Create prototypes and POCs, as needed
  • Design secure and compliant solutions by implementing necessary application security and following various regulatory laws (HIPAA, GDPR)
  • Be a part of the dynamic open-source observational research OHDSI community, participate in workshops, hackathons and collaborate to implement observational research solutions e.g. have a fun time developing cutting-edge solutions

Required Skills

  • Bachelor’s or Master’s degree in Information Technology or Computer Science — or equivalent experience in information technology and software development
  • Complex SQL, including Spark SQL, Hive QL
  • Practical experience with ETL and big data processing
  • Practical experience with RDBMS (any of the following - Oracle, PostgreSQL, SQL Server)
  • Practical experience with MPP databases (any of the following - AWS RedShift, BigQuery, Synapse)
  • Practical experience with Hadoop e.g. Spark, Hive, Impala
  • Practical experience with one of the following platforms: AWS, GCP, and Cloudera 
  • Experience designing and developing solutions using Agile software development approach

Required Experience

  • ETL and Big Data Engineering: 3 years (required)
  • SQL: 3 years (required)
  • Data Warehouse: 3 years (required)

Desired Skills

  • Experience in administering the Big Data platform infrastructure
  • Knowledge of Java and Python
  • Experience in solving DevOps-related tasks
  • Experience with OMOP CDM, OMOP Standardized Vocabularies, methods, and standards developed by the OHDSI community