Senior Data Engineer

About us

ZAGENO was founded in 2015, ZAGENO is eliminating bottlenecks to breakthrough science and research through curated connections linking scientists and their institutions’ procurement offices with leading life science vendors. World-renowned academic and government research facilities, global pharmaceutical companies, and biotechs large and small are already users of our unique online biotech marketplace, currently comprising more than three million products.

What do we do?

  1. We help scientists choose the optimal laboratory kits and materials for each unique experiment setup.
  2. We make purchasing transactions more efficient for both buyers and sellers.
  3. For our vendor partners, we provide a valuable, expanded sales channel.
Our growing ZAGENO team of 60 in Cambridge, USA, and Berlin, Germany, includes experts in science, e-commerce, systems integration, and customer support to enable smarter, faster processes — allowing more time and resources for value-added science and better research results. Find out more on https://zageno.com

About the Role

ZAGENO is looking for a Senior Data Engineer. As a data engineer you will work with a small but growing data team to work across two main areas  (1) to develop pipelines to help clean, understand, and enrich events and content; (2) to develop the data platform that facilitates data analysis, from data warehousing to data science and analytics. Given the variety of data and applications at ZAGENO, you will be collaborating with people and teams across the company. The ideal candidate will be data and data-quality driven, collaborative, and curious.

In this role you will:
  • Architect and develop data pipelines to optimize for performance, quality, and scalability
  • Collaborate with data, engineering, and business teams to build tools and data marts that enable analytics
  • Develop testing and monitoring to improve data observability and data quality
  • Partner with data science to build and deploy predictive models
  • Support code versioning, and code deployments for Data Pipelines.

About You

  • 4+ years of experience as a data engineer
  • Expertise in an object-oriented or functional language is required (we use Python)
  • Strong experience with Data Modeling
  • Strong SQL skills are required (we currently use Postgres and BigQuery)
  • Strong experience with Big data tools / technologies (Hive, Impala, OOZIE, Airflow, NIFI)
  • Experience working with APIs is required
  • Experience orchestrating and scheduling data pipelines is required
  • Expertise with cloud environments is a plus (we are a GCP shop)
  • Experience with cloud data warehouses and distributed query engines is a plus
  • Keen attention to detail
  • Ability to work with ambiguity and drive analytical solutions pragmatically to completion
  • Proven record of designing and developing systems with desired SLAs and data quality metrics
  • Experience with Machine Learning & MLOPs is a big plus

Our Benefits

  • Working for a mission-driven business with a meaningful challenge making a positive impact on the scientific community
  •  A clear growth perspective
  • A professional and dynamic team with a global vision and mindset
  • An exciting, international working environment (we have 45 nationalities!)
  • Competitive compensation
  • Hybrid Working model - you can work from home and from the office
  • A learning and development budget to enable your ambitions to grow 
  • A budget to improve your home office environment