Dataops Engineer
Posted on 4/3/2024
Veeva Systems

5,001-10,000 employees

Cloud computing services for pharmaceutical companies.
Company Overview
Veep's mission is to help R&D, quality, and regulatory teams eliminate inefficiencies and bring high-quality, safe, sustainable products to market without compromising quality. The company builds cloud-based tools for pharmaceutical research.

Company Stage

N/A

Total Funding

$224M

Founded

2007

Headquarters

Pleasanton, California

Growth & Insights
Headcount

6 month growth

4%

1 year growth

22%

2 year growth

39%
Locations
London, UK
Experience Level
Entry
Junior
Mid
Senior
Expert
Desired Skills
Python
Airflow
AWS
Data Analysis
CategoriesNew
Computational Biology
Genomics
Biology & Biotech
Requirements
  • Proficient in Python programming language and PySpark
  • 3+ years of experience working with Apache Spark
  • Previous experience building tools and libraries to automate and streamline data processing workflows
  • Experience running data workflows through DevOps pipelines
  • Experience orchestrating data workflows using state-of-the-art tools (e.g., Airflow, AWS Steps, or similar from other cloud vendors), spawning jobs in a Spark cloud-managed cluster (e.g., EMR, Databricks)
  • Experience with the Delta Lake Architecture and the delta format
Responsibilities
  • Build DataOps tools for data workflows automation and streamline data processing
  • Adopt solutions and tools that adhere to the DataOps best practices
  • Continually strive to reduce wasted effort, identify gaps and correct them, and improve data development and deployment processes
  • Develop the ingestion pipelines for raw data
  • Put in place the building blocks to deliver the data core for life sciences