Full-Time

Principal Software Engineer

Data Pipelines

Updated on 5/17/2024

ZoomInfo

ZoomInfo

1,001-5,000 employees

B2B database and multiplatform revenue tools

Data & Analytics

Senior, Expert

Waltham, MA, USA

Required Skills
Kubernetes
Microsoft Azure
Airflow
Apache Beam
Git
BigQuery
Apache Spark
SQL
Apache Kafka
Java
Docker
AWS
Scala
Jenkins
Data Analysis
Google Cloud Platform
Requirements
  • Bachelor's or Master's degree in Computer Science, Software Engineering, or a related field.
  • Minimum of 10 years of experience in backend software development with a focus on data engineering and big data technologies.
  • Expertise in Apache Spark, Apache Beam, and Airflow.
  • Proficiency in Java, Scala, and SQL.
  • Experience building enterprise-grade software in a cloud-native environment (GCP or AWS).
  • Experience with cloud platforms (AWS, GCP, Azure) and containerization technologies (Docker, Kubernetes).
  • Experience in stream/data processing technologies like Kafka, Spark, Google BigQuery, Google Dataflow, HBase.
  • Familiarity with CI/CD pipelines (Jenkins, Github Actions).
  • Experience with SQL performance optimization, Graph and Vector database or processing frameworks.
  • Knowledge of data modeling, data warehousing, and data integration best practices.
  • Familiarity with streaming data processing, real-time analytics, and machine learning pipelines.
  • Strong problem-solving, analytical, and critical thinking skills.
  • Excellent communication and collaboration skills.
  • Experience in mentoring and leading technical teams.
Responsibilities
  • Architect and develop large-scale, distributed data processing pipelines using Apache Spark, Apache Beam, and Apache Airflow.
  • Design and implement efficient data ingestion, transformation, and storage solutions for structured and unstructured data.
  • Partner with Engineering Leaders, Architects, and Product Managers to provide technical solutions within a larger roadmap.
  • Build and optimize real-time and batch data processing systems ensuring high availability, fault tolerance, and scalability.
  • Collaborate with data engineers, analysts, and scientists to translate business requirements into technical solutions.
  • Implement best practices for data governance, data quality, and data security across the data lifecycle.
  • Mentor and guide junior engineers, fostering a culture of continuous learning and knowledge sharing.
  • Stay up-to-date with trends, technologies, and industry best practices in big data and data engineering domains.
  • Participate in code reviews, design discussions, and technical decision-making processes.
  • Contribute to the development and maintenance of CI/CD pipelines for efficient and reliable deployments.
  • Collaborate with cross-functional teams to ensure successful project delivery.

ZoomInfo is a leading provider of B2B database services and multiplatform operating systems tailored for revenue teams. This company sets industry standards with a comprehensive GTM platform, combining sales, marketing, and conversation intelligence. Its focus on equipping revenue teams with robust tools for prospecting, ABM, conversion, and data hygiene makes it an attractive workplace for professionals looking to impact the business growth landscape.

Company Stage

N/A

Total Funding

$7M

Headquarters

Vancouver, Washington

Founded

2007

Growth & Insights
Headcount

6 month growth

3%

1 year growth

5%

2 year growth

31%