Staff Data Engineer
Posted on 9/27/2023
INACTIVE
Lyft

10,001+ employees

Ridesharing app
Company Overview
Lyft's mission is to improve people's lives with the world's best transportation. The company operates a mobile platform for the ridesharing of cars, bikes, and scooters and serves over a million rides per day.

Company Stage

Series I

Total Funding

$5B

Founded

2012

Headquarters

San Francisco, California

Growth & Insights
Headcount

6 month growth

1%

1 year growth

-1%

2 year growth

3%
Locations
Toronto, ON, Canada
Experience Level
Entry
Junior
Mid
Senior
Expert
Desired Skills
Apache Hive
Apache Spark
Bash
Data Analysis
Hadoop
Airflow
MySQL
Oracle
Postgres
Ruby
SQL
Python
CategoriesNew
Data & Analytics
Requirements
  • 10+ years of relevant professional experience
  • Experience with Hadoop (or similar) Ecosystem (MapReduce, Yarn, HDFS, Hive, Spark, Presto, Pig, HBase, Parquet)
  • Strong skills in a scripting language (Python, Ruby, Bash)
  • Good understanding of SQL Engine and able to conduct advanced performance tuning
  • Proficient in at least one of the SQL languages (MySQL, PostgreSQL, SqlServer, Oracle)
  • 3+ years of experience with workflow management tools (Airflow, Oozie, Azkaban, UC4)
  • Comfortable working directly with data analytics to bridge Lyft's business goals with data engineering
Responsibilities
  • Work across teams to gather requirements and drive alignment on project goals
  • Provide technical leadership to the team for executing projects within scope, time & quality constraints
  • Evangelize best practices within the team through tech talks, demos etc
  • Take initiative and craft projects to improve customer experience & satisfaction by resolving long standing problems or by building important user facing features
  • Owner of the core organization data pipeline, responsible for scaling up data processing flow to meet the rapid data growth at Lyft
  • Evolve data model and data schema based on business and engineering needs
  • Implement systems tracking data quality and consistency
  • Develop tools supporting self-service data pipeline management (ETL)
  • SQL and MapReduce job tuning to improve data processing performance
  • Unblock, support and communicate with internal & external partners to achieve results