Data Infrastructure Engineer
Updated on 5/24/2023
Locations
Sherman Oaks, Los Angeles, CA, USA
Experience Level
Entry
Junior
Mid
Senior
Expert
Desired Skills
Apache Hive
Apache Spark
AWS
Data Analysis
Data Science
Redshift
Scala
SQL
Terraform
Apache Flink
Python
Requirements
  • Bachelor's degree in Computer Science, a related technical field, or equivalent practical experience
  • 4+ years of experience in engineering data solutions using big data technologies (Hive, Presto, Spark, Flink) on large-scale data sets
  • Passionate about building data models, data processing pipeline, and data infrastructure
  • Expertise in Infrastructure as Code with Terraform and Amazon Web Services
  • Strong knowledge of AWS data and analytics solutions (e.g. Kinesis, Glue, Athena, Redshift, S3, Aurora, Lake Formation)
  • Experience with one or more programming languages, including but not limited to: Python, SQL, Scala
  • Good communication and collaboration skills. Able to take vague business and analytics requirements and convert them into scalable data solutions
Responsibilities
  • Design and implement our data warehouse solution on AWS using Terraform (required)
  • Develop data schemas, tables, and views to govern consumer and performance data
  • Build ETL pipelines for data profiling, data cleaning, and data aggregation
  • Ensure data integrity across all data sources
  • Manage data ingestion and processing infrastructure stacks with proper monitoring and alerting
  • Provide underlying data support for all analytics projects
  • Collaborate closely with data analytics engineering and data science teams
Genies
Avatar ecosystem tools