Senior Devops Engineer
Big Data Platform
Posted on 2/13/2024
Roku

1,001-5,000 employees

Streaming platform for television
Company Overview
Roku's mission is to be the TV streaming platform that connects the entire TV ecosystem around the world. The company connects users to the streaming content they love, and we enable content publishers to build and monetize large audiences and provide advertisers with unique capabilities to engage consumers.
Consumer Software
Hardware

Company Stage

Series F

Total Funding

$304.3M

Founded

2002

Headquarters

San Jose, California

Growth & Insights
Headcount

6 month growth

-6%

1 year growth

-3%

2 year growth

13%
Locations
Cambridge, MA, USA
Experience Level
Entry
Junior
Mid
Senior
Expert
Desired Skills
Datadog
Chef
Kubernetes
Microsoft Azure
Python
Airflow
Puppet
React.js
Apache Spark
Docker
AWS
Terraform
Ansible
Hadoop
Development Operations (DevOps)
Linux/Unix
Looker
Google Cloud Platform
CategoriesNew
DevOps & Infrastructure
Requirements
  • Bachelor’s degree or equivalent work experience
  • 8+ years of experience in DevOps or Site Reliability Engineering
  • Experience with Kubernetes or Docker
  • Experience with Terraform, or Chef, or Puppet, or Ansible
  • Experience with Cloud infrastructure such as Amazon AWS, Google Cloud Platform (GCP), Microsoft Azure, or other Public Cloud platforms
  • Experience with at least 3 of the technologies/tools mentioned in the job description
  • Strong background in Linux/Unix shell scripting (or equivalent programming skills in Python)
  • Experience working with monitoring and alerting tools (such as Datadog and PagerDuty) and being part of call rotations
  • Understanding of automation tools like Ansible, Terraform, AWS Opswork
  • Experience with Apache Airflow
Responsibilities
  • Develop best practices around cloud infrastructure provisioning, disaster recovery, and guiding developers on the adoption
  • Collaborate on system architecture with developers for optimal scaling, resource utilization, fault tolerance, reliability, and availability
  • Conduct low-level systems debugging, performance measurement & optimization on large production clusters and low-latency services
  • Create scripts and automation that can react quickly to infrastructure issues and take corrective actions
  • Participate in architecture discussions, influence product roadmap, and take ownership and responsibility over new projects
  • Collaborate and communicate with a geographically distributed team
Desired Qualifications
  • GCP preferred
  • Experience with system engineering around edge cases, failure modes, and disaster recovery
  • Experience scaling production systems running Big Data tools like Spark, Hadoop, Apache Druid, Looker