Search All Jobs

Showing 1 of 1 Jobs

Engineering Manager - Machine Learning

$185k - $220k/yr

Remote in USA

Full-Time

Engineering Manager

Machine Learning

Confirmed live in the last 24 hours

WellSaid Labs

WellSaid Labs

51-200 employees

AI platform for generating voiceovers

Compensation Overview

$185k - $220k/yr

+ Stock Options

Senior

No H1B Sponsorship

Remote in USA

Candidates must be U.S. Citizens or Permanent Residents.

US Citizenship Required

Category
Applied Machine Learning
Deep Learning
AI & Machine Learning
Engineering Management
Required Skills
Machine Learning
AWS
Google Cloud Platform
Requirements
  • Extensive experience in leading high-performing engineering teams
  • 4-5+ years of experience leading or managing a software engineering team and a proven track record
  • 2+ years of experience using Cloud platforms (GCP, AWS, etc)
  • Proficient in software development practices, including specing, documenting, testing, reviewing, deploying, and monitoring
  • Excellent communication and collaboration skills to work effectively with cross-functional teams
  • Flexible and capable of working with a diverse team with various communication styles and needs
  • Create equitable processes across the entire employee lifecycle
  • Be a U.S. Citizen or Permanent Resident
  • Pass a pre-employment background check
Responsibilities
  • Build and lead a high-performing and collaborative machine learning team
  • Collaborate with Product Management and other stakeholders to define, prioritize, and deliver the ML roadmap
  • Ensure that complex projects are effectively scoped, planned, executed, and delivered on time
  • Oversee multiple concurrent machine learning projects simultaneously
  • Own and run team standups, sprint planning, and other delivery ceremonies
  • Work one-on-one with team members to meet their needs, providing regular and constructive feedback
  • Conduct design, code and pull request reviews with the team
  • Champion best practices in machine learning and software engineering
  • Manage team resources effectively by prioritizing projects based on impact
  • Foster customer empathy within the team
Desired Qualifications
  • Knowledge in building deep learning models in PyTorch (not required)
  • Background in Machine learning frameworks (such as PyTorch), and data processing tools (Spark, Pandas, etc.) are a big bonus but not mandatory
  • Experience in text-to-speech, speech-to-text, audio libraries, or cloud storage is a bonus

WellSaid Labs provides a platform that transforms written text into high-quality voiceovers using AI voice generation technology. The service is designed for businesses seeking to create engaging content quickly and affordably, offering a variety of AI-generated voices to suit different needs. Unlike traditional voiceover production, WellSaid Labs leverages advanced speech synthesis to produce natural-sounding voiceovers efficiently. The company operates on a subscription model, aiming to enhance customer engagement and brand presence for its clients.

Company Size

51-200

Company Stage

Series A

Total Funding

$10M

Headquarters

Seattle, Washington

Founded

2018

Simplify Jobs

Simplify's Take

What believers are saying

  • WellSaid Labs won the 2024 Artificial Intelligence Breakthrough Award.
  • The demand for personalized AI voices is increasing, benefiting WellSaid Labs.
  • Their subscription model offers scalable revenue opportunities with various pricing tiers.

What critics are saying

  • ElevenLabs' AI sound effects could divert customers from WellSaid Labs.
  • Competitors' rapid advancements may saturate the AI voice market.
  • High UX expectations could lead to dissatisfaction if not met.

What makes WellSaid Labs unique

  • WellSaid Labs offers life-like synthetic voices with customizable emotional directing.
  • Their HINTS technology allows nuanced AI voice customization for creative professionals.
  • The platform provides user-friendly, high-quality voiceovers for diverse industries.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Health Insurance

Dental Insurance

Vision Insurance

401(k) Retirement Plan

401(k) Company Match

Unlimited Paid Time Off

Parental Leave

Professional Development Budget

Home Office Stipend

Stock Options

Company Equity

Growth & Insights and Company News

Headcount

6 month growth

2%

1 year growth

3%

2 year growth

8%
GeekWire
Jan 15th, 2025
Wellsaid’S New Ai Voice Model Will Introduce ’Emotional Directing’ For Inflection And Tone

GeekWire’s startup coverage documents the Pacific Northwest entrepreneurial scene. Sign up for our weekly startup newsletter , and check out the GeekWire funding tracker and venture capital directory .Brian Cook joined Kirkland-based WellSaid Labs a year ago as CEO. (GeekWire Photo / Todd Bishop)A new AI model from WellSaid Labs will let users guide the emotions, pitch, and pace of AI-generated voice clips in the same way as a human director would coach a voice actor to produce a desired result.The Kirkland, Wash.-based company announced the new model, dubbed Caruso, Wednesday morning in advance of its upcoming launch. Other features include faster audio rendering and improved pronunciation.The idea is to generate AI voices that “say it right the first time, drastically reducing the time and effort that goes into re-rendering audio clips,” said Brian Cook, the company’s CEO, in a post announcing the new model.Originally spun out of Seattle’s AI2 Incubator, WellSaid raised a $10 million Series A round in 2021 led by Fuse. The company focuses on the enterprise AI voice market, with practices and policies designed to encourage safe and ethical AI usage, seeking to distinguish itself in the AI voice market.Cook, the former CEO of Nintex and founder of Incredible Capital, joined WellSaid as CEO a year ago. Matt Hocking is the company’s co-founder and executive chairman

MarTech Cube
Jul 5th, 2024
WellSaid Won 2024 Artificial Intelligence Breakthrough Award

WellSaid won 2024 Artificial Intelligence Breakthrough award.

VentureBeat
May 31st, 2024
Elevenlabs Moves Beyond Speech With Ai-Generated Sound Effects

Time's almost up! There's only one week left to request an invite to The AI Impact Tour on June 5th. Don't miss out on this incredible opportunity to explore various methods for auditing AI models. Find out how you can attend here. After launching tools for text-to-speech and speech-to-speech synthesis, AI voice startup ElevenLabs is moving to the next target. The two-year-old startup founded by former Google and Palantir employees today announced the launch of a new text-to-sound AI offering called Sound Effects.Available starting today on the ElevenLabs website, Sound Effects uses the startup’s in-house foundation model and allows creators to generate different types of audio samples by simply typing a description of their imagined sound. The company first teased the tool in February with a post featuring Sora-generated clips, albeit enhanced with AI sound effects.ElevenLabs partnered with Shutterstock to bring this product to life and expects to see adoption from creators across domains who are looking to enhance their content with immersive soundscapes

GeekWire
Apr 1st, 2024
Vote For Ux Design Of The Year: Finalists Help Users Across Investing, Data Analysis And More

Ophir Ronen, founder and CEO of CalmWave, accepts the award for UX Design of the Year at the 2023 GeekWire Awards. (GeekWire File Photo / Kevin Lisota)The finalists for UX Design of the Year at the 2024 GeekWire Awards are helping end users across home building, investing, AI-generated speech, data analysis, and schedule and expense tracking for gig workers. The award is intended to recognize a product that has a user experience that is meaningful, usable and delightful for those end-users. The five finalists are Digs, Copper, Row Zero, WellSaid Labs, and Solo.Keep scrolling for descriptions of each finalist, provided through the nomination process. And submit your vote here or below for UX Design of the Year, presented by Blink UX.Digs End user: Residential home builders, vendors, and homeowners.Describe how the UX benefits an end user: The UX of Digs takes the complicated process of collaboration on a home build and makes it intuitive and simple to use with familiar UX similar to many favorite social media and collaboration platforms. Add in automatic organization and AI-powered search and the user has a powerful yet easy-to-use experience that enhances the build process for everyone involved.CopperEnd user: Kids, teens, and parents.Describe how the UX benefits an end user: Starting from obtaining SEC approval for teens to be able to build an investment portfolio that is focused on driving financial knowledge through investing real money in the stock market, the team has distilled a tremendous amount of complexity into a user experience that has driven over 2.6 million members to start their journey in wealth building

VentureBeat
Feb 19th, 2024
Hear Your Imagination: Elevenlabs To Launch Model For Ai Sound Effects

After mastering the art of machine learning (ML) based voice cloning and synthesis, ElevenLabs, the two-year-old AI startup founded by former Google and Palantir employees, is moving to expand its portfolio with a new text-to-sound model.Teased a few hours ago, the AI will allow creators to generate sound effects by simply describing their imagination in words. It is expected to enrich content in a new way in the age of AI-driven digital experiences. The model is not available publicly, but ElevenLabs has showcased its capabilities by releasing a minute-long teaser featuring videos produced by OpenAI’s new Sora and enhanced with its own AI sounds. The company has also set up a signup page and is calling potential users to join an early access waitlist for the model.Going beyond voice with AI sound effectsFounded in 2022, ElevenLabs has been researching AI to make audio and video content – from movies to podcasts – accessible across languages and geographies. The company has debuted a range of offerings to further this, including text-to-speech and speech-to-speech models that can produce AI speech from a given piece of content (text/audio/video) in 29 different languages whilst delivering natural voice and emotions (original speaker’s voice in speech-to-speech)