Full-Time

Manager – Technical Delivery

Snorkel AI

Snorkel AI

1,001-5,000 employees

Programmatic AI development platform for enterprises

Compensation Overview

$160k - $240k/yr

+ Equity

Remote in USA + 3 more

More locations: San Francisco, CA, USA | San Carlos, CA, USA | New York, NY, USA

Hybrid

Category
Sales & Solution Engineering (1)
Required Skills
Python
SQL
Machine Learning
Risk Management
Data Analysis
Requirements
  • Experience: 5+ years in technical delivery, operations, or program management, with 3+ years managing people or teams.
  • Technical acumen: Strong grounding in data systems, analytics, or AI/ML concepts; comfort with SQL, Python, and data workflows is a plus.
  • Leadership: Proven track record of developing and scaling high-performing teams in a dynamic environment.
  • Project delivery expertise: Ability to oversee complex, multi-stakeholder technical projects from start to finish.
  • Communication skills: Strong written, verbal, and visual communication skills to translate technical and operational concepts to diverse audiences.
  • Business acumen: Skilled at assessing trade-offs, managing budgets, and aligning resources to business priorities.
  • Mindset: Bias to action, resilience in fast-paced environments, and willingness to roll up your sleeves when needed.
Responsibilities
  • Enable delivery success: Provide oversight on complex customer projects, ensuring delivery is on-time, on-budget, and meets the highest standards of quality.
  • Lead and manage the TDM team: Recruit, mentor, and develop a group of technical delivery managers, ensuring they are set up for success in delivering high-quality projects.
  • Drive operational excellence: Establish best practices, playbooks, and scalable processes across project planning, risk management, and stakeholder communication.
  • Partner with leadership: Collaborate closely with the Head of Delivery to translate strategic goals into operational execution and team priorities.
  • Cross-functional collaboration: Work with research, engineering, product, and client-facing teams to align delivery activities with broader business goals.
  • Champion continuous improvement: Identify opportunities to innovate and streamline delivery methodologies, tooling, and reporting.
  • Customer alignment: Support TDMs in interfacing with clients, helping refine scope, manage expectations, and ensure customer success.
  • Data-driven management: Use metrics and insights to monitor delivery health, team performance, and customer impact.
Desired Qualifications
  • Experience working in data, AI/ML, or enterprise SaaS environments.
  • Exposure to customer-facing delivery or field operations.

Snorkel AI helps large organizations build custom AI faster by turning manual, labor‑intensive AI workflows into programmatic processes. Its platform emphasizes using proprietary data and knowledge to tailor AI to specific workloads, so enterprises can deploy models more quickly than with traditional approaches. The company works with banks, government agencies, and Fortune 500 firms, and has roots in Stanford AI Lab research, with customers including Google, Apple, IBM, and DARPA, as well as backing from major investors. Snorkel AI’s goal is to accelerate enterprise AI deployment by providing data, tooling, and expertise that speed up model development and deployment at scale.

Company Size

1,001-5,000

Company Stage

Series D

Total Funding

$235M

Headquarters

Redwood City, California

Founded

2019

Simplify Jobs

Simplify's Take

What believers are saying

  • Raised $100M Series D at $1.3B valuation in 2025, led by Addition.
  • Accenture Ventures invested to accelerate AI in financial services like NatWest.
  • Launched Snorkel Evaluate and Expert Data-as-a-Service for production AI.

What critics are saying

  • Open-source Snorkel library replicates core weak supervision, eroding moat in 6-12 months.
  • Labelbox captures market with superior UI, surpassing Snorkel in 12-18 months.
  • Scale AI undercuts with 10x cheaper hybrid labeling for Fortune 500 clients.

What makes Snorkel AI unique

  • Snorkel AI uses programmatic labeling from Stanford AI Lab for 10-100x faster ML models.
  • Snorkel Flow encodes domain knowledge via labeling functions for unstructured enterprise data.
  • Trusted by Google, Apple, Intel, and five top US banks for proprietary AI datasets.

Help us improve and share your feedback! Did you find this helpful?

Benefits

Health - Snorkelers and their dependents are covered by comprehensive medical, dental, and vision plans.

Environment - We provide an allowance for Snorkelers to set up workstations however they want.

Wellness - Snorkelers are given a yearly wellness stipend to be used on anything relating to health and well-being.

Growth & Insights and Company News

Headcount

6 month growth

-9%

1 year growth

-7%

2 year growth

-5%
Harvey
Mar 11th, 2026
Introducing BigLaw Bench: research.

Introducing BigLaw Bench: research. BLB: Research helps Harvey AI identify how foundation models are improving on research tasks, enabling investigation of ways to improve AI-based legal research. The second of its major BigLaw Bench (BLB) expansions this quarter is BLB: Research. This dataset focused on hard agentic legal research problems. Working with its data partner Snorkel AI, a leader in creating complex expert data for frontier AI, Harvey AI identified a series of US Case Law research problems that leading models are currently unable to solve - even when provided with search tools like web search. The purpose of BLB: Research is twofold. First, to identify how foundation models are improving on research tasks and the failure modes that continue to cause them to return unsatisfactory answers. Second, to investigate ways to improve AI-based legal research through agentic systems and access to non-public data in addition to foundation model capabilities. By enabling Harvey AI to both identify the best model and to build better infrastructure for those models, BLB: Research helps Harvey AI build deeper and more accurate research capabilities for its customers. Benchmarking beyond models. BLB: Research is its first benchmark that requires models to operate end-to-end over more than documents. Models must use tools to perform searches in order to identify relevant research context and provide a grounded, cited response. While historically Harvey AI had measured search capabilities independently, evolutions in agent building and model deployment led Harvey AI to combine search and response into a unified benchmark. The main driver of this change is the growing consensus on search as the primary means for providing grounded model responses across the industry. On the research side, as models are increasingly trained to use search rather than training data for up-to-date knowledge, benchmarking their capabilities without search unrealistically undersells what they can do. "Building a holistic search benchmark allows us to measure performance in the way that tracks how research capabilities are being developed and how our customers expect them to be delivered." Search is also becoming table stakes in practice, with customers relying on knowledge sources and expecting pin-cited sources as the default for research-driven use cases. Building a holistic search benchmark allows Harvey AI to measure performance in the way that tracks how research capabilities are being developed and how its customers expect them to be delivered. Building realistic complexity. Given its ubiquity as a paradigm, models are, perhaps unsurprisingly, very good at search. They can write booleans and web searches and have learned to use search to explore, understand, and refine their research. A primary objective for BLB: Research was to validate these baseline capabilities, as well as identify areas in which the models were still not capable of turning their search capabilities into meaningful research outcomes for lawyers. In short, Harvey AI wanted to create a hard benchmark. A good way to make benchmarks hard is to make them esoteric. This is unfortunately also a great way to make a benchmark useless. While Harvey AI could show the models fail at a number of tasks, if succeeding at those tasks wouldn't actually help a legal professional, that failure is beside the point. Accordingly, Harvey AI needed to ensure that tasks were difficult and realistic. Realistic tasks are both intrinsically useful and represent a research pattern that is generally useful, such as identifying the best case to cite for a proposition, drafting a research memo, or planning a claim or defense. From these realistic task types Harvey AI then had to find those that were hard. Harvey AI did this not by predicting what should be hard but by actually finding the frontier of model capabilities. First, Harvey AI identified what failure looked like on a structured rubric: How poorly must a model perform before its answer is unhelpful to a legal professional? Harvey AI found that many partly correct answers are still helpful. For example, slightly-off analysis with good citations to principal cases still gives users a research head start. In general, Harvey AI found that model answers become unhelpful once the model completes less than 60% of the required task criteria. On these tasks, models are often missing critical reasoning junctures, making wrong turns in their research, or providing answers that are too surface level to be useful. Here are some research tasks that Harvey will use to benchmark models' research capabilities: Other practice areas included in the benchmark are: Intellectual Property, Commercial Litigation, Constitutional Law, Regulatory, Employment & Labor, Health Law & Life Sciences, Tax, Tort, Real Property, Media and Technology, Immigration, and Family Law. What's next for BLB: Research. Sifting through troves of documents to find answers, narratives, and trends is one of the most common tasks in legal practice. The same search capabilities that make models good at case law research also underpin the ability to search EDGAR, an investigation corpus, a deal room, or even a law firm's own internal knowledge. BLB: Research will build towards measuring all these capabilities and finding the right tools for enabling models to move lawyers from sifting to solving. The next step-change in search capabilities can come from various sources: model capabilities, unique data sources, or the tools to enable the former to interact with the latter. BLB: Research will allow Harvey AI to track and design innovations in all of these spaces, and convert them into the best possible research results for its customers.

TechMarketView
Aug 11th, 2025
Accenture Invests in Snorkel AI, $100M

Accenture has invested in Snorkel AI through Accenture Ventures to enhance AI deployments in the financial services sector. The collaboration aims to create industry-specific solutions for training AI models at scale. Snorkel, originating from Stanford AI Lab, specializes in programmatic data development. Recently, Snorkel raised $100 million in a Series D round, valuing it at $1.3 billion. This investment aligns with Accenture's contract with NatWest Group for AI-driven banking transformation.

AI Demand
Aug 7th, 2025
Accenture Bets Big on Snorkel AI to Supercharge Finance with Smarter Data!

Accenture, through Accenture Ventures, has made a strategic investment in Snorkel AI to accelerate enterprise AI development by enabling faster curation of high-quality datasets for training and evaluating models.

Surperformance
Aug 6th, 2025
Accenture Invests in Snorkel AI Partnership

Accenture has made a strategic investment in Snorkel AI through Accenture Ventures to speed up AI adoption, especially in regulated sectors like financial services. The collaboration aims to develop industry-specific AI solutions using high-quality training data. Snorkel AI will also join Accenture's Project Spotlight, an accelerator for data and AI companies. Financial details of the investment were not disclosed.

Tech in Asia
Jun 4th, 2025
Deepseek’S New Ai Model May Be Trained On Google’S Gemini

👩‍🍳 How we use AI at Tech in Asia, thoughtfully and responsibly.🧔‍♂️ A friendly human may check it before it goes live. More news hereChinese AI lab DeepSeek has released an updated reasoning model, R1-0528, which is reported to perform well in math and coding benchmarks.However, concerns have been raised regarding the potential use of data from Google’s Gemini AI family in training this model.Developer Sam Paech, based in Melbourne, shared evidence on social media indicating that R1-0528 shows similarities to Google’s Gemini 2.5 Pro.Another developer, known for creating SpeechMap, also noted that the reasoning patterns of R1-0528 resemble those of Gemini AI.DeepSeek has not disclosed the sources of data used for training the model.🔗 Source: TechCrunch🧠 Food for thought1️⃣ Model distillation creates an ethical gray area amid fierce AI competitionDistillation, the process of training smaller models using outputs from larger ones, has become a contentious but widespread practice in AI development, especially for companies with limited computing resources.While distillation itself is a legitimate technique, DeepSeek’s alleged use of competitors’ models highlights the intellectual property challenges in AI development, with previous accusations suggesting they used OpenAI’s outputs without authorization1.This case illustrates a technical reality: companies like DeepSeek, which are “short on GPUs and flush with cash,” may find it economically rational to create synthetic data from competitors’ models rather than building everything from scratch2.The increasing adoption of protective measures by major AI labs, such as OpenAI requiring ID verification from countries that exclude China or Google summarizing model traces, demonstrates how seriously these companies view the threat of unauthorized knowledge transfer3.These protective measures reflect a broader industry recognition that model weights represent the culmination of substantial investments, making them valuable intellectual property worth safeguarding4.2️⃣ AI contamination creates attribution challenges for researchers and companiesThe difficulty in definitively proving model copying stems partly from the growing “contamination” of the open web with AI-generated content, making it increasingly challenging to determine a model’s true training sources.As content farms flood the internet with AI-generated text and bots populate platforms like Reddit and X, the lines between human-created content and AI outputs are blurring, complicating efforts to create “clean” training datasets5.This contamination means that similar word choices and expression patterns across different models might simply reflect training on the same AI-generated web content rather than direct copying6.The challenges of attribution are further complicated by the fact that many models naturally converge on similar linguistic patterns due to shared training methodologies and objectives, making it difficult to establish definitive evidence of unauthorized distillation7.These attribution difficulties create significant implications for intellectual property protection in AI, as companies struggle to determine whether similarities between models indicate legitimate convergence or improper copying1.3️⃣ AI security measures signal a shift from open collaboration to competitive protectionThe increasing implementation of security measures by AI labs reflects a significant shift in the industry from open collaboration toward protecting competitive advantages in a high-stakes technological race.Major AI companies are implementing increasingly sophisticated protections, such as OpenAI requiring ID verification, Google “summarizing” model traces, and Anthropic explicitly protecting “competitive advantages,” signaling a new phase of AI development where knowledge protection trumps open sharing8.This defensive posture is emerging in a context where the stakes are enormous. Training a single large AI model can cost millions in computing resources and produce emissions equivalent to five cars’ lifetimes, making the intellectual property extremely valuable9.These protective measures are particularly notable in the context of international AI competition, with some U.S. legislators even proposing criminal penalties for downloading certain Chinese AI models like DeepSeek, highlighting the geopolitical dimensions of AI development10.The tension between collaboration and protection reflects a maturing AI industry where companies increasingly view their training methodologies and model capabilities as critical competitive assets rather than academic research to be openly shared3.Recent DeepSeek developments