
Work Here?
Work Here?
Work Here?
Industries
Data & Analytics
Enterprise Software
Company Size
201-500
Company Stage
Series E
Total Funding
$410M
Headquarters
Santa Clara, California
Founded
2015
Dremio provides a data analytics platform that allows businesses to access and analyze information directly from data lakes and databases without moving or copying it. The platform works by connecting to existing cloud storage on AWS or Azure and using data virtualization to let users run queries through their preferred business intelligence and data science tools. Unlike competitors that require data to be moved into a separate warehouse, Dremio enables "no-copy" analytics to save time and reduce storage costs. The company's goal is to help large enterprises simplify their data architecture and make strategic decisions faster by providing a scalable way to manage massive amounts of information.
Help us improve and share your feedback! Did you find this helpful?
Total Funding
$410M
Above
Industry Average
Funded Over
6 Rounds
Industry standards
Health, Dental, and Vision Insurance
401(k)
Stock Options
Work From Home
Office Events
Parental Leave Benefits
Paid Time Off
Dremio wants to turn Iceberg's open-format victory into a simpler lakehouse pitch. 8 April, 2026 As Apache Iceberg becomes the default table format for more AI and analytics workloads, Dremio is arguing that the real challenge is no longer adoption, but the operational burden that comes after it. Apache Iceberg has effectively won the table-format wars, and Dremio is using that moment to make a sharper case for its own platform: the hard part now is not choosing an open format, but managing it without adding new layers of cost and complexity. Dremio argues that enterprises embraced Iceberg because they wanted interoperability and less lock-in, and that the format has also become increasingly important for AI-era data architectures that need access to structured, semi-structured and unstructured data in one lakehouse. Why this matters for users. For users, the promise of Iceberg is flexibility. Teams can keep data in object storage, use multiple engines, and avoid getting trapped inside a single vendor's proprietary format. But Dremio's post makes the point that openness brings its own operational tax: Iceberg tables fragment over time, metadata grows, snapshots pile up, and performance can degrade unless engineers actively compact files, tune layouts and schedule maintenance jobs. For many data teams, that means time that should go toward new data products, models or business analysis instead gets spent babysitting tables. Dremio's competitive angle is automation. That is where Dremio tries to distinguish itself from competitors like Snowflake and Databricks. The company says it was built around Iceberg from the ground up, rather than adding support later, and is pitching itself as the platform that automates the parts of Iceberg management that users least want to do manually. According to Dremio, its platform continuously optimizes physical data layout with Iceberg Clustering, automatically adapts query acceleration through Autonomous Reflections, and handles file compaction, snapshot expiration, manifest rewriting and orphan file cleanup without manual scheduling. Dremio explicitly contrasts that with Databricks, where it says customers still manage optimization jobs themselves, and with Snowflake, where it says automation is more limited for Snowflake-managed Iceberg tables. The value proposition for customers is straightforward: lower operational overhead and better performance without dedicated maintenance work. Dremio says its autonomous optimization reduces the need for full table rewrites by targeting only degraded regions of data layout, while its reflections system materializes only what is needed based on observed query behavior. The company says this can replace more complex silver-and-gold ETL layering with a more virtualized approach and claims query speeds up to 20 times faster than competing lakehouses on TPC-DS benchmarks. That kind of message is aimed directly at teams that like Iceberg's openness but miss the more hands-off performance tuning of classic cloud warehouses. Interoperability is still the main strategic message. Dremio is also leaning hard on openness as a competitive weapon. The company says it co-founded Apache Polaris, an open catalog standard, and argues that this helps customers avoid a new kind of lock-in at the catalog layer. In the post, Dremio says every table it manages is accessible through compatible engines such as Spark, Trino, Flink, DuckDB and Dremio itself. It contrasts that with Databricks' Unity Catalog-centric approach and Snowflake's managed-table model. For customers building AI and analytics systems across multiple engines and frameworks, Dremio argues that open access to data and metadata is no longer optional. Why Iceberg V3 could matter more than it sounds. The company also uses the post to highlight Apache Iceberg V3, which it describes as the biggest upgrade since row-level deletes in V2. Dremio says it has already shipped V3 table read and write support, including binary deletion vectors that can make updates and deletes faster and less compute-intensive than older position-delete approaches. It also points to new row-level lineage fields, the VARIANT type for semi-structured data, and nanosecond-precision timestamps as features that make Iceberg more suitable for real-time analytics, CDC pipelines, financial services and IoT workloads. Dremio's argument is that these are not incremental additions but features that make Iceberg more practical for the next generation of AI-heavy data systems. What Dremio is really selling. Underneath the format-war framing, Dremio is really making a broader pitch about the future of the lakehouse. It is saying that openness alone is not enough; the winning platform will be the one that keeps Iceberg interoperable while removing the management burden that often comes with it. That gives Dremio a different position from vendors that support Iceberg but still steer customers toward proprietary catalogs, managed layers or heavier operational involvement. Image: Dremio
Dremio, the Agentic Lakehouse company, has announced several developments strengthening its position in the Apache Iceberg ecosystem. The company now offers Apache Iceberg V3 support in Dremio Cloud, featuring deletion vectors for faster change data capture, the VARIANT data type for JSON, and enhanced schema evolution capabilities. Dremio engineer JB Onofre has been elected to the Apache Software Foundation board, following his role shepherding Apache Polaris through incubation. Dremio co-created Polaris, which has now graduated to a top-level Apache project and powers the company's Open Catalog feature. The company's platform includes autonomous reflections for query optimisation, Iceberg clustering using Z-order, and automatic table maintenance. Dremio supports full read and write operations across REST-compatible engines including Spark, Flink, Trino and DuckDB.
Dremio deepens Apache Iceberg leadership with V3 Support, new community appointments, and Polaris momentum. GlobeNewswire | Dremio Corporation Today at 9:05am PDT SAN FRANCISCO, Calif., April 06, 2026 (GLOBE NEWSWIRE) - Dremio, the Agentic Lakehouse company, today highlighted its leadership across the Apache Iceberg ecosystem, including V3 support now available in Dremio Cloud, the election of Dremio engineer JB Onofre to the Apache Software Foundation board, and continued momentum behind Apache Polaris. A longstanding advocate for open-source collaboration and the elimination of vendor lock-in, Dremio has made foundational contributions to projects including Apache Arrow (co-creator and core contributor), Apache Iceberg (contributor and leading educator), and Apache Polaris (co-creator). Reinforcing this commitment, JB Onofre, who shepherded Polaris through incubation, has been elected to the Apache Software Foundation board. Iceberg V3 is designed to support more diverse and complex data types, offer greater control over schema evolution, and deliver performance enhancements for large-scale, high-concurrency environments. Dremio's V3 integration advances handling of semi-structured data, row-level changes, and schema evolution, with full support in Dremio Cloud, including the VARIANT data type for JSON, deletion vectors for faster CDC (change data capture), and improved schema evolution. "The Iceberg lakehouse has become the default architecture for AI and analytics," said Rahim Bhojani, CTO of Dremio. "Most platforms added Iceberg as a feature, but Dremio was built on it from the ground up. Capabilities like Autonomous Reflections, Iceberg Clustering, and now V3 compound on each other, delivering an Iceberg platform that's both the fastest and the easiest to manage." Dremio Continues to Set the Standard for Apache Iceberg Apache Iceberg V3 Support: Dremio delivers full read and write support for the latest Iceberg specification. Deletion vectors accelerate row-level operations for CDC and streaming workloads. The VARIANT type eliminates the schema-on-write bottleneck for semi-structured data. Row-level lineage provides built-in creation and update tracking for regulated industries with no additional tooling required. Arrow-Based SQL Engine for Iceberg: Dremio's query engine was built natively on Apache Arrow, the open columnar standard Dremio co-created, making it uniquely suited for Iceberg workloads. It processes Iceberg and Parquet data in vectorized batches without conversion to a proprietary format, delivering fast, scalable analytics with no lock-in. Autonomous Reflections: Dremio eliminates the management overhead of running an Iceberg lakehouse. Autonomous Reflections observe query patterns and automatically creates, refreshes, and retires materializations, accelerating queries from seconds to sub-second with no code changes or manual tuning. Reflections' incremental refresh keeps data fresh at low resource cost. Iceberg Clustering: uses Z-order to co-locate data across multiple columns simultaneously. ith two-level pruning that skips data at both the manifest and row-group level, it minimizes I/O by running continuously on petabyte-scale tables without full-table rewrites. Automatic table maintenance: compaction, snapshot expiration, and orphan file cleanup run on policy-based schedules with no manual intervention, keeping tables performant and storage costs in check. Enables engineers to focus on building data products, not maintaining tables. Open Catalog (Powered by Apache Polaris): Dremio co-founded Apache Polaris, the open Iceberg catalog standard now graduated to a top-level Apache project. Built on Polaris, Dremio's Open Catalog provides an Iceberg catalog that supports full read and write from any REST-compatible engine, including Spark, Flink, Trino, and DuckDB, all sharing the same Iceberg tables. Governance, including RBAC, row-level filters, column masking, and just-in-time credential vending, is enforced consistently at the catalog layer regardless of which engine is querying. Every Dremio-managed table is accessible to any Iceberg-compatible engine. Ingestion and Transformation: Dremio supports the full range of DML operations on Iceberg tables using standard SQL. Continuous ingestion via CREATE PIPE, batch loads via COPY INTO, and dbt Core integration make Dremio a complete platform for building and maintaining Iceberg-native data pipelines. About Dremio Dremio is the Agentic Lakehouse: the only Iceberg-native data platform built for agents and managed by agents. Every knowledge worker and AI agent gets instant, governed access to enterprise data through any LLM or tool of their choice. Federated queries reach any source without ETL pipelines. An AI Semantic layer adds business context so every agent draws from the same source of truth. The lakehouse manages itself, running clustering, optimization, and compaction autonomously. The result: trusted insights that drive better business outcomes, without the infrastructure complexity or overhead. A lead contributor to Apache Iceberg and co-creator of Apache Arrow and Apache Polaris. Trusted by Shell, TD Bank, Michelin, and thousands of organizations worldwide. Media Contact Chris McCoin or Richard Smith McCoin & Smith Communications Inc. [email protected] | [email protected] Elise Woodard Dremio Corporation 9494632203 [email protected] This is a paid placement. For further inquiries, please contact GlobeNewswire directly.
What to expect at Dremio's Subsurface World Tour 2025. Dremio's Subsurface World Tour 2025 spotlights the latest innovations in the open data lakehouse ecosystem with real-world use cases that bring technical depth and practical insight. What is Dremio's Subsurface World Tour? This year's event shines a spotlight on the latest innovations in the open data lakehouse ecosystem, with real-world use cases that bring technical depth and practical insights together. Subsurface is the global stage for data pioneers, blending cutting-edge research with hybrid lakehouse strategies that are shaping the future of data. This year, the vendor is bringing even more Dremio-focused sessions - offering a closer look at its architecture, performance gains, and role in the modern data stack. Why you should attend. Over 18,000 data engineers, architects, and scientists from around the world have joined us at past events. Their speaker lineup has included experts from industry leaders like Apple, Netflix, Lyft, LinkedIn, TransUnion, Uber, Marsh McLennan, Adobe, AWS, Microsoft, Shell, and Wayfair - as well as emerging players like SpiceAI and Forcemetrics. They'll also feature the original creators behind Apache Arrow, Apache Iceberg, Apache Parquet, Project Nessie, Pandas, and more. Best of all, it's free - and virtual, or coming to a city near you. Don't miss your chance to connect with peers, learn from industry leaders, and be part of the future of data. Featured speakers. * Dremio CEO Sendur Sellakumar * Dremio Founder Tomer Shiran About Dremio Subsurface World Tour. You'll learn from peers who are implementing or managing lakehouse architectures to meet their organizations' data analytics needs. You'll learn from practitioners who are solving common data management and analytics challenges, including data optimization, self-service analytics, and data governance. You'll also gain firsthand insights on important analytics topics including Apache Iceberg and other Open Table Formats, Data Ops, Analytics, Cost Optimization, Data Mesh and Data Fabric architectures, the role of Gen AI in analytics and much more! * What: Dremio Subsurface World Tour 2025 * When: December 10 (Virtual), in-person sessions globally * Where: See registration page to attend free Share this. Tim king. Tim is solutions review's executive editor and leads coverage on data management and analytics. A 2017 and 2018 most influential business journalist and 2021 "who's who" in data management, tim is a recognized industry thought leader and changemaker. Story? Reach him via email at tking@solutionsreview dot com.
Dremio unveils The Agentic Lakehouse: built for agents, managed by agents. Next-generation Dremio Cloud reimagines the Lakehouse for the Agentic Era Subsurface Conference, NEW YORK, NY, Nov. 13, 2025 (GLOBE NEWSWIRE) - Dremio, the agentic lakehouse company, today announced Dremio Cloud, the industry's first agentic lakehouse that ushers in a new generation of data lakehouse platforms to empower artificial intelligence. Unlike other offerings, Dremio Cloud transforms legacy-based data infrastructures into a self-managing platform that continuously learns, adapts, and optimizes without human intervention. This enables data engineers to be 10x more productive and focus on innovation by eliminating mundane and time-consuming work. As enterprises race to deploy AI agents, they face a dual challenge: their data infrastructure processes and methodologies cannot support AI agents, and managing that data platform requires constant manual intervention, training, and education. Dremio Cloud solves both problems by positioning agents as a first-class operator-not a copilot or sidecar-creating an autonomous platform that seamlessly serves AI agents, applications, and users. "At BARC, we see high demand for tool consolidation, automation, and agentic data management. Dremio's Agentic Lakehouse addresses this demand with an array of capabilities that minimize toil and boost productivity for data teams," said Kevin Petrie, vice president of research at BARC. Dremio's agentic platform delivers everything AI agents need to succeed, including: Unified and Governed Data Access: Dremio's Open Catalog, built on the open-source Apache Polaris project-which Dremio co-created-makes it simple to catalog and govern all data across lakehouses and databases. Data is accessible through Dremio's Intelligent Query Engine, which features Zero-ETL federation and both structured and unstructured data processing. AI Semantic Layer: Dremio's deeply integrated semantic layer acts as a living encyclopedia for the business. It provides the crucial context AI needs to avoid hallucinations and deliver accurate insights. AI Agent: With its own AI Agent, Dremio Cloud makes it simple for business users and data analysts to ask questions and get high-quality answers and visualizations in seconds. Agent Choice Through MCP: Dremio Cloud natively supports the open Model Context Protocol (MCP), allowing any MCP-enabled AI agent from providers like Anthropic, OpenAI, and Google to connect to Dremio. Dremio Cloud introduces autonomous capabilities that significantly enhance the productivity of data engineers with: Active Metadata: Dremio's active metadata system continuously analyzes query patterns, data relationships, and usage trends to drive autonomous decision-making. This living intelligence layer enables the platform to predict needs, prevent performance issues, and proactively optimize data layouts-capabilities that lay the groundwork for future agent-driven automation across all platform operations. Enhanced Semantics: Dremio automatically creates wikis and labels for tables, views, and data products, with a human-in-the-loop to govern quality. Materializations and Query Rewrite: The platform learns from every query, automatically creates performance materializations, and rewrites SQL in real time to deliver consistent sub-second performance. Clustering: Dremio intelligently reorganizes data layouts through automated clustering and continuously analyzes access patterns to physically restructure tables for optimal query performance. "Our new agentic lakehouse represents a fundamental shift in how data platforms operate. By embedding intelligent, autonomous capabilities into a managed platform, Dremio Cloud eliminates the manual overhead that plagues data teams while providing the performance and reliability that AI agents demand," said Sendur Sellakumar, CEO at Dremio. "As our agentic autonomous capabilities evolve, they will increasingly operate as coordinated agents themselves, which creates a truly self-managing data ecosystem." Availability: Dremio Cloud, The Agentic Lakehouse, is available immediately. Businesses can sign up for a $400 free trial at dremio.com/get-started and experience the future of autonomous data on a fully managed data platform. About Dremio: Dremio is the pioneer of The Agentic Lakehouse-the only data platform built for agents, managed by agents. Organizations need to transform ideas into actions at unprecedented speed-Dremio delivers this agility by equipping AI agents with federated data access, unstructured data processing, and rich business context through its AI Semantic Layer. In the agentic era, data engineering teams can't manually tune performance for thousands of users and agents asking unpredictable questions every second. Dremio's Agentic Lakehouse autonomously manages itself, removing undifferentiated management tasks, allowing engineers to focus on initiatives that drive business results. Dremio's agentic lakehouse automatically optimizes queries, reorganizes data, and maintains performance at any scale. Dremio is trusted by thousands of global enterprises including Shell, TD Bank, and Michelin, and built on open standards. Dremio co-created Apache Polaris and Apache Arrow, and it's the only lakehouse built natively on Apache Iceberg, Polaris, and Arrow. Learn more atwww.dremio.com. Dremio Corporation
Find jobs on Simplify and start your career today
Industries
Data & Analytics
Enterprise Software
Company Size
201-500
Company Stage
Series E
Total Funding
$410M
Headquarters
Santa Clara, California
Founded
2015
Find jobs on Simplify and start your career today