Jacar mascot — reading along A laptop whose eyes follow your cursor while you read.
Arquitectura

DuckDB in enterprise analytics: concrete cases

DuckDB in enterprise analytics: concrete cases

Actualizado: 2026-05-03

DuckDB is one of those projects that has been growing quietly for two or three years and suddenly appears everywhere. What started as the embedded analytical database for analysts with laptops has, as of October 2025, become a piece appearing in small and medium enterprise data architectures with surprising frequency. This article collects the concrete patterns where it is used, where it displaces more expensive pieces, and where it doesn’t reach yet.

Key takeaways

  • DuckDB is an embedded columnar OLAP engine that reads Parquet from local disk or S3 without importing it, treating files as virtual tables.
  • The most common enterprise pattern is replacing a small cloud warehouse (100-500 GB, hundreds of queries per day) with DuckDB plus Parquet in object storage.
  • Aggregation queries SQLite takes twenty seconds to run, DuckDB resolves in half a second; the difference is the columnar storage model.
  • Limits are clear: no write concurrency, no horizontal scale, and no native governance layer.
  • If your analytical load has less than 500 GB of active data and fewer than 10 queries per second, DuckDB deserves serious evaluation before assuming you need a distributed warehouse.

What makes it different

DuckDB is a columnar OLAP database that runs embedded in the process using it. No server, no deployment, no resource consumption when not queried. Import it as a library, open a database on disk or in memory, and run standard SQL with many analytical extensions.

The key difference with SQLite is the storage engine. SQLite is row-based and optimized for small transactions with many concurrent writes. DuckDB is columnar and optimized for analytical scans over millions of rows. An aggregation query SQLite takes twenty seconds to run, DuckDB runs in half a second. Not magic, but a storage model built for what analytics does.

The second important difference is native Parquet support. DuckDB can read Parquet files from local disk or S3 without importing them first, treating them as virtual tables. This changes the equation of many architectures: Parquet as primary storage format, DuckDB as query engine, no warehouse in between.

Replacing small warehouses

The first pattern appearing often is replacing a small cloud warehouse with DuckDB plus Parquet in object storage. In the cases accompanied, these were architectures where Snowflake or BigQuery were used with small data volumes (between 100 and 500 gigabytes), punctual query patterns (a few hundred queries per day), and monthly costs not matching real usage.

The typical replacement:

  1. Write data as partitioned Parquet in an object storage bucket.
  2. Use DuckDB in a server or function to run queries.

Storage cost drops to cents per gigabyte per month, and compute cost is limited to real query minutes. In one project, the monthly bill dropped from two thousand dollars to under one hundred, without losing real capability.

Analytics engine behind APIs

The second pattern is using DuckDB as the query engine behind data APIs. DuckDB enables an alternative to expensive OLAP replication: generate periodic extracts in Parquet, query them directly with DuckDB from the API service, and serve analytics responses with millisecond latency.

The pattern fits especially well for internal APIs or enterprise dashboards where data doesn’t have to be minute-fresh. Teams coming from Redshift or dedicated ClickHouse for this case have dropped operational complexity to zero while keeping equivalent or better response times.

Processing large files with no infrastructure

The third pattern is the most surprising because of its simplicity. DuckDB can process CSV, JSON, or Parquet files of hundreds of gigabytes from the command line with no infrastructure at all. An analyst with a laptop and DuckDB CLI can run queries on files that previously required a Spark cluster or an ETL job in the warehouse.

The internal mechanism is that DuckDB uses temporary disk spill when memory doesn’t suffice. A JOIN query between two 50-gigabyte files runs, perhaps slowly, on a laptop with 16 gigabytes of RAM.

How it fits with dbt and Python

DuckDB has integrated well with two ecosystems that particularly matter in enterprise: dbt and Python.

For dbt there is an official adapter. Many teams use this for development and testing, keeping Snowflake or BigQuery only for production. For Python, DuckDB integrates with pandas, Polars, and Arrow very smoothly. The Polars plus DuckDB combination is particularly interesting: both are high-performance analytical engines, and going from one to the other via the Arrow format is zero-copy.

Real limits

Not everything is positive. Three areas where DuckDB remains weak in 2025:

  • Write concurrency: not designed for many processes writing simultaneously. For continuous ingestion from several producers, additional patterns or alternatives like ClickHouse are needed.
  • Horizontal scale: runs in a single process, on a single machine. The practical boundary is around one or two terabytes of frequently-queried data.
  • Governance: no native permissions, auditing, or access-control layer. For regulatory compliance cases, these capabilities must be built around it.

My read

DuckDB has earned its spot in the analytics landscape. It doesn’t replace Snowflake or BigQuery in cases where scale justifies them, but it replaces underused warehouses, OLTP databases misused for analytics, Spark clusters oversized for small loads, and ClickHouse installations set up for volumes that didn’t need it.

Conclusion

DuckDB does one thing very well, and that also means there are things it doesn’t do. Recognizing where it fits and where it doesn’t is part of engineering, not a project defect.

Was this useful?
[Total: 13 · Average: 4.2]

Written by

CEO - Jacar Systems

Passionate about technology, cloud infrastructure and artificial intelligence. Writes about DevOps, AI, platforms and software from Madrid.