Updated: February 2026 | Reading time: 16 min
Introduction
Every engineering team that adopts Datadog goes through the same arc. The free trial is smooth. The initial deployment is fast. The dashboards look great. Then the first real invoice arrives, and the true Datadog log management cost does not add up. The bill is two, three, sometimes five times what anyone budgeted. This is not an edge case. It is the norm. "Datadog bill shock" has become such a common experience that it has its own meme in DevOps circles, its own Reddit threads, and its own line item in FinOps team priorities.
The root cause is not that Datadog is a bad product. It is a capable observability platform with deep integrations and a polished user experience. The problem is structural: Datadog's pricing model layers multiple independent charges, per-host fees, per-GB ingestion, per-GB indexing, per-metric costs, and retention premiums, in a way that compounds aggressively as your infrastructure scales. What feels affordable at 10GB per day becomes financially untenable at 500GB per day.
This article breaks down exactly where your money goes when you use Datadog for log management, provides real cost calculations at multiple data volumes, exposes the hidden charges that inflate your bill beyond the headline rates, and shows how an S3-native observability platform like Parseable delivers equivalent functionality at a fraction of the cost.
TL;DR: Datadog log management costs $1.80+/GB when you account for ingestion and indexing. At 500GB/day with 30-day retention, that translates to roughly $1,000,000+ per year. Parseable on S3 stores the same data for approximately $12,600 per year, a 98% cost reduction. This is not a rounding error. It is a fundamentally different cost structure made possible by S3-native architecture.
Understanding Datadog's Log Management Pricing Model
Datadog's pricing page presents a deceptively simple story. But the actual cost of running log management in production involves multiple overlapping charges, each with its own meter and its own bill line. Let us break down every fee that contributes to your total Datadog log management cost.
1. Log Ingestion: $0.10 per GB
Every byte of log data you send to Datadog incurs an ingestion fee of approximately $0.10 per GB. This is the entry toll. It does not matter whether you index the data, archive it, or immediately discard it. If Datadog's agent or API receives the data, you pay.
At first glance, $0.10 per GB sounds trivial. But consider: a moderately sized Kubernetes deployment with 50 services can easily generate 100GB of logs per day. That is $10 per day, $300 per month, or $3,650 per year just for the privilege of sending data to Datadog's endpoint. At 1TB per day, ingestion alone costs $36,500 per year.
2. Log Indexing: $1.70 per GB per Month (15-Day Retention)
This is where the real cost lives. Datadog charges approximately $1.70 per GB per month to index your logs with 15-day retention. Indexed logs are the logs you can actually search, filter, and alert on through Datadog's Log Explorer. If a log event is not indexed, it is effectively invisible to your team.
The critical detail: this is a per-GB-per-month rate, and it applies to the cumulative volume of indexed data stored at any given time. If you are ingesting 100GB per day and retaining 15 days of indexed logs, you have approximately 1,500GB of indexed data in Datadog at any point. Your monthly indexing bill is:
100 GB/day x 30 days x $1.70/GB = $5,100 per month
That is $61,200 per year just for log indexing at 100GB per day.
3. Extended Retention: Additional Charges
The default 15-day retention window on indexed logs is rarely sufficient for compliance, debugging, or trend analysis. Many organizations need 30, 60, or 90 days of searchable logs. Datadog charges additional fees for extended retention, with costs scaling roughly linearly. Extending to 30-day retention approximately doubles your indexing cost.
4. Log Archives and Rehydration
Datadog offers log archiving to external storage (S3, GCS, Azure Blob) for logs that exceed your retention window. Archiving itself is relatively cheap since you are paying your own cloud storage rates. But here is the catch: if you need to search archived logs, you must "rehydrate" them back into Datadog's indexing tier, which incurs the full indexing cost again. You are essentially paying twice: once to index the data originally, and again to re-index it when you need it later.
5. Per-Host Infrastructure Monitoring: $15-$23 per Host per Month
While not strictly a "log management" cost, nearly every team running Datadog log management also runs the Datadog infrastructure agent. The agent collects host-level metrics (CPU, memory, disk, network) and costs between $15 and $23 per host per month depending on your plan and commitment level.
For 50 hosts, that is $9,000 to $13,800 per year. For 200 hosts, $36,000 to $55,200 per year. These costs add up in the background and often get lumped into the same budget as log management.
6. Custom Metrics: $0.05 per Metric per Month
Datadog includes a base allocation of custom metrics, but once you exceed that threshold, each additional metric time series costs approximately $0.05 per metric per month. Applications with rich instrumentation can easily generate 10,000 to 50,000+ custom metrics, translating to $6,000 to $30,000 per year.
The Complete Pricing Picture
| Cost Component | Rate | Notes |
|---|---|---|
| Log Ingestion | ~$0.10/GB | Every GB sent to Datadog |
| Log Indexing (15-day) | ~$1.70/GB/month | Searchable logs |
| Extended Retention (30-day) | ~$2.50/GB/month | Approximate; varies by contract |
| Log Rehydration | ~$0.10/GB | Re-indexing archived logs |
| Infrastructure Agent | $15-$23/host/month | Per monitored host |
| APM Host | $31-$40/host/month | If using distributed tracing |
| Custom Metrics | $0.05/metric/month | Beyond base allocation |
| Indexed Spans (APM) | $1.70/GB/month | Retained trace spans |
The compounding effect of these charges is what produces Datadog bill shock. No single line item looks unreasonable in isolation. But when you multiply ingestion by indexing by retention by host count by metric volume, the total grows exponentially.
Real Datadog Log Management Cost: What You Actually Pay
Let us run the numbers at three common data volumes. These calculations assume 30-day retention on indexed logs, 50 monitored hosts, and infrastructure-only monitoring (no APM). All figures are annual.
Scenario 1: 100 GB per Day
This is a typical volume for a mid-sized SaaS company with 20-50 microservices running on Kubernetes.
| Cost Component | Monthly | Annual |
|---|---|---|
| Log Ingestion (100 GB/day x 30 days x $0.10) | $300 | $3,600 |
| Log Indexing (100 GB/day x 30 x $1.70) | $5,100 | $61,200 |
| Extended Retention Surcharge (15-day to 30-day) | ~$2,400 | ~$28,800 |
| Infrastructure Monitoring (50 hosts x $18 avg) | $900 | $10,800 |
| Custom Metrics (~5,000 metrics x $0.05) | $250 | $3,000 |
| Total | ~$8,950 | ~$107,400 |
Your annual Datadog bill at 100 GB/day: approximately $107,400.
And this is a conservative estimate. It excludes APM, RUM, synthetic monitoring, security monitoring, and any overages or spikes.
Scenario 2: 500 GB per Day
This is the range for larger engineering organizations, typically 100-300 services, multiple environments (staging, production, canary), and compliance requirements that demand comprehensive logging.
| Cost Component | Monthly | Annual |
|---|---|---|
| Log Ingestion (500 GB/day x 30 days x $0.10) | $1,500 | $18,000 |
| Log Indexing (500 GB/day x 30 x $1.70) | $25,500 | $306,000 |
| Extended Retention Surcharge (15-day to 30-day) | ~$12,000 | ~$144,000 |
| Infrastructure Monitoring (150 hosts x $18 avg) | $2,700 | $32,400 |
| Custom Metrics (~15,000 metrics x $0.05) | $750 | $9,000 |
| Total | ~$42,450 | ~$509,400 |
Your annual Datadog bill at 500 GB/day: approximately $509,400.
At this volume, many organizations start making painful trade-offs: reducing log retention, sampling aggressively, excluding non-production environments, or dropping entire log sources to keep costs manageable. These trade-offs create blind spots that only become visible during incidents.
Scenario 3: 1 TB per Day
This is enterprise scale. Large platform teams, financial services companies, e-commerce at scale, or any organization with strict audit and compliance requirements that mandate logging everything.
| Cost Component | Monthly | Annual |
|---|---|---|
| Log Ingestion (1,000 GB/day x 30 days x $0.10) | $3,000 | $36,000 |
| Log Indexing (1,000 GB/day x 30 x $1.70) | $51,000 | $612,000 |
| Extended Retention Surcharge (15-day to 30-day) | ~$24,000 | ~$288,000 |
| Infrastructure Monitoring (300 hosts x $18 avg) | $5,400 | $64,800 |
| Custom Metrics (~30,000 metrics x $0.05) | $1,500 | $18,000 |
| Total | ~$84,900 | ~$1,018,800 |
Your annual Datadog bill at 1 TB/day: over $1 million.
That is seven figures for observability alone. For context, that is the fully loaded salary of 5-8 senior engineers, or the annual infrastructure cost of a large Kubernetes cluster on AWS.
Annual Cost Summary
| Volume | Datadog Annual Cost | Cost per GB (Effective) |
|---|---|---|
| 100 GB/day | ~$107,400 | ~$2.94/GB |
| 500 GB/day | ~$509,400 | ~$2.79/GB |
| 1 TB/day | ~$1,018,800 | ~$2.79/GB |
The effective cost per GB barely decreases with scale. Datadog's pricing does not offer meaningful volume discounts for log management. Whether you are processing 100GB or 1TB per day, you are paying roughly $2.80 per GB in total cost of ownership.
The Hidden Costs Nobody Budgets For
The calculations above account for the primary line items. But Datadog's total cost of ownership includes several charges that consistently surprise teams during the first year.
APM and Distributed Tracing
If your team uses Datadog APM alongside log management (most do, since correlated traces and logs are a core selling point), add $31-$40 per APM host per month. For 100 hosts, that is an additional $37,200 to $48,000 per year. Indexed spans (retained traces) carry the same $1.70/GB/month rate as log indexing.
Real User Monitoring (RUM)
Datadog RUM charges per 10,000 sessions. For a consumer-facing web application with 1 million monthly sessions, the annual RUM cost alone can exceed $12,000. Mobile RUM has its own separate pricing tier.
Synthetic Monitoring
Synthetic tests (API checks, browser tests) are billed per 10,000 test runs. Teams running comprehensive synthetic suites across multiple regions can easily spend $5,000 to $15,000 per year on synthetic monitoring.
Security Monitoring
Datadog's Cloud SIEM product analyzes ingested logs for security signals. It is priced per GB of analyzed logs at approximately $0.20/GB. If you are already ingesting 500GB/day, enabling security monitoring adds another $36,500 per year on top of your existing ingestion and indexing costs.
Custom Metric Overages
Datadog's pricing allocates a base number of custom metrics per host. But modern applications instrumented with Prometheus client libraries or OpenTelemetry can generate thousands of metric series per service. Once you exceed the base allocation, every additional metric series costs $0.05/month. Teams routinely discover they are paying $500 to $2,000 per month in metric overages that nobody anticipated.
The Compounding Effect
Here is what a realistic "full platform" Datadog bill looks like for a 500 GB/day organization with 150 hosts:
| Component | Annual Cost |
|---|---|
| Log Management (ingestion + indexing + retention) | ~$468,000 |
| Infrastructure Monitoring (150 hosts) | ~$32,400 |
| APM (150 hosts + indexed spans) | ~$55,800 |
| Custom Metrics (15,000 series) | ~$9,000 |
| RUM (500K sessions/month) | ~$6,000 |
| Synthetic Monitoring | ~$8,000 |
| Security Monitoring | ~$36,500 |
| Total Annual Spend | ~$615,700 |
This is not a worst-case scenario. This is a team that adopted Datadog's platform broadly and uses the features that Datadog markets as core differentiators. The full-platform cost at scale is what drives the growing interest in alternatives.
The S3-Native Alternative: Parseable Cost Breakdown
Parseable takes a fundamentally different approach to observability economics. Instead of building a proprietary storage layer and charging per-GB premiums, Parseable writes directly to S3-compatible object storage in Apache Parquet columnar format. Your telemetry data lives in your own storage, at your storage provider's rates.
How Parseable's Architecture Changes the Cost Equation
The architectural difference is not incremental; it is structural:
- No ingestion fee: Parseable does not charge per GB ingested. Data flows in via standard HTTP/OTLP endpoints.
- No indexing fee: Parseable uses columnar storage (Parquet) and query-time processing. There is no separate "indexing tier" with its own billing meter.
- No per-host fee: Self-hosted Parseable runs as a single binary. You pay for compute and storage, not per monitored host.
- S3 storage rates: Your data is stored on S3 (or any S3-compatible object store) at the standard rate of approximately $0.023 per GB per month.
Parseable Cost at Scale
Let us run the same three scenarios using Parseable self-hosted on AWS with S3 storage.
Assumptions:
- Parseable runs on EC2 instances sized for the workload
- Log data is stored on S3 Standard
- Parquet columnar format provides approximately 10:1 compression on typical log data
- 30-day retention (same as the Datadog scenarios)
Scenario 1: 100 GB/day on Parseable
| Cost Component | Monthly | Annual |
|---|---|---|
| S3 Storage (100 GB/day x 30 days / 10 compression x $0.023) | ~$6.90 | ~$83 |
| S3 API Costs (PUT/GET requests) | ~$15 | ~$180 |
| EC2 Compute (2x c5.xlarge) | ~$245 | ~$2,940 |
| Data Transfer (internal) | ~$25 | ~$300 |
| Total | ~$292 | ~$3,503 |
Parseable annual cost at 100 GB/day: approximately $3,503.
Compare that to Datadog's $107,400. That is a 97% cost reduction.
Scenario 2: 500 GB/day on Parseable
| Cost Component | Monthly | Annual |
|---|---|---|
| S3 Storage (500 GB/day x 30 days / 10 compression x $0.023) | ~$34.50 | ~$414 |
| S3 API Costs (PUT/GET requests) | ~$50 | ~$600 |
| EC2 Compute (3x c5.2xlarge) | ~$735 | ~$8,820 |
| Data Transfer (internal) | ~$75 | ~$900 |
| Total | ~$895 | ~$10,734 |
Parseable annual cost at 500 GB/day: approximately $10,734.
Compare that to Datadog's $509,400. That is a 98% cost reduction.
Scenario 3: 1 TB/day on Parseable
| Cost Component | Monthly | Annual |
|---|---|---|
| S3 Storage (1,000 GB/day x 30 days / 10 compression x $0.023) | ~$69 | ~$828 |
| S3 API Costs (PUT/GET requests) | ~$100 | ~$1,200 |
| EC2 Compute (4x c5.4xlarge) | ~$1,960 | ~$23,520 |
| Data Transfer (internal) | ~$150 | ~$1,800 |
| Total | ~$2,279 | ~$27,348 |
Parseable annual cost at 1 TB/day: approximately $27,348.
Compare that to Datadog's $1,018,800. That is a 97% cost reduction.
Side-by-Side Annual Cost Comparison
| Daily Volume | Datadog Annual Cost | Parseable Annual Cost | Annual Savings | Savings % |
|---|---|---|---|---|
| 100 GB/day | ~$107,400 | ~$3,503 | $103,897 | 97% |
| 500 GB/day | ~$509,400 | ~$10,734 | $498,666 | 98% |
| 1 TB/day | ~$1,018,800 | ~$27,348 | $991,452 | 97% |
These are not theoretical numbers. They are the direct consequence of replacing per-GB indexing fees with S3 storage rates. The math is straightforward and verifiable by any infrastructure team with access to the AWS S3 pricing calculator.
What Parseable Actually Delivers
Parseable is not a stripped-down logging tool that saves money by cutting features. It is a unified observability platform that handles the same signal types as Datadog.
Full MELT Observability
Parseable supports Logs, Metrics, Events, and Traces (MELT) through a single platform. This is not a collection of loosely integrated tools. All four signal types flow through the same ingestion pipeline, are stored in the same S3 backend, and are queryable through the same SQL interface. Cross-signal correlation, the ability to jump from a log line to a related trace span to the underlying metric, is built in.
SQL-Native Query Language
Parseable uses standard SQL for all queries. There is no proprietary query language to learn, no DQL (Datadog Query Language) equivalent that locks your team's knowledge into a single vendor. Any engineer who knows SQL can be productive with Parseable on day one. This also means your queries, dashboards, and alerts are portable.
OpenTelemetry Native
Parseable supports native OTLP ingestion, which means you can point any OpenTelemetry Collector or SDK directly at Parseable without proprietary agents. If you are already using OpenTelemetry for instrumentation (and in 2026, most teams are), migrating to Parseable requires zero re-instrumentation. Change the OTLP endpoint, and your telemetry flows to Parseable.
Parseable Cloud
For teams that want managed observability without operating infrastructure, Parseable Cloud provides a fully managed option starting at $0.37/GB ingested ($29/month minimum) with a free tier. You get the same S3-native architecture, the same SQL query interface, and the same cost advantages, without managing EC2 instances or S3 buckets yourself.
What You Could Do with the Savings
The cost difference between Datadog and Parseable at scale is not a minor budget optimization. It is transformative. Here is what the savings translate to in concrete terms.
At 100 GB/day: ~$103,900 Annual Savings
- Hire a senior SRE to strengthen your on-call rotation
- Fund your entire staging environment for a year
- Cover the cost of 3-4 additional production Kubernetes nodes
- Invest in load testing infrastructure you have been putting off
At 500 GB/day: ~$498,700 Annual Savings
- Build a 3-person platform engineering team
- Fund a complete migration from legacy infrastructure to Kubernetes
- Cover the annual compute cost of a 50-node production cluster
- Invest in a comprehensive disaster recovery setup
At 1 TB/day: ~$991,500 Annual Savings
- Fund an entire engineering squad (6-8 engineers) for a year
- Build out multi-region infrastructure with full redundancy
- Cover the compute budget for a large-scale ML training pipeline
- Invest in R&D initiatives that have been deprioritized due to budget constraints
The point is not that observability is unimportant and should be minimized. The point is that paying $1 million per year for observability when the same capability is available for $27,000 represents a massive misallocation of engineering budget. Those savings can be redirected to work that actually moves your product forward.
Migration Path: Datadog to Parseable
Migrating from Datadog to Parseable does not require a big-bang cutover. The recommended approach is a gradual, parallel deployment that minimizes risk.
Phase 1: Deploy Parseable (Week 1)
The fastest option is Parseable Cloud — starts at $0.37/GB ingested ($29/month minimum) with a free tier, no infrastructure to manage. To self-host, Parseable deploys as a single binary via Docker, Kubernetes Helm chart, or bare metal. A production-ready instance can be running in under 30 minutes.
# Docker deployment
docker run -p 8000:8000 \
-e P_S3_URL=https://s3.amazonaws.com \
-e P_S3_ACCESS_KEY=your-access-key \
-e P_S3_SECRET_KEY=your-secret-key \
-e P_S3_BUCKET=parseable-logs \
-e P_S3_REGION=us-east-1 \
parseable/parseable:latest \
parseable s3-storePhase 2: Dual-Ship Telemetry (Weeks 2-4)
Configure your OpenTelemetry Collector to export to both Datadog and Parseable simultaneously. This lets your team validate that Parseable is receiving all expected data, that queries return correct results, and that alerting works as expected, all while Datadog remains your primary platform.
# otel-collector config snippet
exporters:
otlphttp/parseable:
endpoint: "https://your-parseable-instance:8000"
datadog:
api:
key: ${DD_API_KEY}
service:
pipelines:
logs:
exporters: [otlphttp/parseable, datadog]Phase 3: Migrate Dashboards and Alerts (Weeks 3-6)
Rebuild your critical dashboards and alerts in Parseable. Since Parseable uses SQL, many Datadog queries translate directly. Start with your top 10 most-used dashboards and your critical alert rules.
Phase 4: Cut Over (Weeks 6-8)
Once your team is comfortable with Parseable and has validated data completeness, stop shipping to Datadog. Cancel your Datadog contract at the next renewal window.
Phase 5: Decommission and Optimize (Weeks 8-12)
Remove Datadog agents, clean up configurations, and optimize your Parseable deployment based on actual query patterns and data volumes. This is also the time to enable features you could not afford on Datadog, such as logging non-production environments, extending retention to 90 or 180 days, or enabling full-fidelity tracing without sampling.
Migration Risk Mitigation
- Data validation: During the dual-ship phase, compare query results between Datadog and Parseable for the same time ranges to ensure data completeness.
- Gradual rollout: Migrate one team or service at a time rather than switching the entire organization at once.
- Rollback plan: Keep your Datadog contract active for one billing cycle after cutover so you can revert if needed.
Frequently Asked Questions
Is Parseable really a Datadog alternative, or is it just a logging tool?
Parseable is a unified observability platform that handles logs, metrics, events, and traces. It competes directly with Datadog on the full observability stack, not just log management. The key difference is architectural: Parseable uses S3-native storage instead of proprietary indexing, which eliminates the per-GB fees that make Datadog expensive. For a detailed comparison of Datadog alternatives, see our guide on Datadog alternatives for log analytics.
How does Parseable's query performance compare to Datadog at scale?
Parseable uses Apache Parquet columnar format, which is highly optimized for analytical queries. For typical observability queries (filtering by time range, service name, log level, and searching for specific patterns), Parseable delivers sub-second response times on datasets of hundreds of millions of rows. The columnar format means Parseable only reads the columns relevant to your query, which makes it efficient even on cold S3 storage. For time-sensitive queries, Parseable maintains a hot cache layer that accelerates the most recent data.
What about Datadog's 750+ integrations? Does Parseable support those?
Parseable supports native OTLP (OpenTelemetry Protocol) ingestion, which means any system instrumented with OpenTelemetry can send data directly to Parseable. Since OpenTelemetry has become the industry standard, the effective integration surface is comparable. For infrastructure metrics, Parseable works with the OpenTelemetry Collector's 200+ receivers, covering all major cloud providers, databases, message queues, and orchestration tools. You do not need proprietary agents.
Can Parseable handle compliance requirements like HIPAA, SOC 2, and GDPR?
Because Parseable stores data in your own S3 buckets, you maintain full control over data residency, encryption, access policies, and retention. This is actually a compliance advantage over Datadog, where your telemetry data lives on a third party's infrastructure. With Parseable self-hosted, your data never leaves your VPC. You can apply your existing S3 bucket policies, KMS encryption, VPC endpoints, and IAM roles to your observability data, which simplifies compliance audits significantly.
What if I want managed observability but not Datadog pricing?
Parseable Cloud provides a fully managed observability experience with the same S3-native architecture. You get managed ingestion, storage, querying, and alerting without operating any infrastructure yourself. Pricing is based on storage consumed rather than per-GB indexing fees, so the cost advantage over Datadog is preserved. Teams looking to compare multiple options should also review our Splunk alternatives guide.
How long does the migration from Datadog to Parseable typically take?
Most teams complete the migration in 6-8 weeks, including a 2-4 week parallel-run period where both systems receive data simultaneously. The actual technical migration (deploying Parseable, configuring OTLP export, rebuilding dashboards) typically takes 1-2 weeks of focused effort. The remaining time is spent on validation, team training, and building confidence before cutting over. Since Parseable uses SQL, engineers find the learning curve minimal compared to learning a new proprietary query language.
The Bottom Line on Datadog Log Management Cost
Datadog's log management pricing is designed for a world where data volumes are small and per-GB fees feel affordable. That world no longer exists. Modern infrastructure generates massive telemetry volumes, and Datadog's pricing model turns that growth into a financial penalty.
The math is clear:
- At 100 GB/day, Datadog costs ~$107,400/year. Parseable costs ~$3,503/year.
- At 500 GB/day, Datadog costs ~$509,400/year. Parseable costs ~$10,734/year.
- At 1 TB/day, Datadog costs ~$1,018,800/year. Parseable costs ~$27,348/year.
The savings come from architecture, not from cutting corners. Parseable delivers full MELT observability (logs, metrics, events, and traces), SQL-native querying, native OpenTelemetry support, and complete data sovereignty, all at S3 storage rates.
If your Datadog bill has become a budget line item that leadership questions every quarter, it is time to evaluate whether you are paying for capability or paying for a pricing model.
Ready to cut your observability costs by 90% or more?
- Start with Parseable Cloud — starts at $0.37/GB, free tier available
- Self-hosted deployment — single binary, deploy in 2 minutes
- Read the docs — guides, API reference, and tutorials
- Join our Slack — community and engineering support
See how Parseable compares to other platforms: Datadog Alternatives | Splunk Alternatives | Open Source Log Management Tools


