Go back to all articles

Benchmark Testing Case Study: How PFLB Validated Chainguard Docker Container Performance

3 min read
author denis sautin preview

Denis Sautin

Author

Denis Sautin

Denis Sautin is an experienced Product Marketing Specialist at PFLB. He focuses on understanding customer needs to ensure PFLB’s offerings resonate with you. Denis closely collaborates with product, engineering, and sales teams to provide you with the best experience through content, our solutions, and your personal journey on our website.

Product Marketing Specialist

Reviewed by Boris Seleznev

boris author

Reviewed by

Boris Seleznev

Boris Seleznev is a seasoned performance engineer with over 10 years of experience in the field. Throughout his career, he has successfully delivered more than 200 load testing projects, both as an engineer and in managerial roles. Currently, Boris serves as the Professional Services Director at PFLB, where he leads a team of 150 skilled performance engineers.

Chainguard specializes in providing highly secure Docker container images — regularly updated, tightly managed, and designed to minimize vulnerabilities. While their reputation for security is well-established, Chainguard wanted to ensure their containers also delivered competitive performance. After all, strong security is crucial, but it shouldn’t slow you down.

Client Goal

chainguard logo

Chainguard needed clear, objective proof that their Docker containers don’t compromise on speed or efficiency. They wanted transparent benchmark testing against popular, widely-used images to validate performance, not just security. To make this credible, Chainguard looked for an unaffiliated expert — someone trusted and independent — to carry out detailed, unbiased testing.

Why Chaingurad Chose PFLB?

Chainguard needed a partner they could trust completely — someone neutral, thorough, and dedicated to rigorous testing. That’s exactly where PFLB shines:

  • Completely Impartial:
    At PFLB, we don’t play favorites. Our testing is transparent, neutral, and never influenced by vendor interests. When we say something performs well, it’s backed purely by facts.
  • Robust and Repeatable Testing:
    We take benchmark tesitng seriously, using meticulously crafted methods that anyone can reproduce. Our approach ensures every test is consistent and credible, giving you results you can rely on.
  • Clear, Actionable Insights:
    Numbers alone don’t tell the full story. That’s why our reports clearly highlight strengths, pinpoint weaknesses, and offer straightforward insights. We help clients quickly understand exactly how their technology stacks up.

Discover PFLB Benchmark Testing Services

Learn more

Scope of Work

To give Chainguard clear, trustworthy answers, PFLB benchmarked a variety of database Docker containers against their counterparts from DockerHub. We selected databases representing different real-world workloads and usage scenarios:

  • PostgreSQL (tested on both ARM and x86 architectures)
  • CockroachDB
  • MariaDB
  • MySQL
  • Redis
  • Memcached
  • ClickHouse
  • MongoDB

Testing PostgreSQL on both ARM and x86 architectures allowed us to offer additional insights, so performance results were relevant across different infrastructure setups and use cases.

Our Benchmark Testing Methodology

At PFLB, we pride ourself on our testing being precise, structured, and repeatable. We ensured that benchmarking for Chainguard was both comprehensive and fair:

Testing Environment Setup

We started by deploying consistent, cloud-based infrastructure using AWS virtual machines. Standardizing our testing environment meant every Docker container had the same opportunity to showcase its true performance — no hidden variables, no surprises.

  • Consistent AWS infrastructure
  • Uniform Docker configurations
  • Standardized benchmarking and monitoring setup

Benchmarking Tools We Used

To accurately reflect real-world database use cases, we selected reliable, industry-recognized benchmarking tools tailored to each database type:

  • Benchbase: For relational databases like PostgreSQL, CockroachDB, MariaDB, MySQL, and ClickHouse, running widely recognized TPC benchmarks:
    • TPC-C: Transactional workloads, measuring real-world database responsiveness.
    • TPC-H: Analytical workloads, evaluating performance for complex data queries.
  • Memtier_benchmark: Specifically chosen for Redis and Memcached, allowing us to run realistic read-heavy scenarios, mirroring typical production environments.

Rigorous Test Execution

We ran each test multiple times — four to five iterations per scenario — to ensure the accuracy and statistical reliability of our results. By varying workloads and scenarios, we covered a wide spectrum of realistic performance demands:

  • Transactional workloads (TPC-C): Simulated everyday business operations, evaluating responsiveness under pressure.
  • Analytical workloads (TPC-H): Complex, resource-intensive queries to test sustained analytical performance.
  • Read-heavy scenarios (Redis, Memcached): Assessing high-volume read performance typical of caching scenarios.

Comprehensive Performance Monitoring

Numbers alone don’t paint the whole picture. We closely monitored multiple dimensions of system performance, giving Chainguard a clear view into how each container performed across several critical metrics:

  • System-level metrics: CPU usage, RAM consumption, disk I/O performance, network throughput.
  • Docker-specific metrics: Individual container CPU and memory utilization, ensuring detailed resource-use insights.
  • Latency and throughput measurements: Including both raw throughput (operations/sec) and “goodput” (successful transactions), alongside detailed latency distribution insights.

Reporting & Conclusions

benchmark testing project

Dedicated Reports

Each database received its own detailed performance report, highlighting key metrics. Clear comparison tables and precise visualizations helped Chainguard quickly understand the strengths and weaknesses of their containers compared to standard Docker images.

Unified Heatmap

We complemented these detailed documents with a unified heatmap, providing a clear, visual summary across all tested databases and scenarios. This heatmap instantly highlighted strengths, weaknesses, and opportunities for improvement.

Consistent Methodology

Uniform testing across every scenario ensured clear and reliable conclusions. Chainguard could confidently compare results across different databases, easily pinpointing areas requiring further optimization or resource allocation.

Client Benefits & the PFLB Advantage

Chainguard received clear, unbiased benchmarking data to confidently inform their product decisions and optimizations. With our rigorous testing and transparent reporting, they gained reliable proof points to support their performance claims.

By partnering with PFLB, Chainguard demonstrated their commitment to credible, neutral validation. Our detailed, actionable insights empowered their teams to optimize effectively, without any vendor bias or promotional influence.

Interested in Benchmarking Your Product?

Related insights in in case studies

Explore what we’ve learned from these experiences

More Case Studies