Skip to content

Add fluss-microbench module with YAML-driven perf framework and JMH benchmarks#3082

Open
platinumhamburg wants to merge 1 commit intoapache:mainfrom
platinumhamburg:feature/fluss-perf
Open

Add fluss-microbench module with YAML-driven perf framework and JMH benchmarks#3082
platinumhamburg wants to merge 1 commit intoapache:mainfrom
platinumhamburg:feature/fluss-perf

Conversation

@platinumhamburg
Copy link
Copy Markdown
Contributor

@platinumhamburg platinumhamburg commented Apr 15, 2026

Unified benchmarking module combining YAML-driven performance testing (write/lookup/scan/mixed workloads with resource sampling and reporting) and JMH micro-benchmarks (Arrow channel, log scanner).

Key features:

  • MICROBENCH_ROOT-based output layout (.microbench/runs/datasets/baselines)
  • Auto-managed latest/previous symlinks per scenario for iterative comparison
  • --diff-previous and --diff-baseline flags for regression detection
  • Forked server log capture (logs/server-0.log) and run.log tee
  • 5 built-in presets: log-append, kv-upsert-get, kv-agg-mixed/listagg/rbm32
  • Contributor guide at website/community/dev/benchmarking.md

CLI design: workload parameters (records, threads, warmup, duration) are YAML-only and cannot be overridden via CLI, ensuring every benchmark report is statistically reliable. Warmup semantics are per-phase global (divided by threads internally). Validator enforces warmup < records/2, effective records per thread >= 10,000, and unique phase names per scenario.

Phase names support type-prefix convention (e.g. write-bulk, write-sustained) resolved via longest-prefix matching to PhaseType. Preset tuning: thread counts capped at 8/4, warmup 500K global for JIT C2 coverage, duration-based write phases (5m) for stable measurement windows, scan warmup for cold-start elimination.

Purpose

Linked issue: close #3080

Brief change log

Tests

API and Format

Documentation

…enchmarks

Unified benchmarking module combining YAML-driven performance testing
(write/lookup/scan/mixed workloads with resource sampling and reporting)
and JMH micro-benchmarks (Arrow channel, log scanner).

Key features:
- MICROBENCH_ROOT-based output layout (.microbench/runs/datasets/baselines)
- Auto-managed latest/previous symlinks per scenario for iterative comparison
- --diff-previous and --diff-baseline flags for regression detection
- Forked server log capture (logs/server-0.log) and run.log tee
- 5 built-in presets: log-append, kv-upsert-get, kv-agg-mixed/listagg/rbm32
- Contributor guide at website/community/dev/benchmarking.md

CLI design: workload parameters (records, threads, warmup, duration) are
YAML-only and cannot be overridden via CLI, ensuring every benchmark report
is statistically reliable. Warmup semantics are per-phase global (divided
by threads internally). Validator enforces warmup < records/2, effective
records per thread >= 10,000, and unique phase names per scenario.

Phase names support type-prefix convention (e.g. write-bulk, write-sustained)
resolved via longest-prefix matching to PhaseType. Preset tuning: thread
counts capped at 8/4, warmup 500K global for JIT C2 coverage, duration-based
write phases (5m) for stable measurement windows, scan warmup for cold-start
elimination.
@platinumhamburg platinumhamburg changed the title Add fluss-microbench module with YAML-driven perf framework and JMH b… Add fluss-microbench module with YAML-driven perf framework and JMH benchmarks Apr 15, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Add fluss-microbench module: YAML-driven performance testing framework

1 participant