New round of X-Series (Uchoa et al.) benchmarks - this time with full results available on our Discord. Next week we will start publishing about enterprise use cases, such as multi-target optimization for heat treatment plants as already advertised on our official SolverForge.ai website. For the benchmarks, link in comments! 👇
We've hit an important milestone with SolverForge v0.8.6. For a while, we had good CVRP benchmark numbers, but there was still an asterisk: those runs were going through manually wired internals in the benchmark harness, not the full public API surface. That is no longer true. The latest results are running through the same user-facing path an actual integration would use: - declarative model - solver.toml - retained runtime - stock list-variable pipeline And the numbers are very strong through that public surface. On the CVRPLIB X dataset (2017, Uchoa et al.): - 100 instances - 1s / 10s / 60s budgets - 300 total SolverForge runs - 0 invalid solutions Average ratio to best-known cost: - 1.0555 @ 1s - 1.0532 @ 10s - 1.0515 @ 60s Other numbers that matter if you care about CVRP: - 288 / 300 runs finished within 10% of best-known - on the large instances (n >= 600), the 60s mean ratio was 1.0491 - on X-n1001-k43, SolverForge came in at 1.0628 @ 1s, 1.0628 @ 10s, 1.0624 @ 60s What is exciting here is not just the benchmark quality. It is that these numbers are now coming from the official API surface of a generic solver, not from hand-wired benchmark internals. This is literally doable with solverforge-cli in a couple of hours! To me, that is the real threshold: a solver is only truly product-ready when the public path preserves the performance of the core engine. That is the standard we are trying to hold in SolverForge: the best path should be the official path.