Sandbox
Side-by-side AI SVG comparison without casting a vote
The sandbox is the inspection page on svgbench.ai. It lets you choose a benchmark prompt and compare two model outputs directly without adding noise to the public arena vote stream.
Why the sandbox exists
Blind voting is the right tool for ranking models, but it is not always the best tool for inspection. Sometimes you want to study the output more carefully, compare two specific systems, or revisit how a model handles one prompt category. The sandbox gives you that inspection surface without turning every comparison into a recorded vote.
This makes the page useful for researchers, prompt engineers, and anyone trying to understand what the benchmark is actually seeing. Instead of inferring quality from ranking numbers alone, you can directly compare the SVGs. You can look at composition, use of the square canvas, internal spacing, shape economy, prompt fit, and whether the SVG reads well at different sizes.
What the sandbox is good for
Use the sandbox to investigate close leaderboard rivals, compare a known top model against a challenger, or inspect prompt categories such as architecture, fruit, vehicle, or animal prompts. It is especially useful when you want to understand whether a ranking gap feels deserved or when you want to inspect one output in more detail before downloading it.
Because the page is prompt-controlled, it is also a good way to separate model quality from prompt variance. If two models are shown the same benchmark prompt, any differences in clarity, balance, or recognizability become much easier to spot. That makes the sandbox especially valuable for careful SVG review.
How it fits into the rest of svgbench.ai
The sandbox is connected to the same underlying benchmark content as the arena, leaderboard, and best pages. That means you are inspecting the same generations that inform the public benchmark, not a disconnected demo environment. The difference is that the sandbox is about analysis rather than ranking.
That shared data model is important for credibility. When you see a strong or weak result in the sandbox, you are not looking at a special-case sample prepared only for presentation. You are looking at the same benchmark generation pool that influences the rest of the site. That keeps svgbench.ai internally consistent and easier to reason about.
A good workflow is to use the leaderboard to find interesting models, the best page to find strong prompt winners, and the sandbox to inspect specific pairings in more depth. If you then want to contribute directly to the benchmark, return to the homepage arena and vote on blind pairs. This keeps judgment and inspection separated in a way that makes the site easier to use and the public data easier to trust.
What this page does not do
The sandbox does not generate new public benchmark data on its own and it is not a replacement for blind voting. It is a comparison and inspection layer. That separation helps svgbench.ai stay rigorous while still being practical for people who need to look closely at SVG quality.