Performance Benchmarks

Performance Benchmark

Chaperone-Thinking-LQ-1.0 delivers world-class reasoning performance — rivaling the best, while being 1.6x faster.

Performance Benchmarks

Chaperone-Thinking-LQ-1.0 is our new LLM engineered for deep reasoning, scientific precision, and real-world deployability. Despite its lightweight, it matches — and in some benchmarks even rivals — the performance of today’s leading reasoning models from OpenAI and DeepSeek. Tested on rigorous datasets like AIME 2024, MATH-500, GPQA Diamond, and MMLU, it delivers state-of-the-art results while remaining efficient and cost-effective. Ideal for research workflows, technical assistants, and high-stakes enterprise applications.

1.6x Higher Throughput

In our side-by-side tests, Chaperone-Thinking-LQ-1 consistently responded faster than DeepSeek-R1-Distill-Qwen-32B, delivering answers in about 11–12 seconds versus roughly 20 seconds for the alternative—around 1.6× quicker overall. While the other model tended to produce slightly longer replies, Chaperone-Thinking-LQ-1 still provided results notably sooner, offering a snappier, more efficient experience for everyday use.

Throughput
Chaperone-Thinking-LQ-1.0
36.86 tok/s
DeepSeek-R1-Distill-Qwen-32B
22.84 tok/s
+61.4% vs DeepSeek
Latency p50
Chaperone-Thinking-LQ-1.0
11.49 s
DeepSeek-R1-Distill-Qwen-32B
20.10 s
−42.8% lower
Latency p95
Chaperone-Thinking-LQ-1.0
13.06 s
DeepSeek-R1-Distill-Qwen-32B
20.11 s
−35.0% lower
Benchmark Metric / Task Chaperone-Thinking-LQ-1.0 DeepSeek-R1-Distill-Qwen-32B
Speed & Latency Throughput (tok/s) 36.86 22.84
Latency p50 (s) 11.49 20.10
Latency p95 (s) 13.06 20.11
Tokens / Request (avg) 435.4 459.0

Averages over 10 trials, with concurrency=1, max_tokens=512, and temperature=0.

Strong Performance

Chaperone-Thinking-LQ-1.0 excels in generative tasks and accuracy-based evaluations, particularly in GPQA benchmarks.

Reasoning Excellence

Demonstrates superior performance in complex reasoning tasks requiring deep understanding and logical processing.

Fast Execution

Optimized for quick processing with minimal overhead, delivering reliable results at high speed across tasks.

Ready to leverage our cutting-edge reasoning model?

Get Started
right-image