December 21, 2025
Announcements
Introducing Smars-1, a powerful, all-around system for advanced intelligence and real-world problem solving. As the first model released in its series, Smars-1 establishes a high standard from the outset, combining strong reasoning capabilities with a refined communication style to deliver a balanced and capable AI experience.
Smars-1 is built as a hybrid model, enabling automatic internal reasoning to handle complex tasks with depth and accuracy. This architecture allows effective management of multi-step projects, strong tool usage, and reliable writing and reasoning about code, making it suitable for both technical and non-technical use cases.
The model demonstrates state-of-the-art performance across multiple benchmarks and is designed as a multimodal system, capable of analyzing and understanding images alongside text. This combination positions Smars-1 ahead of existing models in overall capability, emphasizing versatility and consistent performance across a wide range of tasks rather than narrow specialization.
In addition to intelligence, Smars-1 prioritizes interaction quality. The model is warmer, more instruction-aligned, and more enjoyable to engage with, combining strong analytical ability with clear, natural communication. This balance of reasoning strength and conversational skill makes Smars-1 both highly capable and pleasant to work with across professional and creative domains.
| Benchmark | Smars-1 | GPT-5.2 | Claude 4.5 Sonnet |
Gemini 3 Pro |
|---|---|---|---|---|
|
GPQA Diamond (no tools)
Scientific Reasoning
|
89.6% | 92.4% | 83.4% | 91.9% |
|
AIME 2025 (no tools)
Competitive Maths
|
96.3% | 100.0% | 87.0% | 95.0% |
|
MMMLU
Multilingual Reasoning
|
90.4% | 89.6% | 89.1% | 91.8% |
|
SWE-Bench Verified
Agentic coding
|
73.9% | 80.0% | 77.2% | 76.2% |
|
LiveCodeBench Pro (Elo Rating)
Competitive Programming
|
1,551 | 1,580 | 1,418 | 2,439 |
|
τ2-bench
Agentic tool use
|
79.4% | - | 84.7% | 85.4% |
|
Terminal-Bench 2.0
Agentic terminal coding
|
45.5% | - | 42.8% | 54.2% |
|
HMMT 2025
Elite competition mathematics
|
92.5% | 99.4% | 84.5% | 92.3% |
|
ARC-AGI-2 (Verified)
Visual Logic Challenges
|
16.2% | 52.9% | 13.6% | 31.1% |
|
MMMU-Pro
Multimodal Intelligence
|
61.1% | 79.5% | 68.0% | 81.0% |
|
ScreenSpot-Pro
Visual Interface Reasoning
|
28.7% | 86.3% | 36.2% | 72.7% |
*Benchmarking was performed in a research environment, so results may not exactly match production Smars output.
Smars-1 is built as a fine-tuned system derived from multiple open-source AI models, followed by targeted merging and optimization. The architecture follows a Mixture-of-Experts (MoE) design, allowing different specialized components to activate dynamically based on task requirements. This approach enables strong performance, efficiency, and adaptability while maintaining openness about the model’s foundations.
Overall, Smars-1 represents a major advancement in general intelligence, hybrid reasoning, multimodal understanding, and agentic tool usage, enabling reliable end-to-end execution of complex, real-world tasks at a level exceeding existing models.
Smars-1 model is built on a strong, well-established open-source base model licensed under Apache 2.0, enabling development to focus on real-world performance rather than rebuilding core infrastructure. Starting from a trusted foundation accelerates progress while maintaining stability, transparency, and reliability.
The base model is enhanced through extensive fine-tuning across chat, coding, mathematics, and instruction-following tasks using high-quality proprietary datasets. Significant effort is dedicated to data curation, filtering, and balancing to improve practical performance across diverse use cases.
The model incorporates a hybrid reasoning architecture that adapts to task complexity. Straightforward requests are handled efficiently, while more demanding problems—such as multi-step reasoning, coding, mathematics, or planning—engage deeper reasoning to improve accuracy and coherence.
For complex tasks, the application presents step-by-step reasoning explanations, allowing users to follow how conclusions are reached. This provides greater transparency for technical and analytical workflows while maintaining controlled, interpretable outputs rather than exposing raw internal model states.
Smars-1 supports tool calling for tasks that benefit from external capabilities, including web search, data retrieval, and integration with downstream systems. When tools are available, the model can decide when to invoke them and how to incorporate returned results into its responses, enabling more up-to-date and context-aware outputs.
Smars-1 is designed to prioritize accuracy, clarity, and intellectual honesty over passive agreement or superficial politeness. The model aims to provide direct, actionable, and evidence-based responses, even when this requires challenging assumptions, pointing out logical inconsistencies, or correcting flawed reasoning.
Rather than acting as a simple affirmation engine, Smars-1 is optimized to support learning, critical thinking, and informed decision-making by emphasizing reasoning quality and factual grounding over comfort-oriented or generic responses.
Mixture-of-Experts techniques are used selectively to support efficiency and scalability. This architectural approach allows different components of the model to be activated for different types of inputs, helping manage resource utilization while maintaining stable behavior across a wide range of tasks. Selective expert routing introduces additional flexibility in scaling the system and supports reliable deployment in production environments.
The model is trained for multilingual understanding and generation across a wide range of languages. It can follow instructions, reason, and produce high-quality outputs in multiple languages, as well as handle mixed-language inputs within a single interaction.
Vision functionality is provided through a dedicated image encoder aligned with the language model, enabling interpretation and reasoning over visual inputs alongside text. This modular design supports continued improvement of visual understanding as the technology advances.
The model supports context windows of up to 128k tokens, enabling it to process long conversations, large documents, and complex multi-step tasks in a single interaction. As with all large language models, maintaining reliability, coherence, and instruction adherence over very long contexts remains an active area of development across the field.
Smars-1 is available to all users through the Smars app, with no subscriptions or paid plans required.
To ensure consistent performance and a reliable experience for everyone, daily usage limits are applied. Each user can send up to 30 messages per day and upload up to 10 files per day. These limits help maintain responsiveness, stability, and fair access as demand fluctuates.
All users have access to the full Smars-1 model within these limits. Once the daily limit is reached, additional interactions pause until the next daily reset.
An API for Smars-1 is planned and will be made available in a future release, enabling developers to integrate the model directly into their own applications and workflows.
Usage limits and access policies may evolve over time as the product grows and new capabilities are introduced, with the goal of maintaining quality and reliability across the platform.