Zelili AI

Intern-S1-Pro Is Now Open-Source: Here Is How This 1T MoE Model Can Transform Scientific Research

Intern-S1-Pro Is Now Open-Source

Summary Box Summary Box [In a hurry? Just read this⚔]

  • Shanghai AI Laboratory released Intern-S1-Pro, a massive open-source 1 trillion parameter MoE multimodal model with 22B active parameters.
  • It excels in scientific reasoning and AI4Science tasks, often outperforming or matching top closed-source models on specialized benchmarks.
  • Key innovations include STE Routing, Grouped Routing, FoPE (Fourier Position Encoding), and advanced time-series modeling for heterogeneous signals.
  • The model supports multimodal inputs (images, time-series, text) and integrates with tools like vLLM and SGLang for easy deployment.
  • It shows strong performance across scientific domains (biology, chemistry, physics, earth science) and remains competitive in general knowledge, math, and agent tasks.

Shanghai AI Laboratory has launched Intern-S1-Pro, an advanced open-source multimodal large language model with a massive 1 trillion parameters in a Mixture-of-Experts (MoE) architecture.

This release, announced in early February 2026, positions the model as a leader in scientific reasoning and AI4Science tasks, where it competes effectively against prominent closed-source alternatives.

With an active parameter count of 22 billion, Intern-S1-Pro emphasizes efficiency, stability, and broad applicability across scientific and general domains.

Designed for high-performance multimodal tasks, the model integrates sophisticated techniques to handle complex data types, including images, time-series signals, and text.

It supports ecosystem tools like vLLM and SGLang, facilitating easy deployment and inference.

Developers can access the model weights and code through major platforms, enabling rapid integration into research and production environments.

Key Innovations Driving Performance

Intern-S1-Pro incorporates several cutting-edge features to enhance its capabilities:

  • STE Routing: Enables dense gradient updates for router training, improving overall model efficiency.
  • Grouped Routing: Ensures stable convergence and balanced load distribution across experts during parallel processing.
  • Fourier Position Encoding (FoPE): Combined with enhanced time-series modeling, this allows precise representation of physical signals, supporting heterogeneous sequences from 10^0 to 10^6 points.
  • Multimodal Integration: Strong handling of visual, textual, and sequential data for comprehensive scientific analysis.

These advancements contribute to the model’s top-tier results on advanced reasoning benchmarks, making it a valuable tool for researchers in fields like biology, chemistry, physics, and earth sciences.

Benchmark Performance on Scientific Tasks

Intern-S1-Pro demonstrates superior performance across a range of scientific benchmarks, often outperforming or matching leading models. The following table summarizes scores on key scientific tasks:

BenchmarkDescriptionIntern-S1-Pro (1T-A22B)Qwen3-VL-235B-ThinkingKimi-K2.5 (1T-A32B)GPT-5.2Gemini-3-Pro
SciReasonerScientific Reasoning55.511.915.313.614.7
SFEScientific Multimodal Tasks52.741.453.747.558.9
SmolInstructSmall Molecule74.836.653.548.258.3
MatBenchMaterials Property Prediction72.849.760.053.664.9
Mol-InstructionsBio-molecular Instruction48.88.920.012.334.6
MicroVQABiological Microscopy63.353.855.460.469.0
Biology-InstructionMulti-Omics Sequence52.56.210.710.212.0
ZRemoteBenchRemote Sensing67.851.246.450.451.8
MSEarth-MCQEarth Science56.252.761.962.665.8

Benchmark Performance on General Tasks

In general tasks, Intern-S1-Pro also shows competitive edges, particularly in knowledge, reasoning, and visual grounding:

BenchmarkDescriptionIntern-S1-Pro (1T-A22B)Qwen3-VL-235B-ThinkingKimi-K2.5 (1T-A32B)GPT-5.2Gemini-3-Pro
MMMU-ProKnowledge & Reasoning72.869.978.579.381.0
MMLU-ProKnowledge & Reasoning86.683.487.185.989.3
AIME-2025Math Reasoning93.190.096.1100.095.0
IMO-Answer-BenchMath Reasoning77.372.381.886.381.3
RefCOCO-avgVisual Grounding91.991.187.854.976.2
IFBenchInstruction Following71.258.769.775.470.4
OCRBench V2 (ENG / CHN)OCR Generation60.1 / 60.666.8 / 63.864.2 / 57.456.4 / 54.668.0 / 52.5
SArena (Icon)SVG Generation83.576.377.355.582.6
KodingCode74.372.085.087.786.9
Coder (Text-Only)Agent73.447.879.971.175.5
Tau²-BenchAgent80.957.476.876.685.4
ScreenSpot V2Agent & Grounding93.692.892.449.494.7

Implications for AI Research and Applications

Intern-S1-Pro Huggingface

The benchmarks reveal Intern-S1-Pro‘s strengths in specialized scientific domains, where it often leads in efficiency and accuracy.

For general tasks, it holds its own against giants like GPT-5.2 and Gemini-3-Pro, particularly in math and visual tasks.

This open-source release democratizes access to powerful multimodal AI, fostering innovation in scientific discovery, materials design, and beyond.

Researchers can leverage its MoE efficiency for scalable deployments, while the advanced encodings open new possibilities for time-series analysis in physics and biology.