
Lei Bai, Zhongrui Cai, ..., Zhouqi Hua, ..., Yu Qiao et al.
Technical Report
Intern-S1 is a large multimodal MoE foundation model trained with massive scientific data and mixture-of-rewards reinforcement learning, achieving SOTA performance in scientific reasoning and professional tasks while remaining competitive in general reasoning among open-source models.
Lei Bai, Zhongrui Cai, ..., Zhouqi Hua, ..., Yu Qiao et al.
Technical Report
Intern-S1 is a large multimodal MoE foundation model trained with massive scientific data and mixture-of-rewards reinforcement learning, achieving SOTA performance in scientific reasoning and professional tasks while remaining competitive in general reasoning among open-source models.

Zhouqi Hua, Wenwei Zhang, Chengqi Lyu, Yuzhe Gu, Songyang Gao, Kuikun Liu, Dahua Lin, Kai Chen
International Conference on Learning Representations ICLR 2026
Turing Machine Imitation Learning (TAIL) is a synthetic chain-of-thought framework that instills Turing machine–like execution in LLMs, enabling robust length generalization for computable reasoning. On 18 challenging tasks, a 7B TAIL model outperforms the 671B DeepSeek-R1, establishing a new state of the art.
Zhouqi Hua, Wenwei Zhang, Chengqi Lyu, Yuzhe Gu, Songyang Gao, Kuikun Liu, Dahua Lin, Kai Chen
International Conference on Learning Representations ICLR 2026
Turing Machine Imitation Learning (TAIL) is a synthetic chain-of-thought framework that instills Turing machine–like execution in LLMs, enabling robust length generalization for computable reasoning. On 18 challenging tasks, a 7B TAIL model outperforms the 671B DeepSeek-R1, establishing a new state of the art.