SemCoT: Accelerating Chain-of-Thought Reasoning through Semantically-Aligned Implicit Tokens
Paper
β’
2510.24940
β’
Published
β’
18
SemCoT is a framework that improves the efficiency of Chain-of-Thought (CoT) reasoning by encoding reasoning steps inside hidden representations ("implicit reasoning") instead of generating long textual explanations. This approach significantly speeds up inference while maintaining high reasoning performance by ensuring semantic alignment with ground-truth reasoning.
This specific checkpoint is based on princeton-nlp/Sheared-LLaMA-1.3B and fine-tuned on the ChilleD/SVAMP dataset using the SemCoT framework.
This model is built using PyTorchModelHubMixin. Because SemCoT uses a custom implicit reasoning framework, please refer to the official GitHub repository for instructions on how to load and run the model.
@inproceedings{he2025semcot,
title={SemCoT: Accelerating Chain-of-Thought Reasoning through Semantically-Aligned Implicit Tokens},
author={He, Yinhan and Zheng, Wendy and Zhu, Yaochen and Zheng, Zaiyi and Su, Lin and Vasudevan, Sriram and Guo, Qi and Hong, Liangjie and Li, Jundong},
booktitle={39th Conference on Neural Information Processing Systems (NeurIPS 2025)},
year={2025}
}
Base model
princeton-nlp/Sheared-LLaMA-1.3B