There are 3 repositories under reasoning-models topic.
Open Source Deep Research Alternative to Reason and Search on Private Data. Written in Python.
MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model.
"LightReasoner: Can Small Language Models Teach Large Language Models Reasoning?"
Official repository for EXAONE Deep built by LG AI Research
Official implementation of the NeurIPS 2025 paper "Soft Thinking: Unlocking the Reasoning Potential of LLMs in Continuous Concept Space"
MedReason: Eliciting Factual Medical Reasoning Steps in LLMs via Knowledge Graphs
[NeurIPS 2025] Simple extension on vLLM to help you speed up reasoning model without training.
Official Repository of OmniCaptioner
xVerify: Efficient Answer Verifier for Reasoning Model Evaluations
An interactive thinking and deep reasoning model. It provides a cognitive reasoning paradigm for complex multi-hop problems.
[arXiv 2025] Can MLLMs Guide Me Home? A Benchmark Study on Fine-Grained Visual Reasoning from Transit Maps
[NeurIPS 2025] VeriThinker: Learning to Verify Makes Reasoning Model Efficient
Lightweight replication study of DeepSeek-R1-Zero. Interesting findings include "No Aha Moment", "Longer CoT ≠ Accuracy", and "Language Mixing in Instruct Models".
Pure RL to post-train base models for social reasoning capabilities. Lightweight replication of DeepSeek-R1-Zero with Social IQa dataset.
Awesome-Parallel-Reasoning: Unlocking the reasoning potential of LLMs. Papers, Code, Resources & Survey.
Code for the 2025 ACL publication "Fine-Tuning on Diverse Reasoning Chains Drives Within-Inference CoT Refinement in LLMs"
Implementation and subsequent optimization for "Reviving DSP for Advanced Theorem Proving in the Era of Reasoning Models"
R1-Code-Interpreter: Training LLMs to Reason with Code via Supervised and Reinforcement Learning
AI Lawyer is an intelligent reasoning legal assistant powered by DeepSeek , Ollama RAG and LangChain, designed to streamline legal research and document analysis. By leveraging retrieval-augmented generation (RAG), it provides precise legal insights, and contract summarization. With an intuitive Streamlit-based UI, analyze legal documents.
Using a reasoning LLM to learn a prompt from data
ReasoningShield: Safety Detection over Reasoning Traces of Large Reasoning Models
This repository hosts the instructions and workshop materials for Lab 333 - Evaluate Reasoning Models for Your Generative AI Solutions
Agentic Deep Graph Reasoning Implementation
Explore the evolution of AGI through historical context, reasoning models, and agent systems, while gaining hands-on experience with cutting-edge models like Claude 4, DeepSeek-R1, and OpenAI's o3. Learn to critically evaluate AGI benchmarks, understand their limitations, and identify where current models excel or struggle in reasoning tasks.
Turn stories, strategies, or systems into insight. Auto-generate Dialectical Wheels (DWs) from any text to reveal blind spots, surface polarities, and trace dynamic paths toward synthesis. DWs are semantic maps that expose tension, transformation, and coherence within a system—whether narrative, ethical, organizational, or technological.
State Sandbox is an experimental game for socioeconomic simulation. It uses Large Language Models (o3-mini) to simulate the world and complex policy impacts.
LLM finetuning for Sudoku solving
Predicting drug approval with reasoning.
Sudoku4LLM is a Sudoku dataset generator for training and evaluating reasoning in Large Language Models (LLMs). It offers customizable puzzles, difficulty levels, and 11 serialization formats to support structured data reasoning and Chain of Thought (CoT) experiments.
A simple AI source code that includes chat, reasoning and image features using public APIs like xAI, OpenAI, HuggingFace and Flux.
This repo contains evaluation code for the paper "MANBench: Is Your Multimodal Model Smarter than Human?" [ACL 2025 Findings]
Official code for our paper: "SketchThinker-R1: Towards Efficient Sketch-Style Reasoning in Large Multimodal Models".