There are 0 repository under qwq topic.
[NeurIPS 2025] 🌐 WebThinker: Empowering Large Reasoning Models with Deep Research Capability
Higher performance OpenAI LLM service than vLLM serve: A pure C++ high-performance OpenAI LLM service implemented with GPRS+TensorRT-LLM+Tokenizers.cpp, supporting chat and function call, AI agents, distributed multi-GPU inference, multimodal capabilities, and a Gradio chat interface.
Official PyTorch implementation for Hogwild! Inference: Parallel LLM Generation with a Concurrent Attention Cache
To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning Models
Models based on DeepSeek, Qwen3, ChatGPT, and Ollama call the Golang SDK.
Breaking long thought processes of o1-like LLMs, such as DeepSeek-R1, QwQ
qwq-server,一个轻量级博客系统,基于Go语言和Gin轻量Web框架开发