Yingfei(Jeremy) Xiang (SuperXiang)

SuperXiang

Geek Repo

Company:Baidu, Inc.

Location:Shenzhen, Guangdong, China

Home Page:https://scholar.google.com/citations?user=7n2td58AAAAJ

Twitter:@YingfeiX

Github PK Tool:Github PK Tool

Yingfei(Jeremy) Xiang's repositories

License:Apache-2.0Stargazers:1Issues:0Issues:0

Cherry_LLM

[NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other models

Language:PythonStargazers:1Issues:0Issues:0

CS-Eval

CS-Eval is a comprehensive evaluation suite for fundamental cybersecurity models or large language models' cybersecurity ability.

License:MITStargazers:1Issues:0Issues:0

data-juicer

A one-stop data processing system to make data higher-quality, juicier, and more digestible for (multimodal) LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!

License:Apache-2.0Stargazers:1Issues:0Issues:0

dclm

DataComp for Language Models

Language:HTMLLicense:MITStargazers:1Issues:0Issues:0

deepeval

The LLM Evaluation Framework

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

devika

Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

distillm

Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)

Stargazers:1Issues:0Issues:0

dsir

DSIR large-scale data selection framework for language model training

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

EasyContext

Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.

License:Apache-2.0Stargazers:1Issues:0Issues:0

GPU-Benchmarks-on-LLM-Inference

Multiple NVIDIA GPUs or Apple Silicon for Large Language Model Inference?

Language:Jupyter NotebookStargazers:1Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

MambaInLlama

Official Repository of The Mamba in the Llama: Distilling and Accelerating Hybrid Models

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

Minitron

A family of compressed models obtained via pruning and knowledge distillation

Stargazers:1Issues:0Issues:0

ollama

Get up and running with Llama 2, Mistral, and other large language models.

Language:GoLicense:MITStargazers:1Issues:0Issues:0

OpenDevin

🐚 OpenDevin: Code Less, Make More

License:MITStargazers:1Issues:0Issues:0

opro

official code for "Large Language Models as Optimizers"

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

orpo

Official repository for ORPO

License:Apache-2.0Stargazers:1Issues:0Issues:0

PentestGPT

A GPT-empowered penetration testing tool

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

persona-hub

Official repo for the paper "Scaling Synthetic Data Creation with 1,000,000,000 Personas"

Language:PythonStargazers:1Issues:0Issues:0
Stargazers:1Issues:0Issues:0

Qwen1.5

Qwen1.5 is the improved version of Qwen, the large language model series developed by Qwen team, Alibaba Cloud.

Language:ShellStargazers:1Issues:0Issues:0
Stargazers:1Issues:0Issues:0

simple-one-api

OpenAI 接口接入适配,支持千帆大模型平台、讯飞星火大模型、腾讯混元以及MiniMax、Deep-Seek,等兼容OpenAI接口,仅单可执行文件,配置超级简单,一键部署,开箱即用.

Stargazers:1Issues:0Issues:0

small-LMs-Task-Planning

Can only LLMs do Reasoning?: Potential of Small Language Models in Task Planning

Language:Jupyter NotebookStargazers:1Issues:0Issues:0
Language:PythonLicense:MITStargazers:1Issues:0Issues:0

InternLM

Official release of InternLM2.5 7B base and chat models. 1M context support

License:Apache-2.0Stargazers:0Issues:0Issues:0