DUT-LiuYang's starred repositories
Prompt-Engineering-Guide
🐙 Guides, papers, lecture, notebooks and resources for prompt engineering
ChineseBQB
🇨🇳 Chinese sticker pack,More joy / 表情包的博物馆, Github最有毒的仓库, **表情包大集合, 聚欢乐~
ML-Papers-of-the-Week
🔥Highlighting the top ML papers every week.
accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
awesome-pretrained-chinese-nlp-models
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
GPT2-NewsTitle
Chinese NewsTitle Generation Project by GPT2.带有超级详细注释的中文GPT2新闻标题生成项目。
DeepLearningBookQA_cn
深度学习面试问题 回答对应的DeepLearning中文版页码
BERT-whitening-pytorch
Pytorch version of BERT-whitening
leetcode-java
🎓🎓🎓 Leetcode solution in Java - 536/921 Solved. https://leetcode.com/problemset/all/
BANG
BANG is a new pretraining model to Bridge the gap between Autoregressive (AR) and Non-autoregressive (NAR) Generation. AR and NAR generation can be uniformly regarded as to what extent previous tokens can be attended, and BANG bridges AR and NAR generation by designing a novel model structure for large-scale pretraining. The pretrained BANG model can simultaneously support AR, NAR and semi-NAR generation to meet different requirements.
Meta-Curriculum
Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation (AAAI 2021)
Prune-Tune
Official code repository for AAAI2021 paper Finding Sparse Structures for Domain Specific Neural Machine Translation