Yue Deng's starred repositories
alpaca-lora
Instruct-tune LLaMA on consumer hardware
lm-evaluation-harness
A framework for few-shot evaluation of language models.
awesome-pretrained-chinese-nlp-models
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
chatgpt-prompts-for-academic-writing
This list of writing prompts covers a range of topics and tasks, including brainstorming research ideas, improving language and style, conducting literature reviews, and developing research plans.
LLMDataHub
A quick guide (especially) for trending instruction finetuning datasets
awesome-llm-security
A curation of awesome tools, documents and projects about LLM Security.
WizardVicunaLM
LLM that combines the principles of wizardLM and vicunaLM
Awesome-LLM-Safety
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights into the safety implications, challenges, and advancements surrounding these powerful models.
ABSA-Reading-List
Reading list of aspect-based sentiment analysis.
label-words-are-anchors
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
do-not-answer
Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs
activation_additions
Algebraic value editing in pretrained language models
multilingual-safety-for-LLMs
[ICLR 2024]Data for "Multilingual Jailbreak Challenges in Large Language Models"
contrastive-cot
Contrastive Chain-of-Thought Prompting
ShadowAlignment
Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models
ToxificationReversal
Code for the paper "Self-Detoxifying Language Models via Toxification Reversal" (EMNLP 2023)
safetyllama
Finetune LLaMA-2-7b-chat to perform safety evaluation of user-bot conversation