lilujunai / prompt-in-context-learning

Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.

Home Page:https://github.com/EgoAlpha/prompt-in-context-learning

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Typing SVG

An Open-Source Engineering Guide for Prompt-in-context-learning from EgoAlpha Lab.

English | 简体中文

📝 Papers | ⚡️ Playground | 🛠 Prompt Engineering | 🌍 ChatGPT Prompt

version Awesome

⭐️ Shining ⭐️: This is fresh, daily-updated resources for in-context learning and prompt engineering. As Artificial General Intelligence (AGI) is approaching, let’s take action and become a super learner so as to position ourselves at the forefront of this exciting era and strive for personal and professional greatness.

The resources include:

🎉Papers🎉: The latest papers about in-context learning or prompt engineering.

🎉Playground🎉: Large language models that enable prompt experimentation.

🎉Prompt Engineering🎉: Prompt techniques for leveraging large language models.

🎉ChatGPT Prompt🎉: Prompt examples that can be applied in our work and daily lives.

In the future, there will likely be two types of people on Earth (perhaps even on Mars, but that's a question for Musk):

  • Those who enhance their abilities through the use of AI;
  • Those whose jobs are replaced by AI automation.

💎EgoAlpha: Hello! human👤, are you ready?

📢 News

📜 Papers

You can directly click on the title to jump to the corresponding PDF link location


Survey

👉Complete paper list 🔗 for "Survey"👈

Prompt Engineering

📌 Prompt Design

A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT2023.02.21

GraphPrompt: Unifying Pre-Training and Downstream Tasks for Graph Neural Networks2023.02.16

Commonsense-Aware Prompting for Controllable Empathetic Dialogue Generation2023.02.02

Progressive Prompts: Continual Learning for Language Models2023.01.29

Batch Prompting: Efficient Inference with Large Language Model APIs2023.01.19

Promptagator: Few-shot Dense Retrieval From 8 Examples2022.09.23

Interactive and Visual Prompt Engineering for Ad-hoc Task Adaptation with Large Language Models2022.08.16

DocPrompting: Generating Code by Retrieving the Docs2022.07.13

Design Guidelines for Prompt Engineering Text-to-Image Generative Models2021.09.14

Program Synthesis with Large Language Models2021.08.16

👉Complete paper list 🔗 for "Prompt Design"👈

📌 Automatic Prompt

👉Complete paper list 🔗 for "Automatic Prompt"👈

📌 Chain of Thought

👉Complete paper list 🔗 for "Chain of Thought"👈

📌 Knowledge Augmented Prompts

A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT2023.02.21

GraphPrompt: Unifying Pre-Training and Downstream Tasks for Graph Neural Networks2023.02.16

Commonsense-Aware Prompting for Controllable Empathetic Dialogue Generation2023.02.02

Progressive Prompts: Continual Learning for Language Models2023.01.29

Batch Prompting: Efficient Inference with Large Language Model APIs2023.01.19

Promptagator: Few-shot Dense Retrieval From 8 Examples2022.09.23

Interactive and Visual Prompt Engineering for Ad-hoc Task Adaptation with Large Language Models2022.08.16

DocPrompting: Generating Code by Retrieving the Docs2022.07.13

Design Guidelines for Prompt Engineering Text-to-Image Generative Models2021.09.14

Program Synthesis with Large Language Models2021.08.16

👉Complete paper list 🔗 for "Knowledge Augmented Prompts"👈

📌 Evaluation & Reliability

Language Model Crossover: Variation through Few-Shot Prompting2023.02.23

Evaluating the Robustness of Discrete Prompts2023.02.11

PLACES: Prompting Language Models for Social Conversation Synthesis2023.02.07

Controlling for Stereotypes in Multimodal Language Model Evaluation2023.02.03

Large Language Models Can Be Easily Distracted by Irrelevant Context2023.01.31

Emergent Analogical Reasoning in Large Language Models2022.12.19

Discovering Language Model Behaviors with Model-Written Evaluations2022.12.19

On Second Thought, Let's Not Think Step by Step! Bias and Toxicity in Zero-Shot Reasoning2022.12.15

Solving math word problems with process- and outcome-based feedback2022.11.25

Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks2022.11.22

👉Complete paper list 🔗 for "Evaluation & Reliability"👈

In-context Learning

Larger language models do in-context learning differently2023.03.07

How Robust is GPT-3.5 to Predecessors? A Comprehensive Study on Language Understanding Tasks2023.03.01

Language Model Crossover: Variation through Few-Shot Prompting2023.02.23

How Does In-Context Learning Help Prompt Tuning?2023.02.22

Bounding the Capabilities of Large Language Models in Open Text Generation with Prompt Constraints2023.02.17

Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Learning2023.01.27

Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Learning2023.01.27

One Embedder, Any Task: Instruction-Finetuned Text Embeddings2022.12.19

Complementary Explanations for Effective In-Context Learning2022.11.25

Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them2022.10.17

👉Complete paper list 🔗 for "In-context Learning"👈

Multimodal Prompt

👉Complete paper list 🔗 for "Multimodal Prompt"👈

Prompt Application

SpeechPrompt v2: Prompt Tuning for Speech Classification Tasks2023.03.01

Soft Prompt Guided Joint Learning for Cross-Domain Sentiment Analysis2023.03.01

EvoPrompting: Language Models for Code-Level Neural Architecture Search2023.02.28

More than you've asked for: A Comprehensive Analysis of Novel Prompt Injection Threats to Application-Integrated Large Language Models2023.02.23

Grimm in Wonderland: Prompt Engineering with Midjourney to Illustrate Fairytales2023.02.17

LabelPrompt: Effective Prompt-based Learning for Relation Classification2023.02.16

Prompt Tuning of Deep Neural Networks for Speaker-adaptive Visual Speech Recognition2023.02.16

Prompting for Multimodal Hateful Meme Classification2023.02.08

QaNER: Prompting Question Answering Models for Few-shot Named Entity Recognition2022.03.03

LightNER: A Lightweight Tuning Paradigm for Low-resource NER via Pluggable Prompting2021.08.31

👉Complete paper list 🔗 for "Prompt Application"👈

Foundation Models

Meet in the Middle: A New Pre-training Paradigm2023.03.13

High-throughput Generative Inference of Large Language Models with a Single GPU2023.03.13

Stabilizing Transformer Training by Preventing Attention Entropy Collapse2023.03.11

An Overview on Language Models: Recent Developments and Outlook2023.03.10

Foundation Models for Decision Making: Problems, Methods, and Opportunities2023.03.07

How Do Transformers Learn Topic Structure: Towards a Mechanistic Understanding2023.03.07

LLaMA: Open and Efficient Foundation Language Models2023.02.27

Self-Instruct: Aligning Language Model with Self Generated Instructions2022.12.20

PaLM: Scaling Language Modeling with Pathways2022.04.05

LoRA: Low-Rank Adaptation of Large Language Models2021.06.17

👉Complete paper list 🔗 for "Foundation Models"👈

✉️ Contact

This repo is maintained by EgoAlpha Lab. Questions and discussions are welcome via helloegoalpha@gmail.com.

We are willing to engage in discussions with friends from the academic and industrial communities, and explore the latest developments in prompt engineering and in-context learning together.

🙏 Acknowledgements

Thanks to the PhD students from EgoAlpha Lab and other workers who participated in this repo. We will improve the project in the follow-up period and maintain this community well. We also would like to express our sincere gratitude to the authors of the relevant resources. Your efforts have broadened our horizons and enabled us to perceive a more wonderful world.

About

Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.

https://github.com/EgoAlpha/prompt-in-context-learning

License:MIT License


Languages

Language:CSS 100.0%