Skip to content

Latest commit

 

History

History

techniques

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 

Resources on ChatGPT and Large Language Models

Collection of papers and related works for Large Language Models (ChatGPT, GPT-3, Codex etc.).

Contributors

This repository is contributed by the following contributors.

The automation script of this repo is powered by Auto-Bibfile. If you'd like to commit to this repo, please modify bibtex.bib or related_works.json and re-generate README.md using python scripts/run.py.

This page categorizes the literature by the Techniques

Papers

Outline

Hyperlinks

Evaluation

Survey

In-Context Learning

Instruction Tuning

RLHF

Pre-Training Techniques

Mixtures of Experts

Knowledge Enhanced

Knowledge Distillation

Knowledge Generation

Knowledge Editing

Reasoning

Chain of Thought

Multi-Step Reasoning

Arithmetic Reasoning

Symbolic Reasoning

Chain of Verification

Knowledge Graph Embedding

Federated Learning

Distributed AI

Selective Annotation

  • img Selective Annotation Makes Language Models Better Few-Shot Learners, img img img img img img img
    by Hongjin Su, Jungo Kasai, Chen Henry Wu, Weijia Shi, Tianlu Wang, Jiayi Xin, Rui Zhang, Mari Ostendorf et al.
    This paper proposes a graph-based selective annotation method named vote-k to
    (1) select a pool of examples to annotate from unlabeled data,
    (2) retrieve prompts (contexts) from the annotated data pool for in-context learning.
    Specifically, the selection method first selects a small set of unlabeled examples iteratively and then labels them to serve as contexts for LLMs to predict the labels of the rest unlabeled data. The method selects the predictions with highest confidence (log probability of generation output) to fill up the selective annotation pool.

  • img Selective Data Acquisition in the Wild for Model Charging,
    by Chengliang Chai, Jiabin Liu, Nan Tang, Guoliang Li and Yuyu Luo

Program and Code Generation

Code Representation

Code Fixing

Code Review

Program Generation

Software Engineering

AIGC

Controllable Text Generation

Continual Learning

Prompt Engineering

Natural Language Understanding

Multimodal

Multilingual

Reliability

Robustness

Dialogue System

Recommender System

Event Extraction

Event Relation Extraction

Data Argumentation

Data Annotation

Information Extraction

Domain Adaptive

Question Answering

Application

Meta Learning

  • img Meta-learning via Language Model In-context Tuning, img img img
    by Yanda Chen, Ruiqi Zhong, Sheng Zha, George Karypis and He He

  • img MetaICL: Learning to Learn In Context, img img
    by Sewon Min, Mike Lewis, Luke Zettlemoyer and Hannaneh Hajishirzi
    MetaICL proposes a supervised meta-training framework to enable LMs to more effectively learn a new task in context. In MetaICL, each meta-training example includes several training examples from one task that will be presented together as a single sequence to the LM, and the prediction of the final example is used to calculate the loss.

Generalizability

Language Model as Knowledge Base

Retrieval-Augmented Language Model

Quality

Interpretability/Explainability

Data Generation

Safety

Graph Learning

Knowledge Storage and Locating

Knowledge Fusion

Agent

LLM and GNN

Vision LLM

LLM and KG

Others