A brief overview of ChatGPT: The history, status quo and potential future development
ChatGPT, an artificial intelligence generated content (AIGC) model developed by OpenAI,
has attracted world-wide attention for its capability of dealing with challenging language�…
has attracted world-wide attention for its capability of dealing with challenging language�…
Recent advances in natural language processing via large pre-trained language models: A survey
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks�…
changed the Natural Language Processing (NLP) field. For numerous NLP tasks�…
Scaling instruction-finetuned language models
Finetuning language models on a collection of datasets phrased as instructions has been
shown to improve model performance and generalization to unseen tasks. In this paper we�…
shown to improve model performance and generalization to unseen tasks. In this paper we�…
Visual instruction tuning
Instruction tuning large language models (LLMs) using machine-generated instruction-
following data has been shown to improve zero-shot capabilities on new tasks, but the idea�…
following data has been shown to improve zero-shot capabilities on new tasks, but the idea�…
Judging llm-as-a-judge with mt-bench and chatbot arena
Evaluating large language model (LLM) based chat assistants is challenging due to their
broad capabilities and the inadequacy of existing benchmarks in measuring human�…
broad capabilities and the inadequacy of existing benchmarks in measuring human�…
Llama 2: Open foundation and fine-tuned chat models
In this work, we develop and release Llama 2, a collection of pretrained and fine-tuned large
language models (LLMs) ranging in scale from 7 billion to 70 billion parameters. Our fine�…
language models (LLMs) ranging in scale from 7 billion to 70 billion parameters. Our fine�…
Qlora: Efficient finetuning of quantized llms
We present QLoRA, an efficient finetuning approach that reduces memory usage enough to
finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit�…
finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit�…
A survey of large language models
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for�…
grammatical rules. It poses a significant challenge to develop capable AI algorithms for�…
Instructpix2pix: Learning to follow image editing instructions
We propose a method for editing images from human instructions: given an input image and
a written instruction that tells the model what to do, our model follows these instructions to�…
a written instruction that tells the model what to do, our model follows these instructions to�…
[HTML][HTML] Large language models encode clinical knowledge
Large language models (LLMs) have demonstrated impressive capabilities, but the bar for
clinical applications is high. Attempts to assess the clinical knowledge of models typically�…
clinical applications is high. Attempts to assess the clinical knowledge of models typically�…