In-context autoencoder for context compression in a large language model

T Ge, J Hu, X Wang, SQ Chen, F Wei�- arXiv preprint arXiv:2307.06945, 2023 - arxiv.org
We propose the In-context Autoencoder (ICAE) for context compression in a large language
model (LLM). The ICAE has two modules: a learnable encoder adapted with LoRA from an�…

Soaring from 4K to 400K: Extending LLM's Context with Activation Beacon

P Zhang, Z Liu, S Xiao, N Shao, Q Ye, Z Dou�- arXiv preprint arXiv�…, 2024 - arxiv.org
The utilization of long contexts poses a big challenge for large language models due to their
limited context window length. Although the context window can be extended through fine�…

Adapting language models to compress contexts

A Chevalier, A Wettig, A Ajith, D Chen�- arXiv preprint arXiv:2305.14788, 2023 - arxiv.org
Transformer-based language models (LMs) are powerful and widely-applicable tools, but
their usefulness is constrained by a finite context window and the expensive computational�…

Long-context language modeling with parallel context encoding

H Yen, T Gao, D Chen�- arXiv preprint arXiv:2402.16617, 2024 - arxiv.org
Extending large language models (LLMs) to process longer inputs is crucial for numerous
applications. However, the considerable computational cost of transformers, coupled with�…

Compressing context to enhance inference efficiency of large language models

Y Li, B Dong, C Lin, F Guerin�- arXiv preprint arXiv:2310.06201, 2023 - arxiv.org
Large language models (LLMs) achieved remarkable performance across various tasks.
However, they face challenges in managing long documents and extended conversations�…

Llmlingua: Compressing prompts for accelerated inference of large language models

H Jiang, Q Wu, CY Lin, Y Yang, L Qiu�- arXiv preprint arXiv:2310.05736, 2023 - arxiv.org
Large language models (LLMs) have been applied in various applications due to their
astonishing capabilities. With advancements in technologies such as chain-of-thought (CoT)�…

Fine-tune language models to approximate unbiased in-context learning

T Chu, Z Song, C Yang�- arXiv preprint arXiv:2310.03331, 2023 - arxiv.org
In-context learning (ICL) is an astonishing emergent ability of large language models
(LLMs). By presenting a prompt that includes multiple input-output pairs as examples and�…

Label words are anchors: An information flow perspective for understanding in-context learning

L Wang, L Li, D Dai, D Chen, H Zhou, F Meng…�- arXiv preprint arXiv�…, 2023 - arxiv.org
In-context learning (ICL) emerges as a promising capability of large language models
(LLMs) by providing them with demonstration examples to perform diverse tasks. However�…

Parallel context windows for large language models

N Ratner, Y Levine, Y Belinkov, O Ram, I Magar…�- arXiv preprint arXiv�…, 2022 - arxiv.org
When applied to processing long text, Large Language Models (LLMs) are limited by their
context window. Existing efforts to address this limitation involve training specialized�…

Longrope: Extending llm context window beyond 2 million tokens

Y Ding, LL Zhang, C Zhang, Y Xu, N Shang…�- arXiv preprint arXiv�…, 2024 - arxiv.org
Large context window is a desirable feature in large language models (LLMs). However,
due to high fine-tuning costs, scarcity of long texts, and catastrophic values introduced by�…