Unleashing the Power of Pre-trained Language Models for Offline Reinforcement Learning Paper • 2310.20587 • Published Oct 31, 2023 • 16
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement Paper • 2403.15042 • Published Mar 22 • 25
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization Paper • 1912.08777 • Published Dec 18, 2019 • 2
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks Paper • 1907.12461 • Published Jul 29, 2019 • 1
Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models Paper • 2411.12580 • Published Nov 19 • 2
Studying Large Language Model Generalization with Influence Functions Paper • 2308.03296 • Published Aug 7, 2023 • 12
Multimodal Autoregressive Pre-training of Large Vision Encoders Paper • 2411.14402 • Published Nov 21 • 43
CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation Paper • 2103.06874 • Published Mar 11, 2021 • 1