-
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
Paper • 2403.03853 • Published • 61 -
SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Paper • 2402.09025 • Published • 6 -
Shortened LLaMA: A Simple Depth Pruning for Large Language Models
Paper • 2402.02834 • Published • 14 -
Algorithmic progress in language models
Paper • 2403.05812 • Published • 18
michael
netzkontrast
AI & ML interests
None yet
Organizations
None yet
Collections
4
spaces
1
models
None public yet
datasets
None public yet