netzkontrast
's Collections
Performance
updated
ShortGPT: Layers in Large Language Models are More Redundant Than You
Expect
Paper
•
2403.03853
•
Published
•
66
SLEB: Streamlining LLMs through Redundancy Verification and Elimination
of Transformer Blocks
Paper
•
2402.09025
•
Published
•
9
Shortened LLaMA: A Simple Depth Pruning for Large Language Models
Paper
•
2402.02834
•
Published
•
17
Algorithmic progress in language models
Paper
•
2403.05812
•
Published
•
19
Larimar: Large Language Models with Episodic Memory Control
Paper
•
2403.11901
•
Published
•
33
Qihoo-T2X: An Efficiency-Focused Diffusion Transformer via Proxy Tokens
for Text-to-Any-Task
Paper
•
2409.04005
•
Published
•
19
Self-Discover: Large Language Models Self-Compose Reasoning Structures
Paper
•
2402.03620
•
Published
•
117
Chain-of-Thought Reasoning Without Prompting
Paper
•
2402.10200
•
Published
•
109
CAS-ViT: Convolutional Additive Self-attention Vision Transformers for
Efficient Mobile Applications
Paper
•
2408.03703
•
Published
LLaVA-Mini: Efficient Image and Video Large Multimodal Models with One
Vision Token
Paper
•
2501.03895
•
Published
•
52