-
Enhancing Automated Interpretability with Output-Centric Feature Descriptions
Paper • 2501.08319 • Published • 10 -
Open Problems in Machine Unlearning for AI Safety
Paper • 2501.04952 • Published • 1 -
Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models
Paper • 2412.16247 • Published • 1 -
Inferring Functionality of Attention Heads from their Parameters
Paper • 2412.11965 • Published • 2
Collections
Discover the best community collections!
Collections including paper arxiv:2405.12250
-
Selective Attention Improves Transformer
Paper • 2410.02703 • Published • 24 -
Differential Transformer
Paper • 2410.05258 • Published • 169 -
TidalDecode: Fast and Accurate LLM Decoding with Position Persistent Sparse Attention
Paper • 2410.05076 • Published • 8 -
SeerAttention: Learning Intrinsic Sparse Attention in Your LLMs
Paper • 2410.13276 • Published • 26
-
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Paper • 2312.00752 • Published • 140 -
Elucidating the Design Space of Diffusion-Based Generative Models
Paper • 2206.00364 • Published • 15 -
GLU Variants Improve Transformer
Paper • 2002.05202 • Published • 2 -
StarCoder 2 and The Stack v2: The Next Generation
Paper • 2402.19173 • Published • 137
-
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
Paper • 2405.21060 • Published • 64 -
Your Transformer is Secretly Linear
Paper • 2405.12250 • Published • 151 -
KAN: Kolmogorov-Arnold Networks
Paper • 2404.19756 • Published • 109 -
Trusted Machine Learning Models Unlock Private Inference for Problems Currently Infeasible with Cryptography
Paper • 2501.08970 • Published • 6
-
Autoregressive Model Beats Diffusion: Llama for Scalable Image Generation
Paper • 2406.06525 • Published • 67 -
Husky: A Unified, Open-Source Language Agent for Multi-Step Reasoning
Paper • 2406.06469 • Published • 25 -
Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models
Paper • 2406.04271 • Published • 29 -
Block Transformer: Global-to-Local Language Modeling for Fast Inference
Paper • 2406.02657 • Published • 38
-
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
Paper • 2403.03853 • Published • 62 -
SliceGPT: Compress Large Language Models by Deleting Rows and Columns
Paper • 2401.15024 • Published • 72 -
Your Transformer is Secretly Linear
Paper • 2405.12250 • Published • 151 -
Yi: Open Foundation Models by 01.AI
Paper • 2403.04652 • Published • 62