Growing Through Experience: Scaling Episodic Grounding in Language Models Paper • 2506.01312 • Published Jun 2, 2025
AmoebaLLM: Constructing Any-Shape Large Language Models for Efficient and Instant Deployment Paper • 2411.10606 • Published Nov 15, 2024 • 1
LongMamba: Enhancing Mamba's Long Context Capabilities via Training-Free Receptive Field Enlargement Paper • 2504.16053 • Published Apr 22, 2025
SwiReasoning: Switch-Thinking in Latent and Explicit for Pareto-Superior Reasoning LLMs Paper • 2510.05069 • Published Oct 6, 2025 • 13
LaCache: Ladder-Shaped KV Caching for Efficient Long-Context Modeling of Large Language Models Paper • 2507.14204 • Published Jul 14, 2025
Superficial Self-Improved Reasoners Benefit from Model Merging Paper • 2503.02103 • Published Mar 3, 2025
Mitigating Forgetting Between Supervised and Reinforcement Learning Yields Stronger Reasoners Paper • 2510.04454 • Published Oct 6, 2025
Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less Reasonable Paper • 2503.00555 • Published Mar 1, 2025
Crowdsource, Crawl, or Generate? Creating SEA-VL, a Multicultural Vision-Language Dataset for Southeast Asia Paper • 2503.07920 • Published Mar 10, 2025 • 101
Virus: Harmful Fine-tuning Attack for Large Language Models Bypassing Guardrail Moderation Paper • 2501.17433 • Published Jan 29, 2025 • 10
Gaze-LLE: Gaze Target Estimation via Large-Scale Learned Encoders Paper • 2412.09586 • Published Dec 12, 2024 • 5
PokéLLMon: A Human-Parity Agent for Pokémon Battles with Large Language Models Paper • 2402.01118 • Published Feb 2, 2024 • 32
Harmful Fine-tuning Attacks and Defenses for Large Language Models: A Survey Paper • 2409.18169 • Published Sep 26, 2024 • 1
Vaccine: Perturbation-aware Alignment for Large Language Models against Harmful Fine-tuning Attack Paper • 2402.01109 • Published Feb 2, 2024
Lisa: Lazy Safety Alignment for Large Language Models against Harmful Fine-tuning Attack Paper • 2405.18641 • Published May 28, 2024
Antidote: Post-fine-tuning Safety Alignment for Large Language Models against Harmful Fine-tuning Paper • 2408.09600 • Published Aug 18, 2024
Booster: Tackling Harmful Fine-tuning for Large Language Models via Attenuating Harmful Perturbation Paper • 2409.01586 • Published Sep 3, 2024