-
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper • 2312.11514 • Published • 259 -
AppAgent: Multimodal Agents as Smartphone Users
Paper • 2312.13771 • Published • 53 -
ControlRoom3D: Room Generation using Semantic Proxy Rooms
Paper • 2312.05208 • Published • 8 -
MobileVLM : A Fast, Reproducible and Strong Vision Language Assistant for Mobile Devices
Paper • 2312.16886 • Published • 20
Collections
Discover the best community collections!
Collections including paper arxiv:2312.13771
-
Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision
Paper • 2312.09390 • Published • 33 -
OneLLM: One Framework to Align All Modalities with Language
Paper • 2312.03700 • Published • 21 -
Generative Multimodal Models are In-Context Learners
Paper • 2312.13286 • Published • 35 -
The LLM Surgeon
Paper • 2312.17244 • Published • 9
-
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator
Paper • 2312.04474 • Published • 31 -
Training Chain-of-Thought via Latent-Variable Inference
Paper • 2312.02179 • Published • 9 -
The Unlocking Spell on Base LLMs: Rethinking Alignment via In-Context Learning
Paper • 2312.01552 • Published • 31 -
AppAgent: Multimodal Agents as Smartphone Users
Paper • 2312.13771 • Published • 53
-
GAIA: a benchmark for General AI Assistants
Paper • 2311.12983 • Published • 192 -
ToolTalk: Evaluating Tool-Usage in a Conversational Setting
Paper • 2311.10775 • Published • 8 -
TPTU-v2: Boosting Task Planning and Tool Usage of Large Language Model-based Agents in Real-world Systems
Paper • 2311.11315 • Published • 7 -
An Embodied Generalist Agent in 3D World
Paper • 2311.12871 • Published • 8
-
Technical Report: Large Language Models can Strategically Deceive their Users when Put Under Pressure
Paper • 2311.07590 • Published • 17 -
Exponentially Faster Language Modelling
Paper • 2311.10770 • Published • 118 -
AppAgent: Multimodal Agents as Smartphone Users
Paper • 2312.13771 • Published • 53 -
DocLLM: A layout-aware generative language model for multimodal document understanding
Paper • 2401.00908 • Published • 181