-
Mixture-of-Depths: Dynamically allocating compute in transformer-based language models
Paper • 2404.02258 • Published • 104 -
Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs
Paper • 2403.20041 • Published • 35 -
ViTAR: Vision Transformer with Any Resolution
Paper • 2403.18361 • Published • 54
Prathamesh Sarang
pratos
·
AI & ML interests
None yet
Recent Activity
liked
a dataset
about 2 months ago
vikhyatk/lnqa
liked
a Space
2 months ago
MagicBag/FireFlow
liked
a Space
5 months ago
maxin-cn/Cinemo
Organizations
Collections
2
models
None public yet
datasets
None public yet