-
EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters
Paper • 2402.04252 • Published • 26 -
Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models
Paper • 2402.03749 • Published • 13 -
ScreenAI: A Vision-Language Model for UI and Infographics Understanding
Paper • 2402.04615 • Published • 43 -
EfficientViT-SAM: Accelerated Segment Anything Model Without Performance Loss
Paper • 2402.05008 • Published • 22
Collections
Discover the best community collections!
Collections including paper arxiv:2409.11340
-
IFAdapter: Instance Feature Control for Grounded Text-to-Image Generation
Paper • 2409.08240 • Published • 22 -
IterComp: Iterative Composition-Aware Feedback Learning from Model Gallery for Text-to-Image Generation
Paper • 2410.07171 • Published • 43 -
EvolveDirector: Approaching Advanced Text-to-Image Generation with Large Vision-Language Models
Paper • 2410.07133 • Published • 19 -
OmniGen: Unified Image Generation
Paper • 2409.11340 • Published • 113
-
Qwen2.5-Coder Technical Report
Paper • 2409.12186 • Published • 145 -
Attention Heads of Large Language Models: A Survey
Paper • 2409.03752 • Published • 89 -
Loopy: Taming Audio-Driven Portrait Avatar with Long-Term Motion Dependency
Paper • 2409.02634 • Published • 95 -
OmniGen: Unified Image Generation
Paper • 2409.11340 • Published • 113
-
OmniGen: Unified Image Generation
Paper • 2409.11340 • Published • 113 -
Video-Guided Foley Sound Generation with Multimodal Controls
Paper • 2411.17698 • Published • 9 -
FLOAT: Generative Motion Latent Flow Matching for Audio-driven Talking Portrait
Paper • 2412.01064 • Published • 27 -
OmniFlow: Any-to-Any Generation with Multi-Modal Rectified Flows
Paper • 2412.01169 • Published • 13