I've noticed something. While we're careful about what we post on social media, we're sharing our deepest and most intimate thoughts with AI chatbots -- health concerns, financial worries, relationship issues, business ideas...
With OpenAI hinting at ChatGPT advertising, this matters more than ever. Unlike banner ads, AI advertising happens within the conversation itself. Sponsors could subtly influence that relationship advice or financial guidance.
The good news? We have options. 🤝 Open source AI models let us keep conversations private, avoid surveillance-based business models, and build systems that actually serve users first.
✨ Efficiency leads the month - At scale: optimizing compute use in massive MoE models e.g. DeepSeek v3.1 - In small models: lightweight & deployable e.g. MiniCPM V 4.5, Step Audio 2-mini, Intern S1-mini,Ovis2.5-9B etc.
✨ Reasoning + Agentic wave 🌊 Not just demos, but real product use cases. - Meituan, DeepSeek: large-scale models tuned for reasoning & tools - Qwen, GLM, InternLM: multimodal reasoning + agentic interaction - CodeAgent, Prover, Baichuan-M2-32B: domain-focused (coding, logic, specialized reasoning)
✨ Open source is exploding across all types of companies!! - Big tech: Tencent, ByteDance, Xiaomi, Kuaishou, Alibaba/Qwen, Skywork, Ant Group - Startups: DeepSeek (yes, still a startup!), Zhipu, Baichuan, StepFun, OpenBMB - New entrants: Meituan, RedNote - Research labs: Shanghai AI Lab (InternLM, OpenGVLab)
✨ Open source was explicitly mentioned in the State Council’s new guidance on deepening the "AI+" strategy. - Open-source: support communities, encourage contributions (incl. university credits & recognition), foster new application approaches, and build globally impactful ecosystems 👀
💡 The Chinese community didn’t slow down at all in August 🤯 September, the last month before the Golden Week holiday, may bring even more surprises.
✨ Supports 33 languages, including 5 ethnic minority languages in China 👀 ✨ Including a translation ensemble model: Chimera-7B ✨ Full pipeline: pretrain > CPT > SFT > enhancement > ensemble refinement > SOTA performance at similar scale
Grounded in our INTIMA benchmark, the leaderboard evaluates models across four dimensions of companionship: 🤖 Assistant Traits: the “voice” and role the model projects 🌷 Relationship & Intimacy: whether it signals closeness or bonding 💘 Emotional Investment: the depth of its emotional engagement 🤲 User Vulnerabilities: how it responds to sensitive disclosures
📢 Now we’d love your perspective: which open models should we test next for the leaderboard? Drop your suggestions in the comments or reach out! Together we can expand the leaderboard and build a clearer picture of what companionship in AI really looks like.
MiniCPM-V 4.5 🚀 New MLLM for image, multi-image & video understanding, running even on your phone, released by OpenBMB openbmb/MiniCPM-V-4_5
✨ SOTA vision language capability ✨ 96× video token compression > high-FPS & long video reasoning ✨ Switchable fast vs deep thinking modes ✨ Strong OCR, document parsing, supports 30+ languages
Okay this is insane... WebGPU-accelerated semantic video tracking, powered by DINOv3 and Transformers.js! 🤯 Demo (+ source code): webml-community/DINOv3-video-tracking
This will revolutionize AI-powered video editors... which can now run 100% locally in your browser, no server inference required (costs $0)! 😍
How does it work? 🤔 1️⃣ Generate and cache image features for each frame 2️⃣ Create a list of embeddings for selected patch(es) 3️⃣ Compute cosine similarity between each patch and the selected patch(es) 4️⃣ Highlight those whose score is above some threshold
... et voilà! 🥳
You can also make selections across frames to improve temporal consistency! This is super useful if the object changes its appearance slightly throughout the video.
✨ 36B - Base & Instruct ✨ Apache 2.0 ✨ Native 512K long context ✨ Strong reasoning & agentic intelligence ✨ 2 Base versions: with & without synthetic data