John Smith PRO
John6666
AI & ML interests
None yet
Recent Activity
reacted
to
seawolf2357's
post
with 🔥
5 minutes ago
🎨 Ghibli-Style Image Generation with Multilingual Text Integration: FLUX.1 Hugging Face Edition 🌏✨
Hello creators! Today I'm introducing a special image generator that combines the beautiful aesthetics of Studio Ghibli with multilingual text integration! 😍
https://huggingface.co/spaces/seawolf2357/Ghibli-Multilingual-Text-rendering
✨ Key Features
Ghibli-Style Image Generation - High-quality animation-style images based on FLUX.1
Multilingual Text Rendering - Support for Korean, Japanese, English, and all languages! 🇰🇷🇯🇵🇬🇧
Automatic Image Editing with Simple Prompts - Just input your desired text and you're done!
Two Stylistic Variations Provided - Get two different results from a single prompt
Full Hugging Face Spaces Support - Deploy and share instantly!
🚀 How Does It Work?
Enter a prompt describing your desired image (e.g., "a cat sitting by the window")
Input the text you want to add (any language works!)
Select the text position, size, and color
Two different versions are automatically generated!
💯 Advantages of This Model
No Tedious Post-Editing Needed - Text is perfectly integrated during generation
Natural Text Integration - Text automatically adjusts to match the image style
Perfect Multilingual Support - Any language renders beautifully!
User-Friendly Interface - Easily adjust text size, position, and color
One-Click Hugging Face Deployment - Use immediately without complex setup
🎭 Use Cases
Creating multilingual greeting cards
Animation-style social media content
Ghibli-inspired posters or banners
Character images with dialogue in various languages
Sharing with the community through Hugging Face Spaces
This project leverages Hugging Face's FLUX.1 model to open new possibilities for seamlessly integrating high-quality Ghibli-style images with multilingual text using just prompts! 🌈
Try it now and create your own artistic masterpieces! 🎨✨
#GhibliStyle #MultilingualSupport #AIImageGeneration #TextRendering #FLUX #HuggingFace
Organizations
John6666's activity

reacted to
gkotte's
post with 👀
1 minute ago
If you have any problems with the verification process on the Hugging Face Discord, send a DM to lunarflu
on Discord and HF will respond quickly.
https://discuss.huggingface.co/t/cant-access-hf-discord-after-verification-been-trying-over-a-month/137429/15

reacted to
seawolf2357's
post with 🔥
5 minutes ago
Post
123
🎨 Ghibli-Style Image Generation with Multilingual Text Integration: FLUX.1 Hugging Face Edition 🌏✨
Hello creators! Today I'm introducing a special image generator that combines the beautiful aesthetics of Studio Ghibli with multilingual text integration! 😍
seawolf2357/Ghibli-Multilingual-Text-rendering
✨ Key Features
Ghibli-Style Image Generation - High-quality animation-style images based on FLUX.1
Multilingual Text Rendering - Support for Korean, Japanese, English, and all languages! 🇰🇷🇯🇵🇬🇧
Automatic Image Editing with Simple Prompts - Just input your desired text and you're done!
Two Stylistic Variations Provided - Get two different results from a single prompt
Full Hugging Face Spaces Support - Deploy and share instantly!
🚀 How Does It Work?
Enter a prompt describing your desired image (e.g., "a cat sitting by the window")
Input the text you want to add (any language works!)
Select the text position, size, and color
Two different versions are automatically generated!
💯 Advantages of This Model
No Tedious Post-Editing Needed - Text is perfectly integrated during generation
Natural Text Integration - Text automatically adjusts to match the image style
Perfect Multilingual Support - Any language renders beautifully!
User-Friendly Interface - Easily adjust text size, position, and color
One-Click Hugging Face Deployment - Use immediately without complex setup
🎭 Use Cases
Creating multilingual greeting cards
Animation-style social media content
Ghibli-inspired posters or banners
Character images with dialogue in various languages
Sharing with the community through Hugging Face Spaces
This project leverages Hugging Face's FLUX.1 model to open new possibilities for seamlessly integrating high-quality Ghibli-style images with multilingual text using just prompts! 🌈
Try it now and create your own artistic masterpieces! 🎨✨
#GhibliStyle #MultilingualSupport #AIImageGeneration #TextRendering #FLUX #HuggingFace
Hello creators! Today I'm introducing a special image generator that combines the beautiful aesthetics of Studio Ghibli with multilingual text integration! 😍
seawolf2357/Ghibli-Multilingual-Text-rendering
✨ Key Features
Ghibli-Style Image Generation - High-quality animation-style images based on FLUX.1
Multilingual Text Rendering - Support for Korean, Japanese, English, and all languages! 🇰🇷🇯🇵🇬🇧
Automatic Image Editing with Simple Prompts - Just input your desired text and you're done!
Two Stylistic Variations Provided - Get two different results from a single prompt
Full Hugging Face Spaces Support - Deploy and share instantly!
🚀 How Does It Work?
Enter a prompt describing your desired image (e.g., "a cat sitting by the window")
Input the text you want to add (any language works!)
Select the text position, size, and color
Two different versions are automatically generated!
💯 Advantages of This Model
No Tedious Post-Editing Needed - Text is perfectly integrated during generation
Natural Text Integration - Text automatically adjusts to match the image style
Perfect Multilingual Support - Any language renders beautifully!
User-Friendly Interface - Easily adjust text size, position, and color
One-Click Hugging Face Deployment - Use immediately without complex setup
🎭 Use Cases
Creating multilingual greeting cards
Animation-style social media content
Ghibli-inspired posters or banners
Character images with dialogue in various languages
Sharing with the community through Hugging Face Spaces
This project leverages Hugging Face's FLUX.1 model to open new possibilities for seamlessly integrating high-quality Ghibli-style images with multilingual text using just prompts! 🌈
Try it now and create your own artistic masterpieces! 🎨✨
#GhibliStyle #MultilingualSupport #AIImageGeneration #TextRendering #FLUX #HuggingFace

reacted to
Pendrokar's
post with 🔥
about 5 hours ago
Post
309
🏆 TTS Arena: Added the Orpheus-TTS and CSM TTS model Spaces to the Arena Fork
🏆 Pendrokar/TTS-Spaces-Arena
Orpheus has been able to get into the Top 5 of available models.
Orpheus-TTS 🚀: MohamedRashad/Orpheus-TTS
Conversational Speech Model🌱: sesame/csm-1b
🏆 Pendrokar/TTS-Spaces-Arena
Orpheus has been able to get into the Top 5 of available models.
Orpheus-TTS 🚀: MohamedRashad/Orpheus-TTS
Conversational Speech Model🌱: sesame/csm-1b

reacted to
jeffboudier's
post with 🤗
about 5 hours ago
Post
270
Enterprise orgs now enable serverless Inference Providers for all members
- includes $2 free usage per org member (e.g. an Enterprise org with 1,000 members share $2,000 free credit each month)
- admins can set a monthly spend limit for the entire org
- works today with Together, fal, Novita, Cerebras and HF Inference.
Here's the doc to bill Inference Providers usage to your org: https://huggingface.co/docs/inference-providers/pricing#organization-billing
- includes $2 free usage per org member (e.g. an Enterprise org with 1,000 members share $2,000 free credit each month)
- admins can set a monthly spend limit for the entire org
- works today with Together, fal, Novita, Cerebras and HF Inference.
Here's the doc to bill Inference Providers usage to your org: https://huggingface.co/docs/inference-providers/pricing#organization-billing

reacted to
mrfakename's
post with 🚀
about 5 hours ago
Post
182
Papla P1 from Papla Media is now available on the TTS Arena!
Try out Papla's new ultra-realistic TTS model + compare it with other leading models on the TTS Arena: TTS-AGI/TTS-Arena
Try out Papla's new ultra-realistic TTS model + compare it with other leading models on the TTS Arena: TTS-AGI/TTS-Arena

reacted to
hexgrad's
post with 👀
about 5 hours ago
Post
442
To Meta AI Research: I would like to fold
ylacombe/expresso into the training mix of an Apache TTS model series. Can you relax the Expresso dataset license to CC-BY or more permissive?
Barring that, can I have an individual exception to train on the materials and distribute trained Apache models, without direct redistribution of the original files? Thanks!
CC (Expresso paper authors whose handles I could find on HF) @wnhsu @adavirro @bowenshi @itaigat @TalRemez @JadeCopet @hassid @felixkreuk @adiyoss @edupoux
Barring that, can I have an individual exception to train on the materials and distribute trained Apache models, without direct redistribution of the original files? Thanks!
CC (Expresso paper authors whose handles I could find on HF) @wnhsu @adavirro @bowenshi @itaigat @TalRemez @JadeCopet @hassid @felixkreuk @adiyoss @edupoux

reacted to
fdaudens's
post with 🔥
about 5 hours ago
Post
390
Did we just drop personalized AI evaluation?! This tool auto-generates custom benchmarks on your docs to test which models are the best.
Most benchmarks test general capabilities, but what matters is how models handle your data and tasks. YourBench helps answer critical questions like:
- Do you really need a hundreds-of-billions-parameter model sledgehammer to crack a nut?
- Could a smaller, fine-tuned model work better?
- How well do different models understand your domain?
Some cool features:
📚 Generates custom benchmarks from your own documents (PDFs, Word, HTML)
🎯 Tests models on real tasks, not just general capabilities
🔄 Supports multiple models for different pipeline stages
🧠 Generate both single-hop and multi-hop questions
🔍 Evaluate top models and deploy leaderboards instantly
💰 Full cost analysis to optimize for your budget
🛠️ Fully configurable via a single YAML file
26 SOTA models tested for question generation. Interesting finding: Qwen2.5 32B leads in question diversity, while smaller Qwen models and Gemini 2.0 Flash offer great value for cost.
You can also run it locally on any models you want.
I'm impressed. Try it out: yourbench/demo
Most benchmarks test general capabilities, but what matters is how models handle your data and tasks. YourBench helps answer critical questions like:
- Do you really need a hundreds-of-billions-parameter model sledgehammer to crack a nut?
- Could a smaller, fine-tuned model work better?
- How well do different models understand your domain?
Some cool features:
📚 Generates custom benchmarks from your own documents (PDFs, Word, HTML)
🎯 Tests models on real tasks, not just general capabilities
🔄 Supports multiple models for different pipeline stages
🧠 Generate both single-hop and multi-hop questions
🔍 Evaluate top models and deploy leaderboards instantly
💰 Full cost analysis to optimize for your budget
🛠️ Fully configurable via a single YAML file
26 SOTA models tested for question generation. Interesting finding: Qwen2.5 32B leads in question diversity, while smaller Qwen models and Gemini 2.0 Flash offer great value for cost.
You can also run it locally on any models you want.
I'm impressed. Try it out: yourbench/demo

reacted to
BFFree's
post with 🚀
about 5 hours ago
Post
303
I continue to experiment with taking a sketch to full 3D model using just AI workflow. Yesterday I tried (and loved)
VAST-AI/TripoSG
The results of the texture and accuracy of guessing the rest of a random object are a marvel. This is an insect meets crab character I first created in 2D and the short video attached show the results! Great work!
The results of the texture and accuracy of guessing the rest of a random object are a marvel. This is an insect meets crab character I first created in 2D and the short video attached show the results! Great work!

reacted to
clem's
post with 🤗
about 5 hours ago
Post
341
Now in Enterprise Hub organizations, you can centralize your billing not only for HF usage but also inference through our inference partners.
Will prevent some headaches for your finance & accounting teams haha (so feel free to share that with them).
Will prevent some headaches for your finance & accounting teams haha (so feel free to share that with them).

reacted to
openfree's
post with 🔥
about 15 hours ago
Post
1987
🔥 'Open Meme Studio': Your Creative Meme Factory 🎭✨
Hello everyone! Today I'm introducing 'Open Meme Studio', an amazing space where you can easily create and transform fun and original meme images. 🚀
VIDraft/Open-Meme-Studio
🎯 Taking Meme Creation to the Next Level!
This application leverages the powerful Kolors model and IP-Adapter-Plus to upgrade your meme-making abilities. Go beyond simple image editing and experience a completely new meme world powered by AI!
🛠️ Features You'll Love
📸 Transform and reinterpret existing meme templates
🎭 Freely change expressions and poses
👓 Add props (sunglasses, hats, etc.)
🏞️ Change backgrounds and composite characters
🎨 Apply various artistic styles
💪 Why 'Open Meme Studio' is So Effective
Fast Meme Generation: High-quality memes completed in seconds
Unlimited Creativity: Completely different results just by changing prompts
User-Friendly Interface: Simple prompt input and image upload is all you need
Fine-tuned Control: Adjust how much of the original image characteristics to preserve
Advanced User Options: Freely set seed values, resolution, number of steps, and more
🚀 Streamlined Meme Creation Process
Tasks that previously required complex tools like Photoshop can now be accomplished with just a few simple prompts. Experience intuitive image manipulation through text commands.
🌈 Effective Prompt Examples
😎 "sunglass" - Add cool sunglasses to your character
🏔️ "background alps" - Change the background to Alpine mountains
💃 "dancing" - Transform your character into a dancing pose
😁 "smile" - Change to a smiling expression
🎮 "with Pikachu" - Create a scene with Pikachu
🎨 "3d style" - Convert to 3D style
🔗 Join Our Community
For more meme creation tips and interaction with other users, join our Discord!
https://discord.gg/openfreeai
Start creating unique memes that will shake up social media with 'Open Meme Studio' right now! 🚀💯 It's time for your meme
Hello everyone! Today I'm introducing 'Open Meme Studio', an amazing space where you can easily create and transform fun and original meme images. 🚀
VIDraft/Open-Meme-Studio
🎯 Taking Meme Creation to the Next Level!
This application leverages the powerful Kolors model and IP-Adapter-Plus to upgrade your meme-making abilities. Go beyond simple image editing and experience a completely new meme world powered by AI!
🛠️ Features You'll Love
📸 Transform and reinterpret existing meme templates
🎭 Freely change expressions and poses
👓 Add props (sunglasses, hats, etc.)
🏞️ Change backgrounds and composite characters
🎨 Apply various artistic styles
💪 Why 'Open Meme Studio' is So Effective
Fast Meme Generation: High-quality memes completed in seconds
Unlimited Creativity: Completely different results just by changing prompts
User-Friendly Interface: Simple prompt input and image upload is all you need
Fine-tuned Control: Adjust how much of the original image characteristics to preserve
Advanced User Options: Freely set seed values, resolution, number of steps, and more
🚀 Streamlined Meme Creation Process
Tasks that previously required complex tools like Photoshop can now be accomplished with just a few simple prompts. Experience intuitive image manipulation through text commands.
🌈 Effective Prompt Examples
😎 "sunglass" - Add cool sunglasses to your character
🏔️ "background alps" - Change the background to Alpine mountains
💃 "dancing" - Transform your character into a dancing pose
😁 "smile" - Change to a smiling expression
🎮 "with Pikachu" - Create a scene with Pikachu
🎨 "3d style" - Convert to 3D style
🔗 Join Our Community
For more meme creation tips and interaction with other users, join our Discord!
https://discord.gg/openfreeai
Start creating unique memes that will shake up social media with 'Open Meme Studio' right now! 🚀💯 It's time for your meme

reacted to
AdinaY's
post with 🔥
about 15 hours ago
Post
1001
Dolphin 🐬 an open ASR model released by DataOceanAI, one of the biggest AI data provider in China 🔥
✨ Supports 40 Eastern languages & 22 Chinese dialects
✨ Apache2.0
✨ With 21.2M hours of data (7.4M open data)
Model:
DataoceanAI/dolphin-base
DataoceanAI/dolphin-small
Paper:
Dolphin: A Large-Scale Automatic Speech Recognition Model for Eastern Languages (2503.20212)
✨ Supports 40 Eastern languages & 22 Chinese dialects
✨ Apache2.0
✨ With 21.2M hours of data (7.4M open data)
Model:
DataoceanAI/dolphin-base
DataoceanAI/dolphin-small
Paper:
Dolphin: A Large-Scale Automatic Speech Recognition Model for Eastern Languages (2503.20212)

reacted to
Jaward's
post with 🤗
about 15 hours ago
Post
886
Implements from first-principle recently proposed dynamic tanh as alternative to layernorm. Specifically, we trained a nanoGPT (0.8 M params) on tiny shakespeare with conventional layernorm, RMSNorm and dynamic tanh, then compared performances. Observed performance seems to match or is stable for α = 0.5~ 1.5, might outperform if trained longer.
Code: https://github.com/Jaykef/ai-algorithms/blob/main/Dynamic_Tanh.ipynb
Background music by 周子珺
Code: https://github.com/Jaykef/ai-algorithms/blob/main/Dynamic_Tanh.ipynb
Background music by 周子珺

reacted to
ginipick's
post with 🔥
about 15 hours ago
Post
1825
🏯 Open Ghibli Studio: Transform Your Photos into Ghibli-Style Artwork! ✨
Hello AI enthusiasts! 🙋♀️ Today I'm introducing a truly magical project: Open Ghibli Studio 🎨
ginigen/FLUX-Open-Ghibli-Studio
🌟 What Can It Do?
Upload any regular photo and watch it transform into a beautiful, fantastical image reminiscent of Hayao Miyazaki's Studio Ghibli animations! 🏞️✨
🔧 How Does It Work?
📸 Upload your photo
🤖 Florence-2 AI analyzes the image and generates a description
✏️ "Ghibli style" is added to the description
🎭 Magic transformation happens using the FLUX.1 model and Ghibli LoRA!
⚙️ Customization Options
Want more control? Adjust these in the advanced settings:
🎲 Set a seed (for reproducible results)
📏 Adjust image dimensions
🔍 Guidance scale (prompt adherence)
🔄 Number of generation steps
💫 Ghibli style intensity
🚀 Try It Now!
Click the "Transform to Ghibli Style" button below to create your own Ghibli world! Ready to meet Totoro, Howl, Sophie, or Chihiro? 🌈
🌿 Note: For best results, use clear images. Nature landscapes, buildings, and portraits transform especially well!
💖 Enjoy the magical transformation! Add some Ghibli magic to your everyday life~ ✨
Hello AI enthusiasts! 🙋♀️ Today I'm introducing a truly magical project: Open Ghibli Studio 🎨
ginigen/FLUX-Open-Ghibli-Studio
🌟 What Can It Do?
Upload any regular photo and watch it transform into a beautiful, fantastical image reminiscent of Hayao Miyazaki's Studio Ghibli animations! 🏞️✨
🔧 How Does It Work?
📸 Upload your photo
🤖 Florence-2 AI analyzes the image and generates a description
✏️ "Ghibli style" is added to the description
🎭 Magic transformation happens using the FLUX.1 model and Ghibli LoRA!
⚙️ Customization Options
Want more control? Adjust these in the advanced settings:
🎲 Set a seed (for reproducible results)
📏 Adjust image dimensions
🔍 Guidance scale (prompt adherence)
🔄 Number of generation steps
💫 Ghibli style intensity
🚀 Try It Now!
Click the "Transform to Ghibli Style" button below to create your own Ghibli world! Ready to meet Totoro, Howl, Sophie, or Chihiro? 🌈
🌿 Note: For best results, use clear images. Nature landscapes, buildings, and portraits transform especially well!
💖 Enjoy the magical transformation! Add some Ghibli magic to your everyday life~ ✨

reacted to
AdinaY's
post with 🚀
about 15 hours ago
Post
392
MegaTTS3 📢 an open TTS released by ByteDance
✨ 0.45B with Apache2.0
✨ Support English & Chinese
✨ High quality voice cloning
✨ Accent Intensity Control
ByteDance/MegaTTS3
✨ 0.45B with Apache2.0
✨ Support English & Chinese
✨ High quality voice cloning
✨ Accent Intensity Control
ByteDance/MegaTTS3

replied to
their
post
about 19 hours ago

reacted to
onekq's
post with 👍
1 day ago
Post
1245
Adding MLX version of OneSQL 7B for MacBook (Apple Silicon) users
onekq-ai/OneSQL-v0.1-Qwen-7B-MLX-4bit
This model has the best accuracy among all quantized versions (AWX, GGUF etc.), which I am very happy about.
I tested this model on my MacBook Air with M1 processor and 8GB of RAM, which is the lower bound of Apple Silicon, also the earliest and still the most popular. On average it took 16 seconds to generate a SQL query, and one minute in the worst case. If you own a newer MacBook with M2 or M3, the speed should be considerably faster.
I hope the MLX team will improve inference speed by software tricks (definitely doable) in the future. Meanwhile, if you find the current inference speed acceptable, you are more than welcome to enjoy this model. 🤗
onekq-ai/OneSQL-v0.1-Qwen-7B-MLX-4bit
This model has the best accuracy among all quantized versions (AWX, GGUF etc.), which I am very happy about.
I tested this model on my MacBook Air with M1 processor and 8GB of RAM, which is the lower bound of Apple Silicon, also the earliest and still the most popular. On average it took 16 seconds to generate a SQL query, and one minute in the worst case. If you own a newer MacBook with M2 or M3, the speed should be considerably faster.
I hope the MLX team will improve inference speed by software tricks (definitely doable) in the future. Meanwhile, if you find the current inference speed acceptable, you are more than welcome to enjoy this model. 🤗

reacted to
vincentg64's
post with 🔥
1 day ago
Post
1495
The Rise of Specialized LLMs for Enterprise -https://mltblog.com/3QXXE4I
In this article, I discuss the main problems of standard LLMs (OpenAI and the likes), and how the new generation of LLMs addresses these issues. The focus is on Enterprise LLMs.
LLMs with Billions of Parameters: Most of the LLMs still fall in that category. The first ones (ChatGPT) appeared around 2022, though Bert is an early precursor. Most recent books discussing LLMs still define them as transformer architecture with deep neural networks (DNNs), costly training, and reliance on GPUs. The training is optimized to predict the next tokens or missing tokens. However, this task is remotely relevant to what modern LLMs now deliver to the user, see here. Yet it requires time and intensive computer resources. Indeed, this type of architecture works best with billions or trillions of tokens. In the end, most of these tokens are noise, requiring smart distillation for performance improvement.
The main issues are:
➡️ Performance: Requires GPU and large corpuses as input data. Re-training is expensive. Hallucinations are still a problem. Fine-tuning is delicate (Blackbox). You need prompt engineering to get the best results. Mixtures of experts (multiple sub-LLMs, DeepSeek) is one step towards improving accuracy.
➡️ Cost: Besides the GPU costs, the pricing model charges by the token, incentivizing vendors to use models with billions of tokens.
Read full article describing more issues and how LLM 2.0 addresses them, at https://mltblog.com/3QXXE4I
More links:
- To receive latest updates: https://mltblog.com/4iTvQec
- About LLM 2.0: https://mltblog.com/4g2sKTv
- PowerPoint presentation: https://mltblog.com/43DYviE
- Our company website: https://mlt
In this article, I discuss the main problems of standard LLMs (OpenAI and the likes), and how the new generation of LLMs addresses these issues. The focus is on Enterprise LLMs.
LLMs with Billions of Parameters: Most of the LLMs still fall in that category. The first ones (ChatGPT) appeared around 2022, though Bert is an early precursor. Most recent books discussing LLMs still define them as transformer architecture with deep neural networks (DNNs), costly training, and reliance on GPUs. The training is optimized to predict the next tokens or missing tokens. However, this task is remotely relevant to what modern LLMs now deliver to the user, see here. Yet it requires time and intensive computer resources. Indeed, this type of architecture works best with billions or trillions of tokens. In the end, most of these tokens are noise, requiring smart distillation for performance improvement.
The main issues are:
➡️ Performance: Requires GPU and large corpuses as input data. Re-training is expensive. Hallucinations are still a problem. Fine-tuning is delicate (Blackbox). You need prompt engineering to get the best results. Mixtures of experts (multiple sub-LLMs, DeepSeek) is one step towards improving accuracy.
➡️ Cost: Besides the GPU costs, the pricing model charges by the token, incentivizing vendors to use models with billions of tokens.
Read full article describing more issues and how LLM 2.0 addresses them, at https://mltblog.com/3QXXE4I
More links:
- To receive latest updates: https://mltblog.com/4iTvQec
- About LLM 2.0: https://mltblog.com/4g2sKTv
- PowerPoint presentation: https://mltblog.com/43DYviE
- Our company website: https://mlt

reacted to
hesamation's
post with ❤️
1 day ago
Post
1659
What, How, Where, and How Well? This paper reviews test-time scaling methods and all you need to know about them:
> parallel, sequential, hybrid, internal scaling
> how to scale (SFT, RL, search, verification)
> metrics and evals of test-time scaling
🔗paper: What, How, Where, and How Well? A Survey on Test-Time Scaling in Large Language Models (2503.24235)
If you want to learn what inference-time compute scaling is @rasbt has a great blog post on that:
https://magazine.sebastianraschka.com/p/state-of-llm-reasoning-and-inference-scaling
> parallel, sequential, hybrid, internal scaling
> how to scale (SFT, RL, search, verification)
> metrics and evals of test-time scaling
🔗paper: What, How, Where, and How Well? A Survey on Test-Time Scaling in Large Language Models (2503.24235)
If you want to learn what inference-time compute scaling is @rasbt has a great blog post on that:
https://magazine.sebastianraschka.com/p/state-of-llm-reasoning-and-inference-scaling

posted
an
update
1 day ago
Post
1370
I used up my Zero GPU Quota yesterday (about 12 hours ago). At the time, I got a message saying “Retry at 13:45 (approx.)”, but now it's just changed to “Retry at 03:22”.
Anyway, everyone, let's be careful not to use up our Quota...
Related: https://huggingface.co/posts/Keltezaa/754755723533287#67e6ed5e3394f1ed9ca41dbd
Anyway, everyone, let's be careful not to use up our Quota...
Related: https://huggingface.co/posts/Keltezaa/754755723533287#67e6ed5e3394f1ed9ca41dbd