@hf-staff Do you have any insight into what might be the issue here?
James Clarke
AI & ML interests
Recent Activity
Organizations
Impulse2000's activity


I am not sure if that makes sense, I am under the impression that, if the space is not running(not started), no models can be actively loaded in the space.
Can you share your relevant workflow(docker-compose, app code, etc) so i can see more clearly whats happening?
I might be able to aid in a solution, its possible that there is an issue in the workflow itself.
EDIT: I looked at the spaces, Do you mean this space as an example? 'https://huggingface.co/spaces/nroggendorff/train-llama'
Because this space shows a missing "CUDA_HOME" env var, most your other spaces throwing errors about missing CUDA drivers or are paused. These are configuration errors.
Could you tell me the space and error message?
I might be able to help you fix it.

What policies? all my spaces that use docker work fine.
And i got no email about policy changes.
Please elaborate.
Why tagged as 'Text Generation'?

π© Report: Illegal or restricted content

π© Report: Illegal or restricted content

Add not-for-all-audiances tag to comply with HF guidelines.

I've used it before when i used to be an OpenAI customer, Its good, not as good as you might think though. I assume LMArena, etc will do some benches too, as well as simple bench, etc.
Anyways, DeepSeek R1 is really good and cost effective, and they are allegedly making great progress on R2.

Hello, I've prepared two datasets (raw and cleaned) of the recently declassified documents related to the assassination of President John F. Kennedy.
Raw zlatinb/jfk-2025-raw
Cleaned zlatinb/jfk-2025-cleaned
The 2182 documents cover a vast range of topics, so it may be interesting to train on them to generate insights.

I noticed that Gemma 3 was much more resilient to refusal removal than other models like Qwen 2.5.
I experimented with different recipes and improved the abliteration technique I wrote about last year.
It's still experimental but the refusal rate is super low in my tests. Enjoy!
mlabonne/gemma-3-4b-it-abliterated
mlabonne/gemma-3-12b-it-abliterated
mlabonne/gemma-3-27b-it-abliterated

I made a Colab notebook to automatically abliterate models.
It's quite general, so you can do interesting stuff like blocking a given language in the model outputs.
π» Colab: https://colab.research.google.com/drive/1RmLv-pCMBBsQGXQIM8yF-OdCNyoylUR1?usp=sharing
Diffusers Roadmap?
