Aton Mountlook

AtonMountlook
ยท

AI & ML interests

None yet

Recent Activity

View all activity

Organizations

None yet

AtonMountlook's activity

reacted to Undi95's post with ๐Ÿ‘ 8 days ago
view post
Post
4549
Hi there!

If you want to create your own thinking model or do a better MistralThinker, I just uploaded my entire dataset made on Deepseek R1 and the axolotl config. (well I made them public)

Axolotl config : Undi95/MistralThinker-v1.1

The dataset : Undi95/R1-RP-ShareGPT3

You can also read all I did on those two discord screenshot from two days ago, I'm a little lazy to rewrite all kek.

Hope you will use them!
ยท
reacted to cogwheelhead's post with ๐Ÿ‘ 21 days ago
view post
Post
2514
Me and my team have performed an in-depth investigation comparing o1 to R1 (and other reasoning models)

Link: https://toloka.ai/blog/r1-is-not-on-par-with-o1-and-the-difference-is-qualitative-not-quantitative

It started with us evaluating them on our own university-math benchmarks: U-MATH for problem-solving and ฮผ-MATH for judging solution correctness (see the HF leaderboard: toloka/u-math-leaderboard)

tl;dr: R1 sure is amazing, but what we find is that it lags behind in novelty adaptation and reliability:
* performance drops when updating benchmarks with fresh unseen tasks (e.g. AIME 2024 -> 2025)
* R1-o1 gap widens when evaluating niche subdomains (e.g. university-specific math instead of the more common Olympiad-style contests)
* same with going into altogether unconventional domains (e.g. chess) or skills (e.g. judgment instead of problem-solving)
* R1 also runs into failure modes way more often (e.g. making illegal chess moves or falling into endless generation loops)

Our point here is not to bash on DeepSeek โ€” they've done exceptional work, R1 is a game-changer, and we have no intention to downplay that. R1's release is a perfect opportunity to study where all these models differ and gain understanding on how to move forward from here
reacted to Undi95's post with โค๏ธ 8 months ago
view post
Post
15419
Hello there,

New model released, my goal was to try finetune on the last Llama-3.1-8B-Instruct but not a small train, I wanted to do something useful.
One of the rare model that I didn't made for RP, or in the goal to uncensor it (but I did anyway kek).

The model was trained on 9M Claude conversations ONLY, giving him another writting style.

Undi95/Meta-Llama-3.1-8B-Claude > OG release fp32, it's the epoch 2
Undi95/Meta-Llama-3.1-8B-Claude-bf16 > Base model resharded in bf16 waiting for available quant without issues

Since it's frustrating to be censored using a local model, orthogonal activation steering was used, trying to force the model to never refuse a prompt.

Undi95/Meta-Llama-3.1-8B-Claude-68fail-3000total > Uncensored model, refuse 68 times on 3000 toxic prompt
Undi95/Meta-Llama-3.1-8B-Claude-39fail-3000total > Uncensored model, refuse 39 times on 3000 toxic prompt

It still refuse some prompt but the majority of them is uncensored. OAS can make a model more dumb or make the base perplexity go higher, so I didn't snipe for 0 refusal.

I don't do non-RP model a lot so any feedback is welcome, I would like to re-use this base for some others future project if needed.
ยท
reacted to Undi95's post with ๐Ÿค— 10 months ago
view post
Post
16536
Hey everyone,

Just wanted to shout out a massive thank you to all 2000 of you who've followed me on Hugging Face! ๐ŸŽ‰ It's incredible to have such an awesome crew backing me up as I dive into all these LLM experiments.

Even though not all my models turn out perfect, I've found some real gems and methods along the way ๐Ÿ’Ž. It's like digging for treasure โ€“ sometimes you found nothing, but sometimes you find a pearl, and sometimes you find a new method to try.

Your support and encouragement mean the world to me, and I'm really stoked to keep experimenting and learning. If you told me some years ago I would have so much people following me for what I do, I wouldn't have believed it. Here's to more discoveries and adventures ahead! ๐Ÿš€

Also, big thanks once again, and a huge shoutout to @IkariDev for being there through this journey and supporting me. I'm excited for our future work together and hope we will continue to make people happy! ๐Ÿ‘

I want to thank @Gryphe too, since my early work was heavily inspired from MythoMax and the RP/ERP vibe of it. If I'm here today it's probably because of you ๐Ÿ˜‚

I was so close to forget @chargoddard and his amazing tool too! What will we do without mergekit in our life? Thank you! ๐Ÿ™

See y'all at 3k!
ยท
reacted to abhishek's post with ๐Ÿ”ฅ 11 months ago
view post
Post
3481
With AutoTrain, you can already finetune the latest llama3 models without writing a single line of code. Here's an example finetune of llama3 8b model: abhishek/autotrain-llama3-no-robots
  • 2 replies
ยท
reacted to Undi95's post with ๐Ÿ‘ 12 months ago
view post
Post
Hey, it took some time but I finally moved out and got internet back, so here I am again!
A lot of things to get updated on, I will try to reply to each of you ASAP.
See you soon!
  • 1 reply
ยท
reacted to DmitryRyumin's post with ๐Ÿ‘ 12 months ago
view post
Post
1609
๐Ÿš€๐ŸŽญ๐ŸŒŸ New Research Alert! ๐ŸŒŸ ๐ŸŽญ๐Ÿš€
๐Ÿ“„ Title: FlashFace: Human Image Personalization with High-fidelity Identity Preservation ๐Ÿ”

๐Ÿ“ Description: FlashFace is a personalized photo editing tool that focuses on high-fidelity identity preservation and improved compliance through advanced encoding and integration strategies.

๐Ÿ‘ฅ Authors: Shilong Zhang, Lianghua Huang, @xichenhku et al.

๐Ÿ”— Paper: FlashFace: Human Image Personalization with High-fidelity Identity Preservation (2403.17008)

๐ŸŒ Github Page: https://jshilong.github.io/flashface-page

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #FlashFace #Personalization #HighFidelityIdentity #DeepLearning #Innovation