Neopy

theNeofr

AI & ML interests

RVC, UVR, SDXL, gradio

Recent Activity

updated a model about 17 hours ago
theNeofr/Panty-Anarchy
reacted to v2ray's post with ๐Ÿ‘ about 19 hours ago
GPT4chan Series Release GPT4chan is a series of models I trained on https://huggingface.co/datasets/v2ray/4chan dataset, which is based on https://huggingface.co/datasets/lesserfield/4chan-datasets. The dataset contains mostly posts from 2023. Not every board is included, for example, /pol/ is NOT included. To see which boards are included, visit https://huggingface.co/datasets/v2ray/4chan/tree/main/boards. This release contains 2 models sizes, 8B and 24B. The 8B model is based on https://huggingface.co/meta-llama/Llama-3.1-8B and the 24B model is based on https://huggingface.co/mistralai/Mistral-Small-24B-Base-2501. Why I made these models? Because for a long time after the original gpt-4chan model, there aren't any serious fine-tunes on 4chan datasets. 4chan is a good data source since it contains coherent replies and nice topics. It's fun to talk to an AI generated version of 4chan and get instant replies, and without the need to actually visit 4chan. You can also sort of analyze the content and behavior of 4chan posts by probing the model's outputs. Disclaimer: The GPT4chan models should only be used for research purposes, the outputs they generated do not represent the view of me on the subjects. Moderate the responses before sending it online. Model links: Full model: - https://huggingface.co/v2ray/GPT4chan-8B - https://huggingface.co/v2ray/GPT4chan-24B Adapter: - https://huggingface.co/v2ray/GPT4chan-8B-QLoRA - https://huggingface.co/v2ray/GPT4chan-24B-QLoRA AWQ: - https://huggingface.co/v2ray/GPT4chan-8B-AWQ - https://huggingface.co/v2ray/GPT4chan-24B-AWQ FP8: - https://huggingface.co/v2ray/GPT4chan-8B-FP8
liked a Space about 19 hours ago
Neogrconpet/docs-maker
View all activity

Organizations

Revocalize AI's profile picture Dev Mode Explorers's profile picture Hugging Face Discord Community's profile picture Theneos's profile picture None yet's profile picture Neo' setup's profile picture Web stuff's profile picture

theNeofr's activity

reacted to v2ray's post with ๐Ÿ‘ about 19 hours ago
view post
Post
1770
GPT4chan Series Release

GPT4chan is a series of models I trained on v2ray/4chan dataset, which is based on lesserfield/4chan-datasets. The dataset contains mostly posts from 2023. Not every board is included, for example, /pol/ is NOT included. To see which boards are included, visit v2ray/4chan.

This release contains 2 models sizes, 8B and 24B. The 8B model is based on meta-llama/Llama-3.1-8B and the 24B model is based on mistralai/Mistral-Small-24B-Base-2501.

Why I made these models? Because for a long time after the original gpt-4chan model, there aren't any serious fine-tunes on 4chan datasets. 4chan is a good data source since it contains coherent replies and nice topics. It's fun to talk to an AI generated version of 4chan and get instant replies, and without the need to actually visit 4chan. You can also sort of analyze the content and behavior of 4chan posts by probing the model's outputs.

Disclaimer: The GPT4chan models should only be used for research purposes, the outputs they generated do not represent the view of me on the subjects. Moderate the responses before sending it online.

Model links:

Full model:
- v2ray/GPT4chan-8B
- v2ray/GPT4chan-24B

Adapter:
- v2ray/GPT4chan-8B-QLoRA
- v2ray/GPT4chan-24B-QLoRA

AWQ:
- v2ray/GPT4chan-8B-AWQ
- v2ray/GPT4chan-24B-AWQ

FP8:
- v2ray/GPT4chan-8B-FP8
reacted to odellus's post with ๐Ÿง  10 days ago
view post
Post
1507
Tired: shitposting on bsky
Wired: shitposting on hf
  • 1 reply
ยท
reacted to burtenshaw's post with ๐Ÿš€ 11 days ago
view post
Post
2826
Manic few days in open source AI, with game changing development all over the place. Here's a round up of the resources:

- The science team at @huggingface reproduced and open source the seek r1. https://github.com/huggingface/open-r1
- @qwen released a series of models with 1 million token context! https://qwenlm.github.io/blog/qwen2.5-1m/
- SmolVLM got even smaller with completely new variants at 256m and 500m https://huggingface.co/blog/smolervlm

There's so much you could do with these developments. Especially combining them together into agentic applications or fine-tuning them on your use case.
  • 1 reply
ยท