Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
base_model:
|
4 |
+
- Qwen/Qwen3-32B
|
5 |
+
tags:
|
6 |
+
- '#aha'
|
7 |
+
- '#health'
|
8 |
+
- '#nutrition'
|
9 |
+
- '#medicinalherbs'
|
10 |
+
- '#fasting'
|
11 |
+
- '#faith'
|
12 |
+
- '#healing'
|
13 |
+
- '#bitcoin'
|
14 |
+
- '#nostr'
|
15 |
+
---
|
16 |
+
|
17 |
+

|
18 |
+
|
19 |
+
|
20 |
+
|
21 |
+
Ostrich 32B - bringing you "the knowledge that matters".
|
22 |
+
|
23 |
+
- Health, nutrition, medicinal herbs
|
24 |
+
- Fasting, faith, healing
|
25 |
+
- Liberating technologies like bitcoin and nostr
|
26 |
+
|
27 |
+
This is a similar project to my [Gemma 3 fine tuning](https://huggingface.co/blog/etemiz/fine-tuning-gemma-3-for-human-alignment), but here Qwen3 is the base model. There is a difference in the loop. For Gemma3 I was training in 4 bit, merging loras to the 16 bit model, taking gguf, measuring progress. For Qwen3 I saved in 4 bit, measured in 4 bit bitsandbytes and only worked with 16 bit models for merging using MergeKit.
|
28 |
+
|
29 |
+
Gemma 3 was resisting against faith and fasting domains. This one was happy to learn about those but still resisted in other domains. I will fine tune more and publish more models with proper AHA scores. For now this is a temporary release.
|
30 |
+
|
31 |
+
Purpose of the work: [Curation is All You Need](https://huggingface.co/blog/etemiz/curation-is-all-you-need).
|
32 |
+
|
33 |
+
Check out [the original Ostrich](https://huggingface.co/some1nostr/Ostrich-70B).
|
34 |
+
|
35 |
+
Project sponsored by [PickaBrain](https://pickabrain.ai).
|
36 |
+
|
37 |
+
If you want the better human alignment you can talk to [Kooby](https://pickabrain.ai/ko).
|