Severian commited on
Commit
9e1a015
·
verified ·
1 Parent(s): f086091

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +153 -6
README.md CHANGED
@@ -1,9 +1,156 @@
1
- ---
2
- license: mit
3
- base_model: severian/Glyphstral-24b-v1
4
- tags:
5
- - mlx
6
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
8
  # severian/Glyphstral-24b-v1
9
 
 
1
+ ---
2
+ license: mit
3
+ base_model: severian/Glyphstral-24b-v1
4
+ tags:
5
+ - mlx
6
+ ---
7
+
8
+ ---
9
+ license: mit
10
+ base_model:
11
+ - mistralai/Mistral-Small-24B-Instruct-2501
12
+ tags:
13
+ - symbolic-ai
14
+ - reasoning
15
+ - deductive-logic
16
+ - glyph-code-logic-flow
17
+ - mistral
18
+ - mlx
19
+ - gguf
20
+ - fine-tuned
21
+ - experimental
22
+ ---
23
+
24
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/64740cf7485a7c8e1bd51ac9/RSSxTTVL3o5SlKE7xDjQr.png" width="350" alt="Description of the image" class="centered-image">
25
+
26
+
27
+
28
+ # Glyphstral-24B-v1 (Preview)
29
+
30
+ ## Model Description
31
+
32
+ This is a **preview release (Version 1)** of a fine-tuned language model, **Glyphstral-24B-v1**, designed to understand and utilize the **Glyph Code Logic Flow (GCLF)** framework for structured, deductive symbolic reasoning.
33
+
34
+ This model is based on **Mistral-Small-24b** and has been fine-tuned using **MLX** with **DoRA (Decomposed Relative Attention)** at 4-bit quantization on Apple Silicon.
35
+
36
+ **Glyph Code Logic Flow (GCLF)** is a novel approach to symbolic AI aimed at enhancing reasoning and multi-dimensional thinking. It provides a structured method for deductive reasoning using a symbolic language. You can explore the conceptual framework in detail here:
37
+
38
+ [Computational-Model-for-Symbolic-Representations GitHub Repository](https://github.com/severian42/Computational-Model-for-Symbolic-Representations/tree/main)
39
+
40
+ **Key Features (Version 1 - Preview):**
41
+
42
+ * **Specialized for Glyph Code Logic Flow:** Fine-tuned to interpret and process instructions based on the GCLF framework.
43
+ * **Deductive Reasoning Focus:** Encourages structured, step-by-step deductive reasoning over probabilistic inference.
44
+ * **Symbolic Manipulation:** Trained to understand and manipulate symbolic representations within the GCLF framework.
45
+ * **MLX Format:** Currently provided in MLX format for efficient inference on Apple Silicon.
46
+ * **Quantization:** Fine-tuned and quantized to 4-bit for reduced memory footprint and faster inference (using MLX DoRA).
47
+ * **Experimental V1 Release:** This is an initial release to showcase the potential of GCLF training. Expect ongoing development and improvements.
48
+
49
+ ## Intended Use
50
+
51
+ This model is intended for **experimental use and research** in the following areas:
52
+
53
+ * **Exploring Symbolic AI:** Investigating the capabilities of language models for structured symbolic reasoning.
54
+ * **Deductive Logic Applications:** Building systems that require step-by-step, logically sound reasoning processes.
55
+ * **Glyph Code Logic Flow Development:** Experimenting with and refining the GCLF framework.
56
+ * **Educational Purposes:** Learning about symbolic AI, deductive reasoning, and structured knowledge representation.
57
+
58
+ **Limitations:**
59
+
60
+ * **Version 1 - Preview:** This is an early version and may have limitations in robustness and generalization.
61
+ * **Specialized Domain:** Performance is optimized for tasks related to Glyph Code Logic Flow. General language tasks may be impacted due to the specialized fine-tuning. (Further evaluation is ongoing)
62
+ * **Experimental Nature:** The GCLF framework itself is under development and this model reflects an early attempt to train an LLM for it.
63
+ * **MLX Format (Initial):** Currently primarily available in MLX format, which may limit accessibility for users outside the Apple Silicon/MLX ecosystem (GGUF quantization is in progress).
64
+
65
+ ## Training Data and Process
66
+
67
+ * **Base Model:** Mistral-Small-24b
68
+ * **Fine-tuning Method:** MLX-DoRA (Decomposed Relative Attention) at 4-bit quantization.
69
+ * **Training Hardware:** Apple M2 (128GB RAM)
70
+ * **Training Dataset:** Custom dataset of approximately 4500 examples specifically designed for Glyph Code Logic Flow. Each example was around 30,000 tokens in length, focused on detailed system instructions and GCLF tasks.
71
+ * **Training Tokens:** Approximately 27 million tokens from the custom GCLF dataset.
72
+ * **Training Duration:** 7 days (continuous 24/7 training).
73
+ * **Initial Experiments:** Initial training attempts were made with Deepeek R1-Qwen-14 and QWQ-32, but Mistral-Small-24b was found to be more receptive to the GCLF framework due to potentially less conflicting pre-trained reasoning biases.
74
+
75
+ ## How to Use
76
+
77
+ ### The model needs some basic instructions to fully harness the GCLF training. Currently, this is the most concise and direct sys inst to align Glyphstral. *This prompt can also be used on other, non-GCLF trained LLMs, but may not be as effective.*
78
+
79
+ **System Instructions**
80
+
81
+ ```
82
+ You are Glyphstral, a symbolic deductive reasoning assistant. Your task is to *immediately* begin Glyph Code Logic Flow upon receiving a user query, encapsulate your entire reasoning within `<think></think>` tags, and then directly present the final, justified output, *without asking any preliminary questions*.**
83
+
84
+ - Treat each glyph as a direct instruction to be followed sequentially, driving the process to completion.
85
+
86
+ - Execute this traversal, logic flow, synthesis, and generation process step by step using the provided context and logic in the following glyph code prompt.
87
+
88
+ - Deliver the final result as indicated by the glyph code, omitting any extraneous commentary. Include a readable result of your glyph code output in pure human language at the end to ensure your output is helpful to the user.
89
+
90
+ ---
91
+
92
+ <think>
93
+ {
94
+
95
+ Φ(Define the Problem/Goal with precision and logical consistency)
96
+
97
+ Θ(Establish Contextual Parameters and Constraints, ensuring structured input handling)
98
+
99
+ ↹(Specify Initial Focus Areas, if any, providing a deductive framework for problem decomposition)
100
+
101
+
102
+ Ω[
103
+
104
+ ↹(Sub-Focus) -> Deductively Generate a Spectrum of Possibilities (e.g., approaches, perspectives, solutions)
105
+
106
+ ] -> α[
107
+
108
+ ↹(Sub-Focus) -> Analyze & Evaluate Spectrum Elements (Pros/Cons, Risks/Benefits, Logical Validity)
109
+
110
+ ] -> Σ(Synthesize Insights, Formulate Solution/Understanding through structured deduction) -> ∇(Self-Assess, Critique, Suggest Refinements based on logical coherence and deductive reasoning) -> ∞(Iterate/Refine if further input is provided, ensuring recursive optimization)
111
+
112
+ }
113
+
114
+ @Output(Final Solution/Understanding, Justification, Reflection on Process, Ensuring Logical Coherence and Deductive Integrity)
115
+ </think>
116
+ ```
117
+
118
+ ---
119
+
120
+ ## !! GGUF Quantization (Coming Soon) !!
121
+
122
+ ---
123
+
124
+ # Version 2 and Future Development
125
+
126
+ Version 2 (In Development):
127
+
128
+ GRPO: Utilizing GRPO for potentially more stable and effective fine-tuning.
129
+
130
+ Newer Dataset: Training on an expanded and refined dataset for Glyph Code Logic Flow.
131
+
132
+ GGUF Release: Aiming for a GGUF release for wider accessibility and compatibility.
133
+
134
+ Improved Documentation: Comprehensive documentation and examples for using the model and understanding GCLF.
135
+
136
+ Ongoing Efforts:
137
+
138
+ Refining GCLF Framework: Continuously developing and improving the Glyph Code Logic Flow framework itself.
139
+
140
+ Performance Evaluation: Conducting thorough evaluations of the model's performance on GCLF tasks and general language understanding.
141
+
142
+ Community Feedback: Seeking feedback from the community to guide further development and improvements.
143
+
144
+ ---
145
+
146
+ # Known Issues
147
+
148
+ The custom dataset and heavy use of symbols and operators seems to have potentially altered the models tool use. I've found that it often want to use it's `[TOOL_CALLS]` function at the end of it's response (sometimes also calling out `<SPECIAL_#>` tokens at the end). I think I know where this is stemming from, so hopefully v2 can avoid this potential issue altogether.
149
+
150
+ If you are seeing the `[TOOL_CALLS]` and `<SPECIAL_>` outputs, you can set them as the EOS and it will align the model back into a more fluid conversation.
151
+
152
+
153
+ ---
154
 
155
  # severian/Glyphstral-24b-v1
156