File size: 5,379 Bytes
6bc90b0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
---
license: mit
base_model:
- mistralai/Mistral-Small-24B-Instruct-2501
tags:
- symbolic-ai
- reasoning
- deductive-logic
- glyph-code-logic-flow
- mistral
- mlx
- gguf
- fine-tuned
- experimental
---

# Glyphstral-24B-v1 (Preview)

## Model Description

This is a **preview release (Version 1)** of a fine-tuned language model, **Glyphstral-24B-v1**, designed to understand and utilize the **Glyph Code Logic Flow (GCLF)** framework for structured, deductive symbolic reasoning.

This model is based on **Mistral-Small-24b** and has been fine-tuned using **MLX** with **DoRA (Decomposed Relative Attention)** at 4-bit quantization on Apple Silicon.

**Glyph Code Logic Flow (GCLF)** is a novel approach to symbolic AI aimed at enhancing reasoning and multi-dimensional thinking.  It provides a structured method for deductive reasoning using a symbolic language. You can explore the conceptual framework in detail here:

[Computational-Model-for-Symbolic-Representations GitHub Repository](https://github.com/severian42/Computational-Model-for-Symbolic-Representations/tree/main)

**Key Features (Version 1 - Preview):**

*   **Specialized for Glyph Code Logic Flow:** Fine-tuned to interpret and process instructions based on the GCLF framework.
*   **Deductive Reasoning Focus:**  Encourages structured, step-by-step deductive reasoning over probabilistic inference.
*   **Symbolic Manipulation:**  Trained to understand and manipulate symbolic representations within the GCLF framework.
*   **MLX Format:**  Currently provided in MLX format for efficient inference on Apple Silicon.
*   **Quantization:**  Fine-tuned and quantized to 4-bit for reduced memory footprint and faster inference (using MLX DoRA).
*   **Experimental V1 Release:** This is an initial release to showcase the potential of GCLF training. Expect ongoing development and improvements.

## Intended Use

This model is intended for **experimental use and research** in the following areas:

*   **Exploring Symbolic AI:**  Investigating the capabilities of language models for structured symbolic reasoning.
*   **Deductive Logic Applications:**  Building systems that require step-by-step, logically sound reasoning processes.
*   **Glyph Code Logic Flow Development:**  Experimenting with and refining the GCLF framework.
*   **Educational Purposes:**  Learning about symbolic AI, deductive reasoning, and structured knowledge representation.

**Limitations:**

*   **Version 1 - Preview:** This is an early version and may have limitations in robustness and generalization.
*   **Specialized Domain:**  Performance is optimized for tasks related to Glyph Code Logic Flow. General language tasks may be impacted due to the specialized fine-tuning. (Further evaluation is ongoing)
*   **Experimental Nature:** The GCLF framework itself is under development and this model reflects an early attempt to train an LLM for it.
*   **MLX Format (Initial):** Currently primarily available in MLX format, which may limit accessibility for users outside the Apple Silicon/MLX ecosystem (GGUF quantization is in progress).

## Training Data and Process

*   **Base Model:** Mistral-Small-24b
*   **Fine-tuning Method:** MLX-DoRA (Decomposed Relative Attention) at 4-bit quantization.
*   **Training Hardware:** Apple M2 (128GB RAM)
*   **Training Dataset:** Custom dataset of approximately 4500 examples specifically designed for Glyph Code Logic Flow. Each example was around 30,000 tokens in length, focused on detailed system instructions and GCLF tasks.
*   **Training Tokens:** Approximately 27 million tokens from the custom GCLF dataset.
*   **Training Duration:** 7 days (continuous 24/7 training).
*   **Initial Experiments:**  Initial training attempts were made with Deepeek R1-Qwen-14 and QWQ-32, but Mistral-Small-24b was found to be more receptive to the GCLF framework due to potentially less conflicting pre-trained reasoning biases.

## How to Use

**System Instructions**

## !! GGUF Quantization (Coming Soon) !!



# Version 2 and Future Development

Version 2 (In Development):

GRPO (Gradient Ratio Policy Optimization): Utilizing GRPO for potentially more stable and effective fine-tuning.

Newer Dataset: Training on an expanded and refined dataset for Glyph Code Logic Flow.

GGUF Release: Aiming for a GGUF release for wider accessibility and compatibility.

Improved Documentation: Comprehensive documentation and examples for using the model and understanding GCLF.

Ongoing Efforts:

Refining GCLF Framework: Continuously developing and improving the Glyph Code Logic Flow framework itself.

Performance Evaluation: Conducting thorough evaluations of the model's performance on GCLF tasks and general language understanding.

Community Feedback: Seeking feedback from the community to guide further development and improvements.

---

# Known Issues

The custom dataset and heavy use of symbols and operators seems to have potentially altered the models tool use. I've found that it often want to use it's `[TOOL_CALLS]` function at the end of it's response (sometimes also calling out `<SPECIAL_#>` tokens at the end). I think I know where this is stemming from, so hopefully v2 can avoid this potential issue altogether. 

If you are seeing the `[TOOL_CALLS]` and `<SPECIAL_>` outputs, you can set them as the EOS and it will align the model back into a more fluid conversation.