Glyphstral-24b-v1 / README.md
Severian's picture
Update README.md
6bc90b0 verified
|
raw
history blame
5.38 kB
metadata
license: mit
base_model:
  - mistralai/Mistral-Small-24B-Instruct-2501
tags:
  - symbolic-ai
  - reasoning
  - deductive-logic
  - glyph-code-logic-flow
  - mistral
  - mlx
  - gguf
  - fine-tuned
  - experimental

Glyphstral-24B-v1 (Preview)

Model Description

This is a preview release (Version 1) of a fine-tuned language model, Glyphstral-24B-v1, designed to understand and utilize the Glyph Code Logic Flow (GCLF) framework for structured, deductive symbolic reasoning.

This model is based on Mistral-Small-24b and has been fine-tuned using MLX with DoRA (Decomposed Relative Attention) at 4-bit quantization on Apple Silicon.

Glyph Code Logic Flow (GCLF) is a novel approach to symbolic AI aimed at enhancing reasoning and multi-dimensional thinking. It provides a structured method for deductive reasoning using a symbolic language. You can explore the conceptual framework in detail here:

Computational-Model-for-Symbolic-Representations GitHub Repository

Key Features (Version 1 - Preview):

  • Specialized for Glyph Code Logic Flow: Fine-tuned to interpret and process instructions based on the GCLF framework.
  • Deductive Reasoning Focus: Encourages structured, step-by-step deductive reasoning over probabilistic inference.
  • Symbolic Manipulation: Trained to understand and manipulate symbolic representations within the GCLF framework.
  • MLX Format: Currently provided in MLX format for efficient inference on Apple Silicon.
  • Quantization: Fine-tuned and quantized to 4-bit for reduced memory footprint and faster inference (using MLX DoRA).
  • Experimental V1 Release: This is an initial release to showcase the potential of GCLF training. Expect ongoing development and improvements.

Intended Use

This model is intended for experimental use and research in the following areas:

  • Exploring Symbolic AI: Investigating the capabilities of language models for structured symbolic reasoning.
  • Deductive Logic Applications: Building systems that require step-by-step, logically sound reasoning processes.
  • Glyph Code Logic Flow Development: Experimenting with and refining the GCLF framework.
  • Educational Purposes: Learning about symbolic AI, deductive reasoning, and structured knowledge representation.

Limitations:

  • Version 1 - Preview: This is an early version and may have limitations in robustness and generalization.
  • Specialized Domain: Performance is optimized for tasks related to Glyph Code Logic Flow. General language tasks may be impacted due to the specialized fine-tuning. (Further evaluation is ongoing)
  • Experimental Nature: The GCLF framework itself is under development and this model reflects an early attempt to train an LLM for it.
  • MLX Format (Initial): Currently primarily available in MLX format, which may limit accessibility for users outside the Apple Silicon/MLX ecosystem (GGUF quantization is in progress).

Training Data and Process

  • Base Model: Mistral-Small-24b
  • Fine-tuning Method: MLX-DoRA (Decomposed Relative Attention) at 4-bit quantization.
  • Training Hardware: Apple M2 (128GB RAM)
  • Training Dataset: Custom dataset of approximately 4500 examples specifically designed for Glyph Code Logic Flow. Each example was around 30,000 tokens in length, focused on detailed system instructions and GCLF tasks.
  • Training Tokens: Approximately 27 million tokens from the custom GCLF dataset.
  • Training Duration: 7 days (continuous 24/7 training).
  • Initial Experiments: Initial training attempts were made with Deepeek R1-Qwen-14 and QWQ-32, but Mistral-Small-24b was found to be more receptive to the GCLF framework due to potentially less conflicting pre-trained reasoning biases.

How to Use

System Instructions

!! GGUF Quantization (Coming Soon) !!

Version 2 and Future Development

Version 2 (In Development):

GRPO (Gradient Ratio Policy Optimization): Utilizing GRPO for potentially more stable and effective fine-tuning.

Newer Dataset: Training on an expanded and refined dataset for Glyph Code Logic Flow.

GGUF Release: Aiming for a GGUF release for wider accessibility and compatibility.

Improved Documentation: Comprehensive documentation and examples for using the model and understanding GCLF.

Ongoing Efforts:

Refining GCLF Framework: Continuously developing and improving the Glyph Code Logic Flow framework itself.

Performance Evaluation: Conducting thorough evaluations of the model's performance on GCLF tasks and general language understanding.

Community Feedback: Seeking feedback from the community to guide further development and improvements.


Known Issues

The custom dataset and heavy use of symbols and operators seems to have potentially altered the models tool use. I've found that it often want to use it's [TOOL_CALLS] function at the end of it's response (sometimes also calling out <SPECIAL_#> tokens at the end). I think I know where this is stemming from, so hopefully v2 can avoid this potential issue altogether.

If you are seeing the [TOOL_CALLS] and <SPECIAL_> outputs, you can set them as the EOS and it will align the model back into a more fluid conversation.