Safetensors
GGUF
conversational

Shiny-Phi3.5

Shiny-Phi3.5 is a reflection fine-tune of Phi3.5 using mahiatlinux's dataset.

Recently "Reflection 70B" drew a lot of attention after making claims of massive performance gains via reflection tuning. However, independent testing has been unable to reproduce these results.

I was curious to try it myself, so I made this model. If you'd like to try a smaller reflection model for yourself, or just one that's not associated with the original - then here you go!

What is reflection? Reflection fine-tuning guides the model to generate a plan, and then reflect on the plan before proceeding to the final output. A similar approach has been used by Claude: instructing the model to plan and reflect via system prompts. Reflection tuning "bakes in" the behavior.

Downloads last month
13
GGUF
Model size
3.82B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for johnsoupir/Shiny-Phi3.5

Quantized
(138)
this model

Dataset used to train johnsoupir/Shiny-Phi3.5