File size: 3,815 Bytes
0b62fc4
 
 
 
 
 
75ae202
0b62fc4
d974f51
6f1ca36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f51b04a
6f1ca36
 
 
 
 
 
 
 
 
 
0b62fc4
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
tags:
- phi4-multimodal
- CUDA
- ONNX
- onnxruntime-genai
license: bsd-3-clause
---
![Phinx](media/phinx.png) [![Chat on Discord](https://img.shields.io/discord/754884471324672040?style=for-the-badge)](https://discord.gg/tPWjMwK) [![Follow on Bluesky](https://img.shields.io/badge/Bluesky-tinyBigGAMES-blue?style=for-the-badge&logo=bluesky)](https://bsky.app/profile/tinybiggames.com) [![Reddit](https://img.shields.io/badge/Reddit-Phinx-red?style=for-the-badge&logo=reddit)](https://www.reddit.com/r/Phinx/) [![GitHub](https://img.shields.io/badge/GitHub-Phinx-black?style=for-the-badge&logo=github)](https://github.com/tinyBigGAMES/Phinx)


## A High-Performance AI Inference Library for ONNX and Phi-4

**Phinx** is an advanced AI inference library designed to leverage **ONNX Runtime GenAI** and the **Phi-4 Multimodal ONNX** model for fast,  efficient, and scalable AI applications. Built for developers seeking seamless integration of generative and multimodal AI into their projects, Phinx provides an optimized and flexible runtime environment with robust performance.
### Key Features
- **ONNX-Powered Inference** – Efficient execution of Phi-4 models using ONNX Runtime GenAI.
- **Multimodal AI** – Supports text, image, and multi-input inference for diverse AI tasks.
- **Optimized Performance** – Accelerated inference leveraging ONNX optimizations for speed and efficiency.
- **Developer-Friendly API** – Simple yet powerful APIs for quick integration into Delphi, Python, and other platforms.
- **Self-Contained & Virtualized** – The `Phinx.model` file acts as a **virtual folder** inside the application, bundling the **Phi-4 ONNX model files** and all required dependencies into a single, easily distributable format. 
Phinx is ideal for AI research,  creative applications, and production-ready generative AI solutions. Whether you’re building chatbots,  AI-powered content generation, or multimodal assistants, Phinx delivers the speed and flexibility you need!

### **Phinx Model File Format (`Phinx.model`)**  

The **Phinx.model** format is a specialized file format designed for storing **ONNX-based machine learning models**, optimized for **CUDA-powered inference**. It provides a **structured, efficient, and extensible** way to encapsulate all required components for seamless model execution.

### **Key Features**  

1. **Self-Contained & Virtualized**  
   - The `Phinx.model` file serves as a **virtual folder** inside the application, encapsulating all necessary components for model execution.  
   - It includes the **Phi-4 ONNX model files**, along with all required dependencies, ensuring portability and ease of deployment.

2. **Optimized for CUDA Inference**  
   - Designed to leverage **GPU acceleration**, making it ideal for high-performance AI applications.  
   - Ensures fast loading and execution with CUDA-optimized computations.

3. **Structured & Extensible**  
   - Stores **model weights, metadata, configuration parameters, and dependencies** in a well-organized manner.  
   - Future-proof design allows for extensibility, supporting additional configurations or optimizations as needed.

4. **Simplified Deployment**  
   - By consolidating all required files into a single **`.model`** file, it simplifies model distribution and integration into AI applications.  
   - Eliminates external dependency management, ensuring **plug-and-play** usability.

---
**🚧 Note:** This model is designed to work with **Phinx GenAI library**. You can download the last version from [Github](https://github.com/tinyBigGAMES/Phinx).🚀

---

Phinx – Powering AI with Phi4, ONNX & CUDA, Seamless, Efficient, and Built for Performance! ⚡

<p align="center">
<img src="media/delphi.png" alt="Delphi">
</p>
<h5 align="center">
  
Made with ❤️ in Delphi