# Xinference
Xorbits Inference([Xinference](https://github.com/xorbitsai/inference)) empowers you to unleash the full potential of cutting-edge AI models. ## Install - [pip install "xinference[all]"](https://inference.readthedocs.io/en/latest/getting_started/installation.html) - [Docker](https://inference.readthedocs.io/en/latest/getting_started/using_docker_image.html) To start a local instance of Xinference, run the following command: ```bash $ xinference-local --host 0.0.0.0 --port 9997 ``` ## Launch Xinference Decide which LLM you want to deploy ([here's a list for supported LLM](https://inference.readthedocs.io/en/latest/models/builtin/)), say, **mistral**. Execute the following command to launch the model, remember to replace ${quantization} with your chosen quantization method from the options listed above: ```bash $ xinference launch -u mistral --model-name mistral-v0.1 --size-in-billions 7 --model-format pytorch --quantization ${quantization} ``` ## Use Xinference in RAGFlow - Go to 'Settings > Model Providers > Models to be added > Xinference'.
> Base URL: Enter the base URL where the Xinference service is accessible, like, `http://:9997/v1`. - Use Xinference Models.