Spaces:
Running
Running
# https://huggingface.co/docs/hub/en/spaces-config-reference | |
title: API Inference | |
short_description: Inference on API endpoints | |
emoji: ⚡ | |
colorFrom: blue | |
colorTo: green | |
sdk: streamlit | |
sdk_version: 1.37.1 | |
python_version: 3.11.9 | |
app_file: 0_🏠_Home.py | |
full_width: true | |
pinned: true | |
header: mini | |
license: apache-2.0 | |
models: | |
- stabilityai/stable-diffusion-xl-base-1.0 | |
- black-forest-labs/FLUX.1-schnell | |
- black-forest-labs/FLUX.1-dev | |
- codellama/CodeLlama-34b-Instruct-hf | |
- meta-llama/Llama-2-13b-chat-hf | |
- meta-llama/Meta-Llama-3.1-405B-Instruct-FP8 | |
- mistralai/Mistral-7B-Instruct-v0.2 | |
- NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | |
# ⚡ API Inference | |
[Streamlit](https://streamlit.io) app for running inference on generative AI endpoints. | |
## Configuration | |
Streamlit [configuration](https://docs.streamlit.io/develop/concepts/configuration/options) has the following priority order: | |
1. `streamlit run` flags (highest priority) | |
2. environment variables | |
3. local config file | |
4. global config file | |
So you can deploy the local config and override with environment variables as necessary: | |
```bash | |
STREAMLIT_LOGGER_LEVEL=error | |
STREAMLIT_CLIENT_SHOW_ERROR_DETAILS=false | |
STREAMLIT_CLIENT_TOOLBAR_MODE=viewer | |
STREAMLIT_SERVER_FILE_WATCHER_TYPE=none | |
STREAMLIT_SERVER_RUN_ON_SAVE=false | |
STREAMLIT_BROWSER_SERVER_ADDRESS=adamelliotfields-api-inference.hf.space | |
``` | |
## Secrets | |
```bash | |
FAL_KEY=... | |
HF_TOKEN=... | |
PERPLEXITY_API_KEY=... | |
``` | |
## Installation | |
```sh | |
# clone | |
git clone https://huggingface.co/spaces/adamelliotfields/api-inference.git | |
cd api-inference | |
git remote set-url origin https://adamelliotfields:[email protected]/spaces/adamelliotfields/api-inference | |
# install | |
python -m venv .venv | |
source .venv/bin/activate | |
pip install -r requirements.txt | |
# run | |
python app.py | |
``` | |