Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -111,7 +111,7 @@ from huggingface_hub import InferenceClient
|
|
| 111 |
def llm_chat_response(text):
|
| 112 |
HF_TOKEN = os.getenv("HF_TOKEN")
|
| 113 |
client = InferenceClient(
|
| 114 |
-
provider="
|
| 115 |
api_key=HF_TOKEN,
|
| 116 |
)
|
| 117 |
|
|
@@ -127,7 +127,7 @@ def llm_chat_response(text):
|
|
| 127 |
}]
|
| 128 |
|
| 129 |
response_from_llama = client.chat.completions.create(
|
| 130 |
-
model="meta-llama/Llama-
|
| 131 |
messages=messages,
|
| 132 |
max_tokens=500,
|
| 133 |
)
|
|
|
|
| 111 |
def llm_chat_response(text):
|
| 112 |
HF_TOKEN = os.getenv("HF_TOKEN")
|
| 113 |
client = InferenceClient(
|
| 114 |
+
provider="cerebras", # Use the provider that supports conversational image-text tasks.
|
| 115 |
api_key=HF_TOKEN,
|
| 116 |
)
|
| 117 |
|
|
|
|
| 127 |
}]
|
| 128 |
|
| 129 |
response_from_llama = client.chat.completions.create(
|
| 130 |
+
model="meta-llama/Llama-4-Scout-17B-16E-Instruct",
|
| 131 |
messages=messages,
|
| 132 |
max_tokens=500,
|
| 133 |
)
|