Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -24,7 +24,7 @@ class MessageRequest(BaseModel):
|
|
24 |
message: str
|
25 |
language: str
|
26 |
|
27 |
-
repo_id = "meta-llama/Llama-3
|
28 |
llm_client = InferenceClient(
|
29 |
model=repo_id,
|
30 |
token=os.getenv("HF_TOKEN"),
|
@@ -60,8 +60,8 @@ templates = Jinja2Templates(directory="static")
|
|
60 |
|
61 |
# Configure Llama index settings
|
62 |
Settings.llm = HuggingFaceInferenceAPI(
|
63 |
-
model_name="meta-llama/Llama-3
|
64 |
-
tokenizer_name="meta-llama/Llama-3
|
65 |
context_window=3000,
|
66 |
token=os.getenv("HF_TOKEN"),
|
67 |
max_new_tokens=512,
|
|
|
24 |
message: str
|
25 |
language: str
|
26 |
|
27 |
+
repo_id = "meta-llama/Meta-Llama-3-8B-Instruct"
|
28 |
llm_client = InferenceClient(
|
29 |
model=repo_id,
|
30 |
token=os.getenv("HF_TOKEN"),
|
|
|
60 |
|
61 |
# Configure Llama index settings
|
62 |
Settings.llm = HuggingFaceInferenceAPI(
|
63 |
+
model_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
64 |
+
tokenizer_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
65 |
context_window=3000,
|
66 |
token=os.getenv("HF_TOKEN"),
|
67 |
max_new_tokens=512,
|