Spaces:
Runtime error
Runtime error
Update crew.py
Browse files
crew.py
CHANGED
|
@@ -41,8 +41,8 @@ class NewsletterGenCrew:
|
|
| 41 |
tasks_config = "tasks.yaml"
|
| 42 |
|
| 43 |
def llm(self):
|
| 44 |
-
llm = ChatAnthropic(model_name="claude-3-sonnet-20240229", max_tokens=4096) # ORIGINAL
|
| 45 |
-
#llm = ChatAnthropic(model_name="claude-3-sonnet-20240229",
|
| 46 |
# # max_tokens=4096,
|
| 47 |
# cache=True,
|
| 48 |
# api_key="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"
|
|
@@ -52,6 +52,8 @@ class NewsletterGenCrew:
|
|
| 52 |
# jb_anthropic_key_2_13-06-2024:
|
| 53 |
# ANTHROPIC_API_KEY=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
|
| 54 |
# https://console.anthropic.com/settings/usage
|
|
|
|
|
|
|
| 55 |
|
| 56 |
# llm = ChatGroq(model="llama3-70b-8192")
|
| 57 |
# https://console.groq.com/docs/rate-limits
|
|
@@ -64,7 +66,7 @@ class NewsletterGenCrew:
|
|
| 64 |
# supports many more optional parameters. Hover on your `ChatOllama(...)`
|
| 65 |
# class to view the latest available supported parameters
|
| 66 |
# llm = ChatOllama(model="llama3")
|
| 67 |
-
|
| 68 |
# check if ollama is running and which LLMs can then be used, run this in Anaconda cmd admin window:
|
| 69 |
# ollama list
|
| 70 |
# OUTPUT EXAMPLE:
|
|
|
|
| 41 |
tasks_config = "tasks.yaml"
|
| 42 |
|
| 43 |
def llm(self):
|
| 44 |
+
# llm = ChatAnthropic(model_name="claude-3-sonnet-20240229", max_tokens=4096) # ORIGINAL
|
| 45 |
+
# llm = ChatAnthropic(model_name="claude-3-sonnet-20240229",
|
| 46 |
# # max_tokens=4096,
|
| 47 |
# cache=True,
|
| 48 |
# api_key="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"
|
|
|
|
| 52 |
# jb_anthropic_key_2_13-06-2024:
|
| 53 |
# ANTHROPIC_API_KEY=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
|
| 54 |
# https://console.anthropic.com/settings/usage
|
| 55 |
+
#
|
| 56 |
+
# BadRequestError: Error code: 400 - {'type': 'error', 'error': {'type': 'invalid_request_error', 'message': 'Your credit balance is too low to access the Claude API. Please go to Plans & Billing to upgrade or purchase credits.'}}
|
| 57 |
|
| 58 |
# llm = ChatGroq(model="llama3-70b-8192")
|
| 59 |
# https://console.groq.com/docs/rate-limits
|
|
|
|
| 66 |
# supports many more optional parameters. Hover on your `ChatOllama(...)`
|
| 67 |
# class to view the latest available supported parameters
|
| 68 |
# llm = ChatOllama(model="llama3")
|
| 69 |
+
llm = ChatOllama(model="mistral:latest")
|
| 70 |
# check if ollama is running and which LLMs can then be used, run this in Anaconda cmd admin window:
|
| 71 |
# ollama list
|
| 72 |
# OUTPUT EXAMPLE:
|