Kevin Hu commited on
Commit
ef2346f
·
1 Parent(s): 53b1865

Add more mistral models. (#3826)

Browse files

### What problem does this PR solve?

#3647

### Type of change

- [x] New Feature (non-breaking change which adds functionality)

Files changed (2) hide show
  1. agent/canvas.py +1 -0
  2. conf/llm_factories.json +34 -10
agent/canvas.py CHANGED
@@ -234,6 +234,7 @@ class Canvas(ABC):
234
  except Exception as e:
235
  yield {"content": "*Exception*: {}".format(e), "running_status": True}
236
  logging.exception("Canvas.run got exception")
 
237
 
238
  if ran >= len(self.path[-1]) and waiting:
239
  without_dependent_checking = waiting
 
234
  except Exception as e:
235
  yield {"content": "*Exception*: {}".format(e), "running_status": True}
236
  logging.exception("Canvas.run got exception")
237
+ ran += 1
238
 
239
  if ran >= len(self.path[-1]) and waiting:
240
  without_dependent_checking = waiting
conf/llm_factories.json CHANGED
@@ -618,32 +618,44 @@
618
  },
619
  {
620
  "llm_name": "open-mistral-7b",
621
- "tags": "LLM,CHAT,32k",
622
- "max_tokens": 32000,
 
 
 
 
 
 
 
 
 
 
 
 
623
  "model_type": "chat"
624
  },
625
  {
626
  "llm_name": "mistral-large-latest",
627
- "tags": "LLM,CHAT,32k",
628
- "max_tokens": 32000,
629
  "model_type": "chat"
630
  },
631
  {
632
  "llm_name": "mistral-small-latest",
633
- "tags": "LLM,CHAT,32k",
634
- "max_tokens": 32000,
635
  "model_type": "chat"
636
  },
637
  {
638
- "llm_name": "mistral-medium-latest",
639
  "tags": "LLM,CHAT,32k",
640
  "max_tokens": 32000,
641
  "model_type": "chat"
642
  },
643
  {
644
- "llm_name": "codestral-latest",
645
- "tags": "LLM,CHAT,32k",
646
- "max_tokens": 32000,
647
  "model_type": "chat"
648
  },
649
  {
@@ -651,6 +663,18 @@
651
  "tags": "LLM,CHAT,8k",
652
  "max_tokens": 8192,
653
  "model_type": "embedding"
 
 
 
 
 
 
 
 
 
 
 
 
654
  }
655
  ]
656
  },
 
618
  },
619
  {
620
  "llm_name": "open-mistral-7b",
621
+ "tags": "LLM,CHAT,128k",
622
+ "max_tokens": 128000,
623
+ "model_type": "chat"
624
+ },
625
+ {
626
+ "llm_name": "ministral-8b-latest",
627
+ "tags": "LLM,CHAT,128k",
628
+ "max_tokens": 128000,
629
+ "model_type": "chat"
630
+ },
631
+ {
632
+ "llm_name": "ministral-3b-latest",
633
+ "tags": "LLM,CHAT,128k",
634
+ "max_tokens": 128000,
635
  "model_type": "chat"
636
  },
637
  {
638
  "llm_name": "mistral-large-latest",
639
+ "tags": "LLM,CHAT,128k",
640
+ "max_tokens": 128000,
641
  "model_type": "chat"
642
  },
643
  {
644
  "llm_name": "mistral-small-latest",
645
+ "tags": "LLM,CHAT,128k",
646
+ "max_tokens": 128000,
647
  "model_type": "chat"
648
  },
649
  {
650
+ "llm_name": "codestral-latest",
651
  "tags": "LLM,CHAT,32k",
652
  "max_tokens": 32000,
653
  "model_type": "chat"
654
  },
655
  {
656
+ "llm_name": "mistral-nemo",
657
+ "tags": "LLM,CHAT,128k",
658
+ "max_tokens": 128000,
659
  "model_type": "chat"
660
  },
661
  {
 
663
  "tags": "LLM,CHAT,8k",
664
  "max_tokens": 8192,
665
  "model_type": "embedding"
666
+ },
667
+ {
668
+ "llm_name": "pixtral-large-latest",
669
+ "tags": "LLM,CHAT,32k",
670
+ "max_tokens": 32000,
671
+ "model_type": "image2text"
672
+ },
673
+ {
674
+ "llm_name": "pixtral-12b",
675
+ "tags": "LLM,CHAT,32k",
676
+ "max_tokens": 32000,
677
+ "model_type": "image2text"
678
  }
679
  ]
680
  },