File size: 1,225 Bytes
ad5fc6e
 
 
 
 
 
 
 
 
 
 
6c27626
37db0c7
dbbf795
 
37db0c7
 
 
 
 
dbbf795
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
---
title: Advanced Dual Prompting
emoji: πŸ†
colorFrom: green
colorTo: purple
sdk: gradio
sdk_version: 4.16.0
app_file: app.py
pinned: false
---

# Simple Dual LLM Chatbot

This is a playground for testing out Standford's 'Meta-Prompting' logic ([paper link](https://arxiv.org/abs/2401.12954)), in whcih for every user request, it first passes the request to a 'meta' bot, the 'meta' bot will then generate a system prompt of a field-related 'Expert' bot for answering user's request.  
That is, for each round, the LLM should accordingly assigns the best expert for answering user's specific request.  
Standford claimed that this simple implementation result in a 60%+ better accuracy compared to a standard 'syst_prompt + chat_history' logic.  
Hence, one can't be too curious in checking it out, here is a simple implemnetaion for everybody to play around.  

Something to keep in mind:
1. Currently it requires an api key from chatglm (get one here if you don't have one: [link](https://open.bigmodel.cn/usercenter/apikeys))
2. To balance contextual-understanding and token-saving, the meta bot's logic is modified to have access to only the last round of chat and the current user request when 'generating' an expert.