Update app.py
Browse files
app.py
CHANGED
@@ -164,6 +164,19 @@ if query:
|
|
164 |
|
165 |
relevant_response = pick_relevant_context_chain.invoke({"relevancy_response":response_crisis['relevancy_response']})
|
166 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
167 |
|
168 |
|
169 |
|
@@ -176,6 +189,9 @@ if query:
|
|
176 |
|
177 |
st.subheader("response_crisis['relevancy_response']")
|
178 |
st.json((response_crisis['relevancy_response']))
|
|
|
|
|
|
|
179 |
|
180 |
st.markdown("### Context Relevancy Evaluation")
|
181 |
st.json(response_crisis["relevancy_response"])
|
|
|
164 |
|
165 |
relevant_response = pick_relevant_context_chain.invoke({"relevancy_response":response_crisis['relevancy_response']})
|
166 |
|
167 |
+
relevant_contexts_chain = LLMChain(llm=llm_judge, prompt=context_prompt, output_key="relevant_contexts")
|
168 |
+
|
169 |
+
contexts = relevant_contexts_chain.invoke({"context_number":relevant_response['context_number'],"context":context})
|
170 |
+
|
171 |
+
#temp
|
172 |
+
st.subheader("Relevant Contexts")
|
173 |
+
st.json(contexts['relevant_contexts'])
|
174 |
+
|
175 |
+
response_chain = LLMChain(llm=rag_llm,prompt=final_prompt,output_key="final_response")
|
176 |
+
|
177 |
+
st.subheader("Response Chain")
|
178 |
+
st.json(response_chain)
|
179 |
+
|
180 |
|
181 |
|
182 |
|
|
|
189 |
|
190 |
st.subheader("response_crisis['relevancy_response']")
|
191 |
st.json((response_crisis['relevancy_response']))
|
192 |
+
|
193 |
+
|
194 |
+
print(relevant_response['context_number'])
|
195 |
|
196 |
st.markdown("### Context Relevancy Evaluation")
|
197 |
st.json(response_crisis["relevancy_response"])
|