Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -109,9 +109,12 @@ if generate_button:
|
|
| 109 |
# Generate response using GPT-4 with detailed prompt and debugging logs
|
| 110 |
response = generate_response([{"role": "assistant", "content": question_generation_prompt}, {"role": "user", "content": detailed_prompt}])
|
| 111 |
|
| 112 |
-
# Store generated question in session state for persistence in sidebar
|
| 113 |
st.session_state.generated_question = response
|
| 114 |
|
|
|
|
|
|
|
|
|
|
| 115 |
# Enable follow-up mode after generating the initial question
|
| 116 |
st.session_state.follow_up_mode = True
|
| 117 |
|
|
@@ -129,7 +132,8 @@ if st.session_state.follow_up_mode:
|
|
| 129 |
|
| 130 |
st.session_state.messages.append({"role": "user", "content": user_input})
|
| 131 |
|
| 132 |
-
# Generate assistant's response based on follow-up input using technical_interviewer_prompt as system prompt
|
|
|
|
| 133 |
assistant_response = generate_response(
|
| 134 |
[{"role": "assistant", "content": technical_interviewer_prompt}] + st.session_state.messages
|
| 135 |
)
|
|
|
|
| 109 |
# Generate response using GPT-4 with detailed prompt and debugging logs
|
| 110 |
response = generate_response([{"role": "assistant", "content": question_generation_prompt}, {"role": "user", "content": detailed_prompt}])
|
| 111 |
|
| 112 |
+
# Store generated question in session state for persistence in sidebar and follow-up conversation state
|
| 113 |
st.session_state.generated_question = response
|
| 114 |
|
| 115 |
+
# Add the generated question to the conversation history as an assistant message (to make it part of follow-up conversations)
|
| 116 |
+
st.session_state.messages.append({"role": "assistant", "content": response})
|
| 117 |
+
|
| 118 |
# Enable follow-up mode after generating the initial question
|
| 119 |
st.session_state.follow_up_mode = True
|
| 120 |
|
|
|
|
| 132 |
|
| 133 |
st.session_state.messages.append({"role": "user", "content": user_input})
|
| 134 |
|
| 135 |
+
# Generate assistant's response based on follow-up input using technical_interviewer_prompt as system prompt,
|
| 136 |
+
# including the generated question in context.
|
| 137 |
assistant_response = generate_response(
|
| 138 |
[{"role": "assistant", "content": technical_interviewer_prompt}] + st.session_state.messages
|
| 139 |
)
|