-
Notifications
You must be signed in to change notification settings - Fork 560
Use new prompt for QuestionAnswerTool #645
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Coverage Report •
|
Get review from Project Wednesday before merging |
|
24ce9d5
to
8aaf006
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
bc8d5eb
to
3a7f1e4
Compare
ae6b9bf
to
dbe6c36
Compare
dbe6c36
to
a39dbd5
Compare
- **You cannot list the citation at the end of response. | ||
- Every claim statement you generated must have at least one citation.**""", | ||
"answering_user_prompt": """## Retrieved Documents | ||
{documents} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I understand that the prompt has been copied from Azure OpenAI On Your Data, but this is a very large prompt, and we've been told by data scientists in the past that long prompts don't work very well since the LLM tends to only "remember" the last few sentences and forget what was said to it in the beginning (attention bias, I think it is called). Should we get this reviewed by a data scientist?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My concern is that we don't have a standard process for DS evaluation - what if we make the performance worse.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I agree that we should get as many reviews as possible, do you know who could review this for us?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We could check with Malvina/Eran if they can take a look at this. It doesn't have to be a blocker for this PR though, since users are allowed to change the prompt if they want to. Can totally be done as a follow-up.
config = json.loads(config_file) | ||
|
||
# These properties may not exist in the config file as they are newer | ||
config["prompts"]["answering_system_prompt"] = config["prompts"].get( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am planning to add tests to cover these scenarios once the core of the PR has been reviewed
Will create new pull request, due to vast number of changes since opening |
Required by #322 , closes #648
Purpose
main_prompt
used in On Your Data.The new prompt + few-shot example are configurable in the Admin app. There is JSON schema validation on the retrieved documents to warn against uploading invalid JSON.
Does this introduce a breaking change?
Pull Request Type
What kind of change does this Pull Request introduce?
How to Test
What to Check