-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question in reproducing experimental results. #2
Comments
I am unsure whether you can directly input a sentence (without tokenization) into a model. Can you run the RULER experiments? |
One more question for you, how can I stat the concrete pruning rate used in the decoding phase 🤔? As MagicPIG uses dynamic retrieval, it is not like Quest, which uses a fixed token budget. By the way, if I got the concrete pruning rate, can I use the following formula to calculate the overall equivalent token budget 🤔?
|
I think your understanding is correct. BTW, we will release v0.2 next week. Maybe make it easier for you to evaluate. |
Thanks a lot! |
HI😊, we are reproducing your experimental results as the baseline of our method.
We are confused by the following questions. 🤔
First, why running the following code needs more than 1 minutes? Since the prompt is short, it is expected to finish very quickly.
Second, after waiting for over 1 min, we finally got the results looks like this:
We think this output is not reasonable and want to know if there are some improper configurations in the following scripts?
If there are improper configurations for short prompt generation, we want to further know what is the most suitable configuration under different prompt length? e.g. 1K, 2K, 4K, 8K.
The text was updated successfully, but these errors were encountered: