Skip to content

Inconsistent Output: Clear Results in Notebook vs. Noisy Responses in Inference Client #34

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
humblenginr opened this issue Dec 8, 2024 · 1 comment

Comments

@humblenginr
Copy link

humblenginr commented Dec 8, 2024

When using the notebook, the results are clear and seem aligned with the prompts. However, with the inference client, I mostly encounter noise or responses that don’t appear to understand the input. The output often lasts only 3–4 seconds before stopping.

I’m unsure if this is the intended behavior of the base model, requiring training for better performance, or if there might be an issue on my end, such as problems with microphone input affecting the model’s performance.

Any guidance would be greatly appreciated!

This was referenced Dec 8, 2024
@humblenginr
Copy link
Author

btw I've tried it on different GPUs (RTX 4090, A6000, H100)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant