Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error message is not helpful when the input to a model exceeds the limit. #355

Open
gamer-mitsuha opened this issue Dec 2, 2024 · 1 comment
Assignees

Comments

@gamer-mitsuha
Copy link

Context: I tried to send chunked texts to an embedding model using Ollama and got the following error:

.venv/lib/python3.12/site-packages/ollama/_client.py", line 120, in _request_raw
    raise ResponseError(e.response.text, e.response.status_code) from None
ollama._types.ResponseError: {}

It was not clear what happened and I had to check the failed batch of examples and examine what happened. Eventually I was able to figure out the reason is that the chunk limit is higher than the allowed input size, and thus the server returns error for such long inputs.

I'm wondering if it's possible to produce more helpful error message in such cases?

@ParthSareen
Copy link
Contributor

Hey @gamer-mitsuha - sorry about that. Seems like the Ollama server didn't give an output back or we should have captured that. Thanks for raising - will take a look!

@ParthSareen ParthSareen self-assigned this Dec 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants