-
Notifications
You must be signed in to change notification settings - Fork 20
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] Allow using a tag to exclude a document from processing and auto-add the tag to the document when processing fails #77
Comments
In a comment you wrote you increased it from 10000 to 40000. I don't know what to do but shorten the content. What is your suggestion? |
I just looked into the ollama documentation. You have to configure the context sizes there. It doesn't matter what is passed via the api as parameter. |
I disagree, i'm editing it in the JS file i attached a screenshot to and am seeing changes in the context window sizes without any changes on the ollama side. I'd attempted to change it on ollama but it continued to be set to what is set in the above JS file |
the way i'm editing it is via docker exec -it paperless-ai /bin/bash, then apt update && apt install vim -y, then editing the config file inside of the container itself, this doesn't persist between container rebuilds, but it works if i just stop and start the container. |
What if you increase the context size to 128k? That's the maximum llama3.2 can handle. Would be interesting to see what happens and if the doc processes. |
the doc is so large that i actively run out of VRAM already with it set this high on those documents, ollama isn't apparently smart enough to keep itself from running out of memory. I also had to shut down stable diffusion to give ollama the full 12GB of VRAM i've got. |
Hmmm okey. I will remove the 10k ctx value. But there will be no future solution to process these files if the are so big. I will add them to processed to not retry them later, after failure. |
I don't mind the value being there, but it would be good to have it be configurable rather than hard-coded, just like the ollama URL is. just another field there that can set the paramter, and have it with a sane default (like the 10000) |
For OpenAI it's quite simple as I can use tiktoken library to truncate to the maximum token size. I will integrate your thoughts as a feature next release. Probally the best solution. |
I have multiple documents that due to context tokens i'm unable to process, and it is repeatedly crashing ollama when attempted, this returns a 500 error to paperless-ai but paperless ai retrys the document on the next run. it would be good to have a way to note documents that have attempted processing but failed and not re-attempt them automatically without the tag being removed to have them back in the queue
The text was updated successfully, but these errors were encountered: