-
Notifications
You must be signed in to change notification settings - Fork 136
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Please update tokenizers and transformers version #607
Comments
We currently offer the following
Both of these are currently only available for Python 3.8. To change the Python version of your app, see here. As you can see from its setup.py file, indic-punct pins all of its requirements to specific versions. With packages that do this, it's sometimes possible to get them working by specifying whatever is the closest version available in the Chaquopy repository:
In this case I've used the closest newer version of each requirements, but sometimes you might need to use the closest older one. |
Unfortunately, the current version of indic-punct (2.1.4) also has a native requirement which Chaquopy doesn't support at all (pynini). It's possible that one of the older versions of indic-punct doesn't have this requirement, but the release history is confusing (8 releases in one day, and no tags on GitHub), so that's something you'd have to look into by yourself. See also #608. |
We're not planning to update this package in the near future, but if you'd like to try building the new version yourself, follow the instructions here. However, our package build tool doesn't currently have working support for Rust – see #1030 for details. If anyone else needs a newer version of tokenizers, please click the thumbs up button above, and post a comment explaining why you need it. |
Hello, |
Sorry, we have no update planned in the near future. But if you'd like to try updating it yourself, see the links in my previous comment. Our current |
Looks like I also need an updated version of tokenizers package for working with manga-ocr (Requires transformers >= 4.25.0
I've added my thumbs up and might lo0ok at the instructions to install myself later if I have time. |
Thanks – I haven't checked, but you may be able to work around this by using an older version of manga-ocr. |
In my case I need version 0.13.3 because it is a requirement of faster-whisper.
However I am blocked due to the build-wheel.sh script setting I wonder if someone knows the reasoning for setting that env variable and/or the consequences of unsetting it or setting it to a different value that conforms to the usual |
I don't remember exactly why we added that variable; you can probably find out from the Git history. But going by the |
I needed a module in a more recent version of I tried building a wheel for Error log
Not sure how to proceed from here. Any help is appreciated. |
This appears to be caused by the |
@mhsmith Here's the Dockerfile
docker-compose.yaml
Other changes
I also found this comment suggesting adding the Hope you can help me solve this error. Thanks! |
Sorry, I don't have time to look into this in any detail. But |
Hey, Hope you are doing well. I am facing issues while trying to pip install anthropic which has a dependency of tokenizer>=0.13. I tried with 0.13 version, but i get the attached errors. Could you please guide as to how we can work around this issue. |
You could try using an older version of anthropic. Looking back through the blame of anthropic's pyproject.toml, the last version which didn't require such a new version of tokenizers was anthropic 0.2.10. That came out less than a year ago, but this is obviously a fast-moving package, so I don't know if that would be acceptable for you. |
Luckily, the tokenizer version 0.10.3 has worked with the latest anthropic package so far. I thought to test it regardless of the incompatibility error during build and run, and it worked. Yeah, anthropic older versions are not available to newer users as per their api docs, because of huge changes/improvements in their latest offering "opus". So far so good.. |
Please add tokenizers version 0.11.1, it is a requirement for some of the latest python modules like indic-punct.
The text was updated successfully, but these errors were encountered: