Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Using t5-large in t5 notebook, the translation result is invalid #135

Open
brevity2021 opened this issue Sep 8, 2022 · 1 comment
Open

Comments

@brevity2021
Copy link

Hi,

First, thank you for the great work!
I was playing with the t5 notebook in demo/generative-model. I build a docker image through Makefile, and run the notebook from the container.

I changed little things from the notebook, only a few printings. When I ran using t5-small it runs fine. But when I switched to use t5-large, the translation result in the Benchmark section becomes empty. I further printed out the tokens generated, and the results are

text generated by ONNX:
Onnx tokens:
tensor([0, 2, 0, 1], device='cuda:0')

which is obviously not correct.

I attach the notebook here for your reference. I doubt if there are any instability issues when you convert to fp16, since that method depends on the randomly generated data.

My experiment was running on a g5.2xlarge instance.

@ayoub-louati
Copy link
Contributor

@brevity2021 we are working on adding support for t5 conversion using the convert script. I think it should cover the precision for different T5 models (including t5-large).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants