Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it normal that A10 inference speed is lower than 2080ti? #523

Open
qinbo23 opened this issue Nov 7, 2023 · 1 comment
Open

Is it normal that A10 inference speed is lower than 2080ti? #523

qinbo23 opened this issue Nov 7, 2023 · 1 comment

Comments

@qinbo23
Copy link

qinbo23 commented Nov 7, 2023

hello?I tested the Transformer-base inference speed on different devices. It's weird that A10 speed is lower than 2080ti speed.

MODEL: Transformer-base
DATA: fp16
SPEED: (number of src characters / second)
3090 7.5k/s
2080 4.5k/s
A10 2.0K/s

@Mr-wang2016
Copy link

me too.
can anyone help?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants