Inference time on 8vcpu/32GB Ram or T4 30GBRAM, 16GB VRAM, 8vcpu

#18
by NeevrajKB - opened

Planning on deploying on a server, first time user, so asking for guidance. What's max concurrent requests possible with the aforementioned specs with a low inference time?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment