Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does 2.18.0 serving has tensorRT support? #4018

Closed
Yadan-Wei opened this issue Jan 18, 2025 · 3 comments
Closed

Does 2.18.0 serving has tensorRT support? #4018

Yadan-Wei opened this issue Jan 18, 2025 · 3 comments
Assignees
Labels
stale This label marks the issue/pr stale - to be closed automatically if no activity stat:awaiting response

Comments

@Yadan-Wei
Copy link

Yadan-Wei commented Jan 18, 2025

  • TensorFlow 2.18.0 has dropped tensorRT support, how about TensorFlow Serving?

I tried to host a tensorRT model with 2.18.0-gpu container, it fails with log below.
025-01-17 19:47:04.267837: I external/org_tensorflow/tensorflow/cc/saved_model/loader.cc:466] SavedModel load for tags { serve }; Status: fail: NOT_FOUND: Op type not registered 'CreateTRTResourceHandle' in binary running on 9b2753a90ca1. Make sure the Op and Kernel are registered in the binary running in this process. Note that if you are loading a saved graph which used ops from tf.contrib (e.g. tf.contrib.resampler), accessing should be done before importing the graph, as contrib ops are lazily registered when the module is first accessed.. Took 2505621 microseconds. 2025-01-17 19:47:04.267889: E tensorflow_serving/util/retrier.cc:40] Loading servable: {name: tftrt_saved_model version: 1} failed: NOT_FOUND: Op type not registered 'CreateTRTResourceHandle' in binary running on 9b2753a90ca1. Make sure the Op and Kernel are registered in the binary running in this process. Note that if you are loading a saved graph which used ops from tf.contrib (e.g. tf.contrib.resampler), accessing should be done before importing the graph, as contrib ops are lazily registered when the module is first accessed.

While host with 2.17.0-gpu, it is successful.

Dockerfile of both version have tensorrt relevant library and ENV.

  • If we want to build with tensorRT support, how should we do?

I've tried to set TF_NEED_TENSORRT=1, but it doesn't work.

Thanks.

@janasangeetha janasangeetha self-assigned this Feb 24, 2025
@malla456 malla456 assigned malla456 and unassigned janasangeetha Mar 25, 2025
@malla456
Copy link

Hi @Yadan-Wei ,

Apologies for the delay. As stated in the official documentation, TensorRT support has been removed from CUDA builds for the purpose of improving code health. Consequently, TensorFlow 2.18 does not support TensorRT. To serve a TensorRT model with TensorFlow Serving, you will need to downgrade to TensorFlow 2.17.0.

Thank you.

Copy link

github-actions bot commented Apr 2, 2025

This issue has been marked stale because it has no recent activity since 7 days. It will be closed if no further activity occurs. Thank you.

@github-actions github-actions bot added the stale This label marks the issue/pr stale - to be closed automatically if no activity label Apr 2, 2025
Copy link

This issue was closed due to lack of activity after being marked stale for past 7 days.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
stale This label marks the issue/pr stale - to be closed automatically if no activity stat:awaiting response
Projects
None yet
Development

No branches or pull requests

3 participants