Replies: 1 comment
-
@angelayi Could you take a look at let us know if the above workflow is a potential solution ? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
TL;DR
We want to use AOTInductor to compile and package Torch-TensorRT compiled modules to produce shared object files which can be used in Pythonless deployment. This could be especially beneficial for Jetson platforms.
Reference: https://pytorch.org/docs/main/torch.compiler_aot_inductor.html
Goal(s)
Ensure we support AOTInductor to export Torch-TRT compiled programs using
torch._inductor.aoti_compile_and_package
.Usecases
Proposed APIs/UX
We could probably get this feature for free (atleast from a frontend perspective). Here's the workflow I have in my mind.
We have recently added the re-export feature #3262 which uses
torch.export
to create first classExportedProgram
s from Torch-TensorRT compilation. As long as these exported programs are valid and torchbind objects are supported by AOTI, I would think the above workflow would work.Limitations
Considerations
Data Structures
None at this time
Details specific for TorchScript Support
Torchscript support won't be affected.
Details specific for FX support
All this will be applicable to dynamo frontend. ir=fx isn't active so it doesn't get affected.
Implementation Phases
Prototype -
MVP
(<TARGET RELEASE VERSION>)
Extension Phase 1
(<TARGET RELEASE VERSION>)
Extension Phase 2
(<TARGET RELEASE VERSION>)
Beta Was this translation helpful? Give feedback.
All reactions