Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Running custom Encoder Decoder model #2491

Open
AvivSham opened this issue Nov 24, 2024 · 1 comment
Open

[Question] Running custom Encoder Decoder model #2491

AvivSham opened this issue Nov 24, 2024 · 1 comment
Labels
question Further information is requested triaged Issue has been triaged by maintainers

Comments

@AvivSham
Copy link

Hi All,
Thank you for your amazing work.
We have an encoder decoder model we want to run using TensorRT-LLM. We made an architectural modification by pooling the encoder's output dim using stacked MLP layers.
What is the recommended way of modifying the code to support the new architecture? We assume that we need to change the code to convert the model (to a static computation graph) and run it.

Please advice,

@hello-11 hello-11 added question Further information is requested triaged Issue has been triaged by maintainers labels Nov 25, 2024
@hello-11
Copy link
Collaborator

@AvivSham you can follow this guide.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested triaged Issue has been triaged by maintainers
Projects
None yet
Development

No branches or pull requests

2 participants