Skip to content
This repository has been archived by the owner on Jun 7, 2023. It is now read-only.

JRNN batch size to speed up inference? #69

Open
max-markov opened this issue May 2, 2022 · 0 comments
Open

JRNN batch size to speed up inference? #69

max-markov opened this issue May 2, 2022 · 0 comments

Comments

@max-markov
Copy link

It seems like the out-of-the box version provided for JRNN (building on the original published model) is very slow, and inference for a single sentence takes quite a while, even when using a GPU (around 1 minute).

Increasing the batch size doesn't help since the inference code is buggy and only supports single samples. Changing the dimensionality in the graph definition file (graph-2016-09-10.pbtxt) leads to dimensionality mismatches downstream.

Has anyone been able to make inference faster for this model?

Thank you :)

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant