UniversalVocoding icon indicating copy to clipboard operation
UniversalVocoding copied to clipboard

Inference speed comparison

Open sujeendran opened this issue 6 years ago • 1 comments

Hi! Could you share some details about the inference speed compared to Griffin-Lim/WaveNet/WaveRNN?

sujeendran avatar Jul 26 '19 09:07 sujeendran

Hi @sujeendran,

Sorry about the delay. With a GeForce GTX 1080 Ti I'm getting around 3700 samples a second (so almost 0.25x real-time with 16kHz audio) and on an Intel Core i7-8700K CPU @ 3.70GHz I'm getting about 1700 samples a second.

So definitely faster than vanilla WaveNet but not quite real-time. However, since there is only a single forward GRU on the autoregressive path I'm sure you could get better than real-time with some engineering e.g. sparsifying the GRU, designing a custom cuda kernel, etc.

bshall avatar Jul 29 '19 06:07 bshall