-
Notifications
You must be signed in to change notification settings - Fork 92
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multi-Utterance Mini-Batch #7
Comments
Updating from multiple utterances would likely help to smooth out the On Fri, Oct 30, 2015 at 3:05 PM, mschonwe [email protected] wrote:
|
My motivation is based on a lecture from Adam Coates describing the roofline model, and how to increase system throughput by saturating both computational and bandwidth resources. I was looking to increase the arithmetic intensity (and therefore hopefully throughput) by feeding mini-batches into costAndGrad. The profiling I did indicates that nearly all the processing time is spent in cudamat calls (in costAndGrad), so I wasn't thinking the CTC calculation was substantially rate limiting. (runsnake image below: ctc_loss is the oval at the bottom right). I think we'll be ok looping over each utterance and averaging the gradients for each mini-batch. |
Do you have a suggestion for supporting processing of mini-batches of multiple utterances at a time?
We have refactored our data to have feature files of fixed frame lengths. We can have dataLoader load in the .bin features as utterances(rows) x frame--features(columns), but it seems we would need to modify ctc_fast.pyx to loop over the utterances, and somehow combine the gradients. The loop over the utts seems easy enough, but not sure how to combine gradients.
Have you already tested multi-utterance mini-batches and decided that they are not appropriate for the task?
The text was updated successfully, but these errors were encountered: