Skip to content
This repository has been archived by the owner on Oct 13, 2022. It is now read-only.

Multi-GPU training #76

Open
hegc opened this issue Jan 19, 2021 · 5 comments
Open

Multi-GPU training #76

hegc opened this issue Jan 19, 2021 · 5 comments

Comments

@hegc
Copy link

hegc commented Jan 19, 2021

Hi, these examples are excellent, and can we training on multi-GPUs with K2/lhotse? Just like the DDP in pytorch?

@danpovey
Copy link
Contributor

danpovey commented Jan 19, 2021 via email

@csukuangfj
Copy link
Collaborator

There is a WIP pull-request about multi-GPU training: #71

@hegc
Copy link
Author

hegc commented Jan 19, 2021

Thanks, I'll try it. @danpovey @csukuangfj

@pzelasko
Copy link
Collaborator

BTW I intend to finish that PR so that we can switch between single-GPU and multi-GPU training when I find some spare time. I will also need to make sure that it does the right thing when aggregating things like validation loss between the GPUs (I saw both processes return slightly different values) and storing/loading checkpoints. If you have more time and can pick it up before I do, you're welcome to do it :)

@danpovey
Copy link
Contributor

danpovey commented Jan 20, 2021 via email

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants