You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Recently Intel's Habana Labs announced AWS nodes for their Gaudi Accelerator. As described in the reddit post there seems to be a lot of potential for model parallel solution that require fast interconnect between processors.
Looking at the pytorch migration guide there is room for lightning to assist in making this process seamless, helping users to not have to worry about this details (especially around the KCRS -> RSCK format for convolutions).
I suggest we introduce a new accelerator type such that we can enable this in lightning through the accelerator flag, i.e:
trainer=Trainer(accelerator='hpu')
HPU is what the accelerator seems to be called within the migration guide.
Uh oh!
There was an error while loading. Please reload this page.
🚀 Feature
Recently Intel's Habana Labs announced AWS nodes for their Gaudi Accelerator. As described in the reddit post there seems to be a lot of potential for model parallel solution that require fast interconnect between processors.
Looking at the pytorch migration guide there is room for lightning to assist in making this process seamless, helping users to not have to worry about this details (especially around the KCRS -> RSCK format for convolutions).
I suggest we introduce a new accelerator type such that we can enable this in lightning through the accelerator flag, i.e:
HPU is what the accelerator seems to be called within the migration guide.
cc @Borda
The text was updated successfully, but these errors were encountered: