You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to use standardized metric such as images/sec to arrive at relative training speeds for different frameworks(e.g. PyTorch and TF). Have you computed such a metric, if not, what do you think of this approach.
modified run.py to compute images_per_sec rate. lines modified highlighted with "[KAL]"
If you run the code you'll see tqdm progress bar which shows average time per batch, elapsed time and approximate time to finish. Second progress bar shows same for epochs.
However you might prefer to use a distributed framework for pytorch, which is supposed to provide better performance even on a single PC
I am trying to use standardized metric such as images/sec to arrive at relative training speeds for different frameworks(e.g. PyTorch and TF). Have you computed such a metric, if not, what do you think of this approach.
modified run.py to compute images_per_sec rate. lines modified highlighted with "[KAL]"
=========================================
`
def train(model, loader, epoch, optimizer, criterion, device, dtype, batch_size, log_interval, scheduler):
The text was updated successfully, but these errors were encountered: