README.md

    Introduction

    This repository holds NVIDIA-maintained utilities to streamline mixed precision and distributed training in Pytorch. Some of the code here will be included in upstream Pytorch eventually. The intention of Apex is to make up-to-date utilities available to users as quickly as possible.

    Full API Documentation: https://nvidia.github.io/apex

    GTC 2019 and Pytorch DevCon 2019 Slides

    Contents

    1. Amp: Automatic Mixed Precision

    apex.amp is a tool to enable mixed precision training by changing only 3 lines of your script. Users can easily experiment with different pure and mixed precision training modes by supplying different flags to amp.initialize.

    Webinar introducing Amp (The flag cast_batchnorm has been renamed to keep_batchnorm_fp32).

    API Documentation

    Comprehensive Imagenet example

    DCGAN example coming soon...

    Moving to the new Amp API (for users of the deprecated "Amp" and "FP16_Optimizer" APIs)

    2. Distributed Training

    apex.parallel.DistributedDataParallel is a module wrapper, similar to torch.nn.parallel.DistributedDataParallel. It enables convenient multiprocess distributed training, optimized for NVIDIA's NCCL communication library.

    API Documentation

    Python Source

    Example/Walkthrough

    The Imagenet example shows use of apex.parallel.DistributedDataParallel along with apex.amp.

    Synchronized Batch Normalization

    apex.parallel.SyncBatchNorm extends torch.nn.modules.batchnorm._BatchNorm to support synchronized BN. It allreduces stats across processes during multiprocess (DistributedDataParallel) training. Synchronous BN has been used in cases where only a small local minibatch can fit on each GPU. Allreduced stats increase the effective batch size for the BN layer to the global batch size across all processes (which, technically, is the correct formulation). Synchronous BN has been observed to improve converged accuracy in some of our research models.

    Checkpointing

    To properly save and load your amp training, we introduce the amp.state_dict(), which contains all loss_scalers and their corresponding unskipped steps, as well as amp.load_state_dict() to restore these attributes.

    In order to get bitwise accuracy, we recommend the following workflow:

    # Initialization
    opt_level = 'O1'
    model, optimizer = amp.initialize(model, optimizer, opt_level=opt_level)
    
    # Train your model
    ...
    with amp.scale_loss(loss, optimizer) as scaled_loss:
        scaled_loss.backward()
    ...
    
    # Save checkpoint
    checkpoint = {
        'model': model.state_dict(),
        'optimizer': optimizer.state_dict(),
        'amp': amp.state_dict()
    }
    torch.save(checkpoint, 'amp_checkpoint.pt')
    ...
    
    # Restore
    model = ...
    optimizer = ...
    checkpoint = torch.load('amp_checkpoint.pt')
    
    model, optimizer = amp.initialize(model, optimizer, opt_level=opt_level)
    model.load_state_dict(checkpoint['model'])
    optimizer.load_state_dict(checkpoint['optimizer'])
    amp.load_state_dict(checkpoint['amp'])
    
    # Continue training
    ...

    Note that we recommend restoring the model using the same opt_level. Also note that we recommend calling the load_state_dict methods after amp.initialize.

    Requirements

    Python 3

    CUDA 9 or newer

    PyTorch 0.4 or newer. The CUDA and C++ extensions require pytorch 1.0 or newer.

    We recommend the latest stable release, obtainable from https://pytorch.org/. We also test against the latest master branch, obtainable from https://github.com/pytorch/pytorch.

    It's often convenient to use Apex in Docker containers. Compatible options include:

    • NVIDIA Pytorch containers from NGC, which come with Apex preinstalled. To use the latest Amp API, you may need to pip uninstall apex then reinstall Apex using the Quick Start commands below.
    • official Pytorch -devel Dockerfiles, e.g. docker pull pytorch/pytorch:nightly-devel-cuda10.0-cudnn7, in which you can install Apex using the Quick Start commands.

    See the Docker example folder for details.

    Quick Start

    Linux

    For performance and full functionality, we recommend installing Apex with CUDA and C++ extensions via

    git clone https://github.com/NVIDIA/apex
    cd apex
    pip install -v --disable-pip-version-check --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" ./

    Apex also supports a Python-only build (required with Pytorch 0.4) via

    pip install -v --disable-pip-version-check --no-cache-dir ./

    A Python-only build omits:

    • Fused kernels required to use apex.optimizers.FusedAdam.
    • Fused kernels required to use apex.normalization.FusedLayerNorm.
    • Fused kernels that improve the performance and numerical stability of apex.parallel.SyncBatchNorm.
    • Fused kernels that improve the performance of apex.parallel.DistributedDataParallel and apex.amp. DistributedDataParallel, amp, and SyncBatchNorm will still be usable, but they may be slower.

    Pyprof support has been moved to its own dedicated repository. The codebase is deprecated in Apex and will be removed soon.

    Windows support

    Windows support is experimental, and Linux is recommended. pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" . may work if you were able to build Pytorch from source on your system. pip install -v --no-cache-dir . (without CUDA/C++ extensions) is more likely to work. If you installed Pytorch in a Conda environment, make sure to install Apex in that same environment.

    项目简介

    🚀 Github 镜像仓库 🚀

    源项目地址

    https://github.com/nvidia/apex

    发行版本

    当前项目没有发行版本

    贡献者 87

    全部贡献者

    开发语言

    • Python 43.3 %
    • C++ 31.3 %
    • Cuda 25.0 %
    • Shell 0.3 %
    • C 0.2 %