cuDNN
About
NVIDIA cuDNN
The NVIDIA CUDA® Deep Neural Network library (cuDNN) is a GPU-accelerated library of primitives for deep neural networks. cuDNN provides highly tuned implementations for standard routines such as forward and backward convolution, pooling, normalization, and activation layers.
Deep learning researchers and framework developers worldwide rely on cuDNN for high-performance GPU acceleration. It allows them to focus on training neural networks and developing software applications rather than spending time on low-level GPU performance tuning. cuDNN accelerates widely used deep learning frameworks, including Caffe2, Chainer, Keras, MATLAB, MxNet, PyTorch, and TensorFlow. For access to NVIDIA optimized deep learning framework containers that have cuDNN integrated into frameworks, visit NVIDIA GPU CLOUD to learn more and get started.
Key Features
- Tensor Core acceleration for all popular convolutions including 2D, 3D, Grouped, Depth-wise separable, and Dilated with NHWC and NCHW inputs and outputs
- Optimized kernels for computer vision and speech models including ResNet, ResNext, SSD, MaskRCNN, Unet, VNet, BERT, GPT-2, Tacotron2 and WaveGlow
- Supports FP32, FP16, and TF32 floating point formats and INT8, and UINT8 integer formats
- Arbitrary dimension ordering, striding, and sub-regions for 4d tensors means easy integration into any neural net implementation
- Speed up fused operations on any CNN architecture
cuDNN is supported on Windows and Linux with Ampere, Turing, Volta, Pascal, Maxwell, and Kepler GPU architectures in data center and mobile GPUs.
Specification
You May Also Like
Related products
-
RAPIDS
SKU: N/AACCELERATED DATA SCIENCE The RAPIDS suite of open source software libraries gives you the freedom to execute end-to-end data science and analytics pipelines entirely on GPUs. SCALE OUT ON GPUS Seamlessly scale from GPU workstations to multi-GPU servers and multi-node clusters with Dask. PYTHON INTEGRATION Accelerate your Python data science toolchain with minimal code changes ...More Information -
DOCKER
SKU: N/ADocker is the de facto developer standard for building and sharing apps that enable simplicity, agility and choice for software development across any infrastructure so that you can get your job done and deploy your applications faster. Docker provides developer-friendly, CLI-based workflow and makes it easy to build, share, and run containerized applications. Even your most complex applications can be containerized. You can build locally, deploy to the cloud, and run anywhere. -
Caffe
SKU: N/ACaffe is a deep learning framework made with expression, speed, and modularity in mind. It is developed by Berkeley AI Research (BAIR) and by community contributors. Yangqing Jia created the project during his PhD at UC Berkeley. Caffe is released under the BSD 2-Clause license. Expressive architecture encourages application and innovation. Models and optimization are defined by configuration without ...More Information
Our Customers
























