Below you will find pages that utilize the taxonomy term “NVIDIA”
Post
BERT on AMD RadeonGPU
Introduction It is often asked by natural language processors on AMD Radeon GPU whether BERT works, so I will introduce the story when I verified it.
Results Confirming Installation Official Documentation: https://github.com/google-research/bert
It seems that it should be possible to move from TensorFlow 1.11 to 1.12.0. As ROC m2.3 update has come, I will set up on an experimental machine and run it on Radeon VII.
Originally, on the TensorFlow 1.
Post
How to setup Caffe on AMD Radeon GPU
Introduction Currently, PyTorch / Caffe2 base is at the cutting edge, but some of the sources on github are still based on old Caffe versions. We will look at how to set up hipCaffe (ROCm-Caffe) based on Ubuntu 16.04 + ROCm Write
Installation Requirements -Ubuntu 16.04 -Complete set of basic packages required for build -ROCm driver -MIOpen library (CUDA simulation layer) -OpenCV2 or OpenCV3 -hipCaffe
Here is a script that compiles with one command line.
Post
PyTorch-ROCm on AMD Radeon GPU
Introduction PyTorch supports ROC m2.1!
This is an installation guide for running on AMD Radeon GPUs.
Installation AMDGPUdriver 2.1 supports PyTorch1.x.x It was announced that PyTorch was officially supported.
https://rocm.github.io/dl.html
Deep Learning on ROCm TensorFlow: TensorFlow for ROCm – latest supported version 1.13
MIOpen: Open-source deep learning library for AMD GPUs – latest supported version 1.7.1
PyTorch: PyTorch for ROCm – latest supported version 1.0
Installation Issues(2019/03/01) The official page only describes Docker-based installation methods, and there is no documentation for installing from scratch.
Post
Improved way to install tensorflow-rocm
Introduction Summary of how TensorFlow works with AMDGPU Radeon
Installation Dependencies sudo apt update sudo apt -y install software-properties-common curl wget # for add-apt-repository Install Python 3.5.2 Python3.6/Python3.7 can be unstable with AMD GPUs so we will be using Python 3.5.2 For Ubuntu 18 since the default version is Python3.6 executing the following script will configure 3.5.2
PYTHON35=false if [[ `python3 --version` == *"3.5"* ]] ; then echo 'python3.
Post
Benchmark CIFAR10 on TensorFlow with ROCm on AMD GPUs vs CUDA9 and cuDNN7 on NVIDIA GPUs
Introduction I’m going to continue my description of the CIFAR10 benchmark, from where I left off.
Related articles Mar 7, 2018 Benchmarks on MATRIX MULTIPLICATION | A comparison between AMD Vega and NVIDIA GeForce series Mar 20, 2018 Benchmarks on MATRIX MULTIPLICATION | TitanV TensorCore (FP16=>FP32)
CIFAR10 Average examples pre second
Introduction I took the CIFAR10 dataset, which is widely used throughout the world in competitions and benchmarks, and used the public release of TensorFlow in order to measure its training speed.
Post
Benchmarks on MATRIX MULTIPLICATION | TitanV TensorCore (FP16=>FP32)
Introduction This is continued from the last article. I’ll be writing about benchmarks for the multiplications of matrices, including for the TensorCore (FP16=>FP32), which was first incorporated in the Volta.
Matrix Multiplication with TensorCore NVIDIA TitanV’s TensorCore (FP16=>FP32) and other FP32 benchmarks. I used the same setup as before: Ubuntu16.04 incorporating Python3.5+390.30, CUDA9.0, cuDNN7, and TensorFlow1.6.
https://devblogs.nvidia.com/programming-tensor-cores-cuda-9/ *It’s not really ideal to make comparisons against FP32 where they should be made against FP16, and give exaggerated descriptions the way they do on NVIDIA’s official site, but please note that this graph does make comparisons between FP32 and TensorCore (FP16).
Post
VGG-19 on Keras/PlaidML backend
PLAIDML, which is rumored to be faster than HIP-TENSORFLOW Introduction Hello!
HIP-TensorFlow is a library implemented by performing an CUDA simulation of TensorFlow, but since its execution speed is still under development or based on the old TensorFlow, there is a speed difference when compared against the latest NVIDIA + TensorFlow in the DeepLearning. Also, since it works at the same speed for RX 580 as for superior GPUs like Vega 56 and Vega 64, it is still an immature library in that it cannot demonstrate the potential of the Vega series.
Post
Benchmarks on MATRIX MULTIPLICATION | A comparison between AMD Vega and NVIDIA GeForce series
Introduction ACUBE Corp. graciously allowed us to borrow a Radeon Pro WX9100, so we have decided to make a report on the card and a record of the results here on our company blog. We would like to extend our heartfelt gratitude to ACUBE Corp. for this opportunity.
This report focuses on the Radeon Pro WX9100 card and makes comparisons with the Radeon RX560/580 and RadeonVega56/64/Frontier Edition from the same manufacturer, as well as with the GeForce series from NVIDIA.
Post
Wednesday, March 7, 2018, The World's First AMD GPU-based Cloud Instances for Deep Learning
IRVINE, CALIF. (PRWEB) MARCH 06, 2018 California startup Pegara, Inc. launched the world’s first set of deep learning instances based on GPUs from American chip maker AMD through its “GPU EATER” heterogeneous cloud computing service. Although much of today’s deep learning research is conducted using GPUs (graphics processing units) from NVIDIA, in conjunction with libraries it provides, such as CUDA and cuDNN, the revision of the company’s EULA (End User License Agreement) content for its consumer graphics drivers around December 2017 has raised strong voices of concern among researchers and developers at domestic and overseas universities and enterprises about potential termination of research projects and delays in their practical application.