With distributed training on cluster of Xeons, one can utilize larger memory to
- (i) train with larger batch size,
- (ii) train on full-image without tiling which can speed up time-to-train and improve accuracy.
- Fast and Accurate Training of an AI Radiologist
- Training Speech Recognition Models on HPC Infrastructure
- Large Minibatch Training on Supercomputers with Improved Accuracy and Reduced Time to Train
- Distributed Training of Generative Adversarial Networks for Fast Detector Simulation
- Densifying Assumed-Sparse Tensors
- Using deep neural network acceleration image analysis drug discovery
- Efficient neural network training on Intel Xeon-based supercomputers
- Adverse Effects of Image Tiling on Convolutional Neural Networks
- Adverse Effects Of Image Tiling For Automatic Deep Learning Glioma Segmentation In MRI
- Systematic Evaluation of Image Tiling Adverse Effects on Deep Learning Semantic Segmentation
- How DeepVariant Uses Intel’s AVX-512 Optimizations
- Federated learning in medicine:facilitating multi‑institutional collaborations without sharing patient data
- Application of a Tiled Fully Convolutional Network to Whole Image Predictions Leads to Lower Latency Inference
- Addressing the Memory Bottleneck in AI Model-Training for Healthcare
- Slidecast: Dell EMC Using Neural Networks to “Read Minds”
- Paving a New Path to AI-Driven Neuroscience