John Stone (Research Staff, The Beckman Institute) points out that improvements in the AVX-512 instruction set in the Intel Xeon Phi (and latest generation Intel Xeon processors) can deliver significant performance improvements for some time consuming molecular visualization kernels over most existing Intel Xeon CPUs. Based on his recent results using the Intel Xeon Phi … [Read more...]
SURFsara Achieves Accuracy and Performance Breakthroughs for Both Deep Learning and Wide Network Training
Sponsored Content SURFsara posted the best accuracy and an under 40 minute training time on some popular deep learning architectures and data sets to establish new single-model state-of-the-art results using only general-purpose CPU-based hardware, as opposed to special accelerators. Specifically SURFsara reports under 40 minutes to train the ResNet50 model on the ImageNet-1k … [Read more...]
Accelerating Python and Deep Learning
Sponsored Content “For deep learning to have a meaningful impact and business value, the time to train a model must be reduced from weeks to hours,” observed Ananth Sankaranarayanan, Intel’s director of engineering, analytics and AI solutions. Demonstrating the performance benefits of Intel Xeon and Intel Xeon Phi hardware and new Intel Architecture (IA) optimized software … [Read more...]
Come learn and meet the experts at the 2016 Intel HPC Developer Conference – Just Before SC16!
Sponsored Content Come learn and meet the experts at the Intel 2016 HPC Developer Conference - just before SC16! Arrive early at Supercomputing 2016 and fuel your insight with focused technical sessions and hands-on labs brought to you by your peers, industry experts, academic/governments institutions and Intel technologists. Plus it’s free! Register now for the 2016 … [Read more...]
Up To Orders of Magnitude More Performance with Intel’s Distribution of Python
Intel has created a freely downloadable, optimized Python distribution that can greatly accelerate Python codes. Benchmarks show that two order of magnitude speedups (over 100x) can be achieved by using the Intel Distribution for Python. The Intel® Distribution for Python 2017 Beta program (product release will be in September) provides free access to this optimized version. … [Read more...]
NVIDIA – “[Intel] Should Get Their Facts Straight” on Machine Learning Benchmarks
NVIDIA responds to the machine learning benchmark results presented by Intel at ISC'16, "It’s great that Intel is now working on deep learning. This is the most important computing revolution with the era of AI upon us and deep learning is too big to ignore. But they should get their facts straight." (Source: NVIDIA) NVIDIA notes further that, "While we can correct each of … [Read more...]
Pascal upgrade and Intel Xeon Phi at the Swiss National Supercomputing Centre
It was announced at GTC 2016 that the Piz Daint system at the Swiss National Supercomputing Center (CSCS) in Lugano, Switzerland will be upgraded to 4,500 Pascal GPUs. No date was stated for when the upgrade will be completed. The upgrade is expected to more than double the performance of Piz Daint, which currently provides 7.8 PF/s of performance. The announcement provides a … [Read more...]
Code Modernization Efforts Deliver a 32.9X Speedup in STAC-A2™ Financial Industry Standard Benchmarks
Global financial industry leaders such as Citi and J.P. Morgan have acknowledged they are currently modernizing their code via collaborative efforts with the Intel Software and Solutions Group. Results reported in the recent presentation, Intel’s New STAC A2 Results and Speeding Up FX LSV Monte Carlo case study[1], demonstrate an overall 32.9X speedup in the financial industry … [Read more...]
PGI/NVIDIA To Develop LLVM-based Open-Source Fortran Compiler for NNSA and Possibly OpenPower
PGI (The Portland Group) is working with the NNSA to create an open-source Fortran compiler designed to integrate with the widely used LLVM compiler infrastructure. Recently, PGI announced comparable x86 multicore OpenACC performance as compared the the Intel compiler using OpenMP. This move ensure that PGI will have access to the latest and best information about the AVX-512 … [Read more...]
Now Online – Presentations and Videos from Houston 2015 Directives and Tools for Accelerators Workshop
The presentations and videos from the October 12 University of Houston "2015 Directives and Tools for Accelerators" workshop are now available online. As with last year's workshop, this workshop was also considered a complete success Topics covered include: Usage of directives’ that preserves a single code base, accelerates for GPUs, and offers cross-platform … [Read more...]