Tuesday, January 29, 2019

Accelerating AI and Deep Learning with Dell EMC Isilon and NVIDIA GPUs - Dell EMC Certifications


Over the last few years, Dell EMC and NVIDIA have established a strong partnership to help organizations accelerate their AI initiatives. For organizations that prefer to build their own solution, we offer Dell EMC’s ultra-dense PowerEdge C-series, with NVIDIA’s TESLA V100 Tensor Core GPUs, which allows scale-out AI solutions from four up to hundreds of GPUs per cluster. For customers looking to leverage a pre-validated hardware and software stack for their Deep Learning initiatives, we offer Dell EMC Ready Solutions for AI: Deep Learning with NVIDIA, which also feature Dell EMC Isilon All-Flash storage.  Our partnership is built on the philosophy of offering flexibility and informed choice across a broad portfolio.

To give organizations even more flexibility in how they deploy AI with breakthrough performance for large-scale deep learning Dell EMC and NVIDIA have recently collaborated on a new reference architecture that combines the Dell EMC Isilon All-Flash scale-out NAS storage with NVIDIA DGX-1 servers for AI and deep learning (DL) workloads.

To validate the new reference architecture, we ran multiple industry-standard image classification benchmarks using 22 TB datasets to simulate real-world training and inference workloads. This testing was done on systems ranging from one DGX-1 server, all the way to nine DGX-1 servers (72 Tesla V100 GPUs) connected to eight Isilon F800 nodes.

This blog post summarizes the DL workflow, the training pipeline, the benchmark methodology, and finally the results of the benchmarks.

Key components of the reference architecture shown in figure 1 include:

  • Dell EMC Isilon All-Flash scale-out NAS storage delivers the scale (up to 33 PB), performance (up to 540 GB/s), and concurrency (up to millions of connections) to eliminate the storage I/O bottleneck keeping the most data hungry compute layers fed to accelerate AI workloads at scale.
  • NVIDIA DGX-1 servers which integrate up to eight NVIDIA Tesla V100 Tensor Core GPUs fully interconnected in a hybrid cube-mesh topology. Each DGX-1 server can deliver 1 petaFLOPS of AI performance, and is powered by the DGX software stack which includes NVIDIA-optimized versions of the most popular deep learning frameworks, for maximized training performance.


Benchmark Methodology Summary


In order to measure the performance of the solution, various benchmarks from the TensorFlow Benchmarks repository were carefully executed. This suite of benchmarks performs training of an image classification convolutional neural network (CNN) on labeled images. Essentially, the system learns whether an image contains a cat, dog, car, train, etc.

The well-known ILSVRC2012 image dataset (often referred to as ImageNet) was used. This dataset contains 1,281,167 training images in 144.8 GB[1]. All images are grouped into 1000 categories or classes. This dataset is commonly used by deep learning researchers for benchmarking and comparison studies.

When running the benchmarks on the 148 GB dataset, it was found that the storage I/O throughput gradually decreased and became virtually zero after a few minutes. This indicated that the entire dataset was cached in the Linux buffer cache on each DGX-1 server. Of course, this is not surprising since each DGX-1 server has 512 GB of RAM and this workload did not significantly use RAM for other purposes. As real datasets are often significantly larger than this, we wanted to determine the performance with datasets that are not only larger than the DGX-1 server RAM, but larger than the 2 TB of coherent shared cache available across the 8-node Isilon cluster. To accomplish this, we simply made 150 exact copies of each image archive file, creating a 22.2 TB dataset.

Conclusion


Here are some of the key findings from our testing of the Isilon and NVIDIA DGX-1 server reference architecture:

Achieved compelling performance results across industry standard AI benchmarks from eight through 72 GPUs without degradation to throughput or performance.
Linear scalability from 8-72 GPUs delivering up to 19.9 GB/s while keeping the GPUs pegged at >97% utilization.
The Isilon F800 system can deliver up to 96% throughput of local memory, bringing it extremely close to the maximum theoretical performance limit an NVIDIA DGX-1 system can achieve.
Isilon-based DL solutions deliver the capacity, performance, and high concurrency to eliminate the IO storage bottlenecks for AI. This provides a rock-solid foundation for large scale, enterprise-grade DL solutions with a future proof scale-out architecture that meets your AI needs of today and scales for the future.

Our experts say about Dell EMC Certification Exams



Thursday, January 17, 2019

Sony Entertainment & Technology, and the Digital Transformation of Entertainment - Dell EMC Certifications


Ahead of the holidays, I had the chance to make one last customer/partner visit to the Sony Pictures studios lot, and what I saw blew me away. In the land of fantasy, celluloid and big ideas we met with Innovation Studios and saw yet another example of why one of this year’s biggest trends will be how data is continuing to reshape industries.

Dell Technologies, Intel, Deloitte and Sony Pictures Entertainment came together in 2018 to help create Innovation Studios.  My trip was to follow up on this and see Sony’s vision of the future and how they were using our technology to make it happen. Dell Technologies supported Innovation Studios with PowerEdge Servers, Dell EMC Networking, Isilon Storage, Dell workstations and VR gear. I saw just how much the analog and digital worlds are blurring through volumetric image acquisition, a process that produces a photo-realistic three-dimensional recreation of a physical environment.

That description fails to capture the extent of what I saw. Imagine the room you’re now standing in, pick an object and gaze at it. Now move your head from side to side and up and down. See how the light hits it and illuminates it differently as your focus shifts? Now imagine walking behind that object and disappearing or having the object cast a shadow onto you. Volumetric image acquisition allows a person to interact with an object that isn’t there. The scene is shot against a green screen, a method most are familiar with from major blockbusters. But unlike CGI being inserted during post-production, this technology displays the virtual background and the physical actor/props in real-time allowing the director to see the finished product as they are shooting.

This is a revolutionary way to tackle filmmaking that will increasingly find its way into blockbusters and indie films alike. Glenn Gainor, president of Innovation Studios, hopes to democratize storytelling by using technology to bring down the costs of filmmaking while creating immersive experiences that have previously been impossible to create.

The studio looks pretty standard at first sight: to the right is a mini data center humming full of our equipment, and to the left is production equipment, cameras and a small stage with a green screen. When Glenn flipped on the camera, the “Shark Tank” exit interview set magically appeared the monitors (this virtual set had just made its debut on the show a few days earlier). As the camera moved around the empty studio, it appeared we were right there on the “Shark Tank” set, our eyes completely fooled. A Variety reporter recently noted a similar reaction from a “Shark Tank” production manager.

Right now the rules are being broken. Inspirational leaders like Glenn are attempting to rewrite the story on how things are done. Imagine what this means for the entertainment industry. No more waiting until the lighting is just right. No more having to pay massive amounts to shoot on location. The ability to quickly spin up an environment for reshoots. This approach will deliver so much creative freedom and fundamentally change the way movies are made.

Every industry will go through a similar transformation, but incorporating technology into the culture and skill set of a company can be difficult. Organizations need strong technology partners who are committed to helping them achieve these digital outcomes. In the case of Innovation Studios, that meant a partnership with Dell Technologies, Deloitte and Intel to deliver the technology solutions that made this possible. Volumetric capture is a very CPU and GPU intensive process and reproducing it live requires substantial bandwidth in compute, networking and high-performance storage. The solution deployed features Intel Xeon-powered PowerEdge servers, Dell EMC Networking and Isilon storage on the back end, plus Dell Precision Workstations. Our team has worked very closely with Innovation Studios to ensure they had the right hardware configured to deliver the performance and scale this solution required. I can’t wait to see how this reshapes the future of filmmaking. These are exciting times!

Our experts say about Dell EMC Certification Exams



Thursday, January 3, 2019

Best Practices for Robust VMware Data Protection in Data Environment - Dell EMC Certifications


Everyone has heard the numbers.  There is more and more data and it keeps growing at an ever increasing rate.  According to the latest IDC estimates, the amount of data is expected to grow tenfold by 2025 to 163 zettabytes.  To thrive in today’s economy, organizations must make the most of their data and must also ensure that the data is protected and can be recovered quickly and cost effectively. And, since majority of enterprise workloads run on virtualized environments, with most of those workloads running on VMware, having robust VMware data protection is essential to the success of most organizations.

A number of best practices deployed as part of your data protection strategy can help you achieve this:

Automate protection of virtual workloads


With more data and more VMs being spun up at an ever increasing rate, you cannot rely on manual processes to ensure that all your new applications and VMs are protected.  You do not want busy application owners or vAdmins to have to manually create new protection policies or assign new workloads or VMs to existing policies.  And, you certainly do not want the lag time from the creation of a workload to raising a request with a backup or storage admin to configure backups of the workload.

Modern data protection solutions automate protection processes with support for dynamic mapping of vSphere objects to existing protection policies.  This means that your new VMs, based on criteria you define (such as DS cluster, data center, tags, VMname, etc.), are automatically assigned and protected upon creation with no human intervention necessary.

Self-service data protection


Your application owners and vAdmins are busy making sure that your business and mission critical applications are up and performing as intended.  

They do not want any delays with respect to data protection related tasks (backup policy changes, file and/or VM recoveries, etc.) by going through a backup or infrastructure admin and they certainly do not want to learn or log into another system or UI to do it on their own.  Enabling your application and vAdmins to perform data protection tasks directly from the native vSphere UI they are familiar with can go a long way in making their jobs, with respect to data protection, easier.

Modern data protection solutions provide deep integration with vSpehre to deliver self-service data protection workflows, enabling application owners and vAdmins the freedom to perform the vast majority of data protection related tasks directly from the native and familiar vSphere UI.

Distributing data across Software Defined Data Center nodes


Most data protection solutions handle data deduplication and distribution of data at the edge of the SDDC, after the data has been transferred from the source storage to the target backup storage.  This results in extensive network traffic and bandwidth usage.  It can also lead to the requirement of a separate expensive backup network and inefficient use of storage space.

Modern data protection solutions perform deduplication processing within the SDDC before the data is transferred to backup storage.  This allows the solution to scale-out with the SDDC and ensures data processing requirements do not fall behind as the SDDC scales. In addition, an ideal data protection solution should also utilize horizontally scaled out pools of protection storage, which results in more efficient allocation of software defined storage.  The result is less bandwidth usage and less storage capacity consumed for backing up your data.

Change block tracking restore capabilities


Change block tracking for backups is very common among today’s data protection solutions.  This means that for each subsequent VM backup, the solution only copies data blocks that have changed since the last backup.  This results in faster backups and less bandwidth consumed.

What is not very common, but important is change block tracking restore.  This allows the solution to track the difference between the current version of the VM and the version you are trying to restore.  It then only restores the changed blocks between the two.  The result is much faster VM restores and much less bandwidth consumed for those restores.  A solution that supports change block tracking restore enables new architectures without compromising recovery time objectives. One such architecture is the ability to backup and restore VMs across wide area networks. A modern data protection solution will have support for change block tracking restores.

Performance disaggregated from capacity


There is a trend in data protection solutions towards simplicity in the form of converged appliances that combine data protection software and backup storage, along with all the ancillary compute and networking, in a single converged appliance.  While this trend towards simplicity is admirable, for many data protection solutions, it has come at the expense of capacity and/or performance efficiency.

For many of these simple, converged solutions capacity and processing are linked.  Once you run out of processing power on an appliance – too many VMs to backup, too many backup streams needed to comply with SLAs, etc. – you need another appliance to add more processing even if you did not need the additional storage capacity.  Similarly, if you run out of storage capacity, you need another appliance and end up paying for the additional processing that you did not need. This is a common dilemma of architectures that scale capacity and performance linearly.

A modern data protection solution, whether it is in the traditional HW/SW form factor, SDDC or a converged appliance, disaggregates performance from capacity and allows each to scale independently.  Ideally, the backup processes should be load balanced across your entire SDDC environment and there should be no need to add or pay for additional processing if you need more capacity.

Dell EMC delivers software defined, automated data protection


With Dell EMC Data Protection, you don’t have to compromise on performance, automation, architecture or efficiency for the sake of simplicity.  Our solutions, whether you are deploying our Data Protection software and Data Domain backup storage or our new Integrated Data Protection Appliances, each utilize our software defined data protection architecture to deliver on each of the above best practices.

Dell EMC Data Protection solutions deliver:


  • Automation of VM protection with dynamic mapping to protection policies
  • Leading self-service data protection for VMware environments with our native vSphere integration
  • Minimal bandwidth usage and maximum storage efficiency with our industry leading edge and target deduplication technology that delivers up to 55:1 storage efficiency and up to 98% bandwidth efficiency
  • Change block restore for lighting fast VM recovery
  • Disaggregated performance from capacity with distributed virtual proxies that are automatically provisioned, deployed, and load balanced across the SDDC environment

Success Secrets: How you can Pass Dell EMC Certification Exams in first attempt