vScaler has launched its AI optimised Architecture including vScaler Cloud Storage, which centralizes all I/O and places active files on the fastest tier (Flash/NVMe) but also leverages the cost benefits of an erasure coded HDD tier and provides policy-based movement between these tiers— all within a single platform.
vScaler, an optimised cloud platform built with AI and Deep Learning workloads in mind – provides a production ready environment with integrated Deep Learning application stacks, RDMA accelerated fabric and optimised NVMe storage, eliminating the administrative burden of setting up these complex AI environments manually.
The development of AI solutions is complex and varies greatly depending on the application. From data capture and ingestion, to preprocessing and algorithm development – the AI lifecycle involves many stages and technologies.
The dynamic nature of these AI environments typically means that the datasets are constantly growing, and algorithms are being optimised for greater accuracy. An infrastructure that can handle growth seamlessly and grow as your requirements grow is crucial.
The NVIDIA DGX-2 system (as well as previous generation DGX-1) have been integrated into vScaler as the building block for intensive workloads, providing the power of 16x V100 32GB GPUs.
Featuring GPUDirect RDMA thanks to compatible, high performance RDMA fabric from Mellanox, the platform provides a significant decrease in GPU-GPU communication latency and completely offloads the CPU.
The AI optimised Architecture includes vScaler Cloud Storage, which centralizes all I/O and places active files on the fastest tier (Flash/NVMe) but also leverages the cost benefits of an erasure coded HDD tier and provides policy-based movement between these tiers— all within a single platform.
David Power, CTO at vScaler said,
“With AI becoming more prevalent in almost every industry, it’s clear that a platform that can manage the complete AI development and production life cycle will become more and more critical for organisations wishing to adopt an AI strategy. Whether you are starting out on your AI journey or looking to scale an existing infrastructure, we can help with simplifying the configuration and management of all software and storage in a cost-effective and easy to use environment”.
The vScaler platform features containerised stacks (including the RAPIDS suite of software libraries) which enable users to spin up application specific environments with the appropriate Deep Learning frameworks installed and ready for use, including TensorFlow, Caffe and Theano (and many more) orchestrated on individual VMs or in a scale-out manner using Kubernetes on bare metal.
These frameworks are accelerated using the world’s fastest GPUs, purpose-built to dramatically reduce training time for Deep Learning and Machine Learning algorithms and AI simulations.
You can download a free copy of vScaler’s AI Reference Architecture White paper here.