ZeroStack announced that administrators of its Self-Driving Cloud platform can provide single-click deployment of GPU resources and deep learning frameworks like TensorFlow, PyTorch, and MXNet, taking care of all the OS and CUDA library dependencies so users can focus on AI development.
Furthermore, users can enable GPU acceleration with dedicated access to multiple GPU resources for an order-of-magnitude faster inference latency and user responsiveness. GPUs within hosts can be shared across users in a multi-tenant manner.
Artificial Intelligence and Machine Learning products and solutions are quickly becoming commonplace and are shaping our experiences in computing like no other time in history, and AI applications and solutions are now more viable than ever with the availability of modern machine learning and deep learning frameworks such as TensorFlow, Caffe, etc., along with access to GPUs that are built specifically to perform parallel operations on large amounts of data. However, one significant challenge remains: deploying, configuring, and executing these complex tools and managing their interdependencies and versioning and compatibility with servers and GPUs.
ZeroStack’s AI-as-a-service capability gives customers powerful features to automatically detect GPUs and make them available for users to run their AI applications. In order to maximize utilization of this powerful resource, cloud admins can configure, scale, and allow fine-grained access control of GPU resources to end users.
“ZeroStack is offering the next level of cloud by delivering a collection of point-and-click service templates,” said Michael Lin, Director of Product Management at ZeroStack. “Our new AI-as-a-service template automates provisioning of key AI tool sets and GPU resources for DevOps organizations.”