Exafunction
Exafunction optimizes your deep learning inference workload, delivering up to a 10x improvement in resource utilization and cost. Focus on building your deep learning application, not on managing clusters and fine-tuning performance. In most deep learning applications, CPU, I/O, and network bottlenecks lead to poor utilization of GPU hardware. Exafunction moves any GPU code to highly utilized remote resources, even spot instances. Your core logic remains an inexpensive CPU instance. Exafunction is battle-tested on applications like large-scale autonomous vehicle simulation. These workloads have complex custom models, require numerical reproducibility, and use thousands of GPUs concurrently. Exafunction supports models from major deep learning frameworks and inference runtimes. Models and dependencies like custom operators are versioned so you can always be confident you’re getting the right results.
Learn more
IBM Watson Machine Learning Accelerator
Accelerate your deep learning workload. Speed your time to value with AI model training and inference. With advancements in compute, algorithm and data access, enterprises are adopting deep learning more widely to extract and scale insight through speech recognition, natural language processing and image classification. Deep learning can interpret text, images, audio and video at scale, generating patterns for recommendation engines, sentiment analysis, financial risk modeling and anomaly detection. High computational power has been required to process neural networks due to the number of layers and the volumes of data to train the networks. Furthermore, businesses are struggling to show results from deep learning experiments implemented in silos.
Learn more
Zebra by Mipsology
Zebra by Mipsology is the ideal Deep Learning compute engine for neural network inference. Zebra seamlessly replaces or complements CPUs/GPUs, allowing any neural network to compute faster, with lower power consumption, at a lower cost. Zebra deploys swiftly, seamlessly, and painlessly without knowledge of underlying hardware technology, use of specific compilation tools, or changes to the neural network, the training, the framework, and the application. Zebra computes neural networks at world-class speed, setting a new standard for performance. Zebra runs on highest-throughput boards all the way to the smallest boards. The scaling provides the required throughput, in data centers, at the edge, or in the cloud. Zebra accelerates any neural network, including user-defined neural networks. Zebra processes the same CPU/GPU-based trained neural network with the same accuracy without any change.
Learn more
NVIDIA DIGITS
The NVIDIA Deep Learning GPU Training System (DIGITS) puts the power of deep learning into the hands of engineers and data scientists. DIGITS can be used to rapidly train the highly accurate deep neural network (DNNs) for image classification, segmentation and object detection tasks. DIGITS simplifies common deep learning tasks such as managing data, designing and training neural networks on multi-GPU systems, monitoring performance in real-time with advanced visualizations, and selecting the best performing model from the results browser for deployment. DIGITS is completely interactive so that data scientists can focus on designing and training networks rather than programming and debugging. Interactively train models using TensorFlow and visualize model architecture using TensorBoard. Integrate custom plug-ins for importing special data formats such as DICOM used in medical imaging.
Learn more