Abhi Gupta, Senior Product Analyst
Sayce Falk

Sep 19, 2022

Reduce AI/ML Production Workload Costs More Than 70% by Automating Hardware Independence with OctoML

In the current economic climate, our customer conversations have begun to shift from AI innovation to AI/ML budgets, as technology and business leaders explore all avenues for cost savings.

Featured

Fast-track to deploying machine learning models with OctoML CLI and NVIDIA Triton Inference Server

Today, we introduce the OctoML CLI, a Command Line Interface that automates the deploying deep learning models - model containerization and acceleration. One of the key technologies that ties our containerization and acceleration together is NVIDIA Triton Inference Server.

Sameer Farooqui
André Kang-Moeller, Staff Software Engineer, OctoML

Jun 21, 2022

All Posts

Abhi Gupta, Senior Product Analyst
Sayce Falk

Sep 19, 2022

Abhi Gupta, Senior Product Analyst
Sayce Falk

Sep 19, 2022

Reduce AI/ML Production Workload Costs More Than 70% by Automating Hardware Independence with OctoML

In the current economic climate, our customer conversations have begun to shift from AI innovation to AI/ML budgets, as technology and business leaders explore all avenues for cost savings.

OctoML Logomark Container
OctoML Logomark Container

10 tips for OctoML CLI power users to fast-track your model deployments

The OctoML CLI is a sophisticated and powerful tool to fast-track your machine learning deployments. We wanted to share our favorite tips to get the most out of your accelerated ML containers.

Jared Roesch

Jared Roesch

Jun 21, 2022

Jared Roesch

Jared Roesch

Jun 21, 2022

New AI/ML Tool for Fans of Docker Compose and Kubernetes–OctoML CLI

We’re excited to share with you our first public release of OctoML CLI (v0.4.4) which provides an ML deployment workflow which should feel very familiar to anyone who uses Docker, Docker Compose and Kubernetes.

Sameer Farooqui
André Kang-Moeller, Staff Software Engineer, OctoML

Jun 21, 2022

Sameer Farooqui
André Kang-Moeller, Staff Software Engineer, OctoML

Jun 21, 2022

Fast-track to deploying machine learning models with OctoML CLI and NVIDIA Triton Inference Server

Today, we introduce the OctoML CLI, a Command Line Interface that automates the deploying deep learning models - model containerization and acceleration. One of the key technologies that ties our containerization and acceleration together is NVIDIA Triton Inference Server.

Adrian Sampson

Adrian Sampson

Apr 6, 2022

Adrian Sampson

Adrian Sampson

Apr 6, 2022

Beware the Tensor From Hell: How to Avert the Combinatorial Complexity Explosion in ML Engineering

In an earlier era, deploying ML systems was not exactly easy, but at least the problem was contained. Engineers could focus on a few important models, all of which built on similar libraries and tools. That era of uniformity is gone.

1

...

Accelerate Your AI Innovation