OctoAI launches OctoStack for enterprises to customize, deploy private AI models

by | Apr 2, 2024 | Technology

Join us in Atlanta on April 10th and explore the landscape of security workforce. We will explore the vision, benefits, and use cases of AI for security teams. Request an invite here.

Seattle-based OctoAI has a new offering called OctoStack, designed to help those in the enterprise deploy private generative AI models. Companies can use this “turn-key production platform” in a virtual private cloud or on-premises and will have access to highly optimized inference, model customization and asset management. In doing so, OctoAI wants to give companies the freedom to build and run gen AI applications in the way they see fit.

“Enabling customers to build viable and future-proof Generative AI applications requires more than just affordable cloud inference,” Luis Ceze, OctoAI’s chief executive, said in a statement. “Hardware portability, mode onboarding, fine-tuning, optimization, load-balancing — these are full-stack problems that require full-stack solutions.”

OctoStack supports fine-tuning and deployment of a range of open source and commercial AI models, such as Meta’s Llama family, Mistral’s 8x8B and Stable Diffusion models. However, it doesn’t include Anthropic’s Claude, because the AI is only offered in the cloud via Anthropic. “But we offer a lot of these super capable open source models that you can fully control and customize for,” Ceze said.

How the OctoStack platform works in the enterprise. Image credit: OctoAI

From Fully Managed to Do-It-Yourself

This isn’t the first attempt by the startup to provide companies with a packaged AI offering. Last year, OctoAI released its self-optimizing infrastructure service. As Ceze explains, the difference is that the feature introduced back then is now a fully managed solution. “That means that you call our APIs, offers highly efficient inference, and we have support for customizing the model,” he told VentureBeat. “We have support for building model cocktails and so on, all with the enterprise and production in mind.”

VB Event
The AI Impact Tour – Atlanta

Continuing our tour, we’re headed to Atlanta for the AI Impact Tour stop on April 10th. This exclusive, i …

Article Attribution | Read More at Article Source

[mwai_chat context=”Let’s have a discussion about this article:nn
Join us in Atlanta on April 10th and explore the landscape of security workforce. We will explore the vision, benefits, and use cases of AI for security teams. Request an invite here.

Seattle-based OctoAI has a new offering called OctoStack, designed to help those in the enterprise deploy private generative AI models. Companies can use this “turn-key production platform” in a virtual private cloud or on-premises and will have access to highly optimized inference, model customization and asset management. In doing so, OctoAI wants to give companies the freedom to build and run gen AI applications in the way they see fit.

“Enabling customers to build viable and future-proof Generative AI applications requires more than just affordable cloud inference,” Luis Ceze, OctoAI’s chief executive, said in a statement. “Hardware portability, mode onboarding, fine-tuning, optimization, load-balancing — these are full-stack problems that require full-stack solutions.”

OctoStack supports fine-tuning and deployment of a range of open source and commercial AI models, such as Meta’s Llama family, Mistral’s 8x8B and Stable Diffusion models. However, it doesn’t include Anthropic’s Claude, because the AI is only offered in the cloud via Anthropic. “But we offer a lot of these super capable open source models that you can fully control and customize for,” Ceze said.

How the OctoStack platform works in the enterprise. Image credit: OctoAI

From Fully Managed to Do-It-Yourself

This isn’t the first attempt by the startup to provide companies with a packaged AI offering. Last year, OctoAI released its self-optimizing infrastructure service. As Ceze explains, the difference is that the feature introduced back then is now a fully managed solution. “That means that you call our APIs, offers highly efficient inference, and we have support for customizing the model,” he told VentureBeat. “We have support for building model cocktails and so on, all with the enterprise and production in mind.”

VB Event
The AI Impact Tour – Atlanta

Continuing our tour, we’re headed to Atlanta for the AI Impact Tour stop on April 10th. This exclusive, i …nnDiscussion:nn” ai_name=”RocketNews AI: ” start_sentence=”Can I tell you more about this article?” text_input_placeholder=”Type ‘Yes'”]

Share This