Cloudera has made two of its artificial intelligence products available to run in on-premise data centres, making its AI Inference service and AI Studios available for customers to run on their own private infrastructure.
Cloudera AI Inference services provides inference services with embedded Nvidia NIM microservice capabilities to help deploy and manage AI production.
AI Studios provides low-code templates for building generative AI applications and agents.
The company stated that by bringing the benefits of Cloudera Data Services to an organisation’s data centre, users can "significantly reduce infrastructure costs and streamline data lifecycles, accelerate workload deployment, enhance security by automating complex tasks, and achieve faster time to value for AI deployment".
Bringing private AI on-premises also means "giving enterprises secure, GPU-accelerated generative AI capabilities behind their firewall ... with built-in governance and hybrid portability, organisations can now build and scale their own sovereign data cloud in their own data centre," Cloudera said.
Cloudera chief product officer Leo Brunnick said the approach provides “agility and efficiency without sacrificing security or control”.
“This release is a significant step forward in data modernisation, moving from monolithic clusters to a suite of agile, containerised applications," he said.