hosted·ai v2.0.1 – now it’s easy to optimize GPUaaS

We’re excited to announce the availability of hosted·ai v2.0.1, with new features to make GPUaaS even easier to manage and sell. Let’s get into it!

Screenshot of the hosted·ai control panel showing GPU infrastructure orchestration

#BuildingInPublic – the platform story so far

What’s new in v2.0.1?

1. Tune your GPU pools for different workloads


Introducing… the new GPU optimization slider.

When you create a GPU pool, you assign GPUs to the pool and choose the sharing ratio – i.e. how many tenants the resources of the pool can be allocated/sold to. For any setting above 1, the new optimization slider becomes available.

Behind this simple slider is a world of GPU cloud flexibility. The slider enables providers to configure each GPU pool to suit different customer use cases. Here’s a quick demo from Julian Chesterfield, CTO at hosted·ai:

  • GPUaaS optimized for security
    Temporal scheduling is used. The hosted·ai scheduler switches user tasks completely in and out of physical GPUs in the pool, zeroing the memory each time. At no point do any user tasks co-exist on the GPU. This is the most secure option, but comes with more performance overhead.
  • GPUaaS optimized for performance
    Spatial scheduling is used. The hosted·ai scheduler assigns user tasks simultaneously to make optimal use of the GPU resources available. There is no memory zeroing. This is the highest-performance option, but it doesn’t isolate user tasks – they are allocated to GPUs in parallel.
  • Balanced GPUaaS
    Temporal scheduling is used, but without fully enforced memory zeroing. This provides a blend of performance and security.

2. Self-service GPU / end user enhancements


Also in this release, some handy improvements for end users running their applications in your hosted·ai environments:

GPU application service exposure

We’re made it easier to expose ports for end user applications and services through the hosted·ai admin panel (and coming soon, through the user panel).

Now your customers can choose how they present their application services to the outside world, through configurable ports

Self-service GPU pool management

We’ve added new management tools for your customers too. Each GPU resource pool they subscribe to can be managed through their user panel, with visibility of the status of each pod; the ability to start, stop and restart pods; and logs with information about the applications using GPU.

3. Furiosa device integration


Now service providers can create regions with clusters based on Furiosa, as well as NVIDIA. Once a region has been set up for Furiosa, it can be managed, priced and sold using the same tools hosted·ai makes available for NVIDIA – and in future, other accelerator devices.

More information:


Coming next:


  • Full stack KVM – complete implementation, replacing Nexvisor
  • Scheduler credit system – expanding GPU optimization with a credit system to deliver consistent performance for inference in mixed-load environments
  • Billing enhancements – more additions to the hosted·ai billing and metering engine – more ways to monetize your service
  • Infiniband support

Subscribe to get updates