GPU Instances in the Cloud - Rent Nvidia L4

Powerful performance for AI training, AI inference, rendering, high-intensity calculations, and multimedia.

Run advanced workloads directly in the cloud – without buying and operating your own hardware. Our GPU instances are built for enterprises, developers, and creators who demand high performance, low latency, and full control.

Why Hexabyte GPU?

Modern hardware – Nvidia L4 (PCIe Gen 4) with 24GB GDDR6 and energy-efficient Ada Lovelace architecture.

Flexible instances – Scale up or down as needed.

Local infrastructure – Low response times within the EU, servers located in Umeå, Sweden.

Transparent pricing – Also the most performance for the money.

Perfectly suited for

NVIDIA L4

24GB GDDR6

Massive graphics memory for large models and datasets.

PCIe Gen 4 Passthrough

Full control over the entire GPU, no sharing with other customers.

Up to 30.3 TFLOPS FP32

Raw computing power for AI and rendering.

AI-optimized Ada Lovelace architecture

Specialized in AI inference, AI training and video processing.

Energy-efficient performance

2.5x better performance per watt compared to previous generations.

Low response times within the EU

Server hall in Umeå, Sweden.

Pricing example Nvidia L4 GPU instances

The L4 family consists of GPU servers with Nvidia L4 GPUs, AMD EPYC 7713 CPUs, DDR4 RAM, and high-IOPS local NVMe disks in RAID.

l4.gpu1

3.91 SEK
/ hour
or
2627 SEK
/ month

l4.gpu2

4.28 SEK
/ hour
or
2875 SEK
/ month

l4.gpu3

4.82 SEK
/ hour
or
3239 SEK
/ month

l4.dual-gpu

10.69 SEK
/ hour
or
7187 SEK
/ month
Calculated at 672 hours per month. For a full month of use on the same instance, the cost does not exceed the fixed monthly price.

Own public IP addresses

1 IPv4 address and 1 IPv6 address are included per instance.

Renewable electricity

Our servers are powered by electricity from Umeå Energi, which supplies renewable electricity. In 2022, the energy consisted of 72% hydropower, 10% wind power, 16% biopower and 2% solar energy, read more at Umeå Energi's website.

NVMe SSD instead of hard drives

We know that our customers prioritize speed, so all data storage is done on NVMe SSDs of enterprise models. Faster and more power efficient! All data is replicated on three different disks.

We are our own internet provider

Internet for our services is delivered by us via our own AS number. Stable and secure connection - Without traffic restrictions!

FAQ

A GPU instance in the cloud is a virtual machine at a cloud provider with a GPU (graphics card) attached to the virtual machine. The GPU can be either a shared vGPU or a full GPU via PCIe passthrough. 

A regular VPS with a CPU is fine for general-purpose applications, but some workloads require the parallel processing power of a GPU. With a GPU instance, you can:

  • Train and run AI and ML models significantly faster
  • Render graphics and 3D in real time
  • Running video transcoding at scale
  • Take advantage of CUDA, TensorRT, and other NVIDIA-optimized libraries
    In short: CPU for general work, GPU for accelerated computing.

All our GPU instances come with dedicated NVMe storage locally in the serverThis disk is used as a scratch disk or primary storage and delivers extremely high IOPS and low latency, compared to traditional network storage. For workloads like AI training, data mining, and video processing, this means you get much faster access to data.

We offer flexible hourly billing where you only pay for the time your instance is running. For customers who want long-term operation, we also show the maximum monthly price. The debiting of the inserted payment card is done automatically and there is no binding times – you can start and stop your GPU instances whenever you want.

Companies can apply for an invoice as a payment method by submitting a support ticket in the cloud platform.

You get full root access to your GPU instance and can install your own drivers, libraries and software. This means you can set up everything from Docker containers with PyTorch/TensorFlow to full virtual workstations (vWS) with e.g. Remote Desktop or NICE DCV.

Support is handled via tickets directly in our cloud system, and you can write in both Swedish and English. Our technicians are located in Sweden and will help you with operational issues, instance management and troubleshooting.

No. All of our GPU plans only include dedicated NVIDIA L4 GPUs, connected via PCIe passthrough. This means you don't share GPU resources with other customers – unlike some cloud providers that offer lower performance “shared vGPU”. With us, performance is predictable and guaranteed.

Absolutely. NVIDIA L4 is great for both AI inference and model trainingThe card supports FP32, FP16, and INT8, making it efficient for both training and optimized inference. You can use popular frameworks like TensorFlow, PyTorch, and JAX without any problems.

We use cookies.