Skip to Content

Fullstack engineer (AI platform & APIs)

1 open position

Berget AI is building a developer-first, sovereign AI platform on open infrastructure. We run our own GPU hardware and provide Kubernetes-native inference APIs and platform services. We’re expanding the core product team and are looking for a backend developer who enjoys owning systems end-to-end and making complex infrastructure feel simple and reliable.

What you’ll work on

You will design, build, and operate backend services powering Berget AI’s platform, including inference APIs, OpenAI-compatible endpoints, internal product APIs, and usage and billing flows.

You’ll own features from idea to production: architecture, implementation, testing, deployment, monitoring, and iteration. You’ll keep releases flowing, fix bugs quickly, and continuously improve reliability and performance.

You’ll build Kubernetes-native services with strong focus on observability, security, and operational robustness. You’ll actively participate in incident response and ongoing security work, treating operational safety as a core product concern.

You’ll work closely with inference, platform, and community roles to translate real user needs into stable, developer-friendly APIs.

What you bring

You have strong experience building backend services in TypeScript/Node.js and designing clean, reliable APIs. We are using FluxCD as our GitOps platform and run everything i Kubernetes and we are obsessed with automating everything including testing, deployment so we can trust the system to heal itself when needed. You are used to using coding agents to accelerate your speed and accuracy.

You’re comfortable owning systems from PR to production and understand distributed systems fundamentals, failure modes, and secure service design.

You care about quality, reliability, and making infrastructure approachable for developers. Experience with Kubernetes-native application patterns is a strong plus. Familiarity with Python or inference runtimes (vLLM, SGLang) is a bonus, not a requirement.

You’re pragmatic, collaborative, and enjoy working in small teams with high ownership.

Why Berget

You’ll help shape the core APIs of a European sovereign AI platform, with real ownership, fast feedback loops, and close collaboration across engineering, product, and community.

Stockholm, Sweden
Full-Time

Infrastructure & operations engineer (hardware, network & storage)

1 open position

Berget AI operates its own GPU-based infrastructure across multiple data center sites. We’re looking for an infrastructure & operations engineer who owns the physical and low-level systems that power our platform: hardware, networking, storage, and on-site operations.

What you’ll work on

You will design, deploy, and operate the physical infrastructure powering Berget AI’s GPU data centers: servers, racks, storage, networking, power, and cooling.

You’ll plan and operate high-availability network architectures (L2/L3, routing, switching, firewalls), manage storage systems for performance and reliability, and oversee vendor and on-site work such as rack/stack, cabling, hardware swaps, labeling, and audits.

You’ll handle day-to-day operations including hardware lifecycle management, firmware updates, monitoring, and incident response. You’ll automate provisioning and configuration using scripting and infrastructure-as-code tools, and work closely with the platform team to ensure hardware, storage, and topology fit Kubernetes and inference needs.

Security, operational safety, and documentation are core responsibilities.

What you bring

You have hands-on experience operating data center or on-prem infrastructure in production. You understand servers, networking gear, storage systems, and how they fail.

You have strong networking fundamentals (VLANs, routing, BGP/OSPF, firewalls), solid Linux knowledge, and practical experience with storage systems (NVMe, networked or distributed storage, backups).

You’re systematic, calm under pressure, and enjoy making physical systems reliable and easy to operate.

Why Berget

You’ll help build and scale Europe’s sovereign AI compute infrastructure, with real ownership over hardware, networks, and storage in a small, highly technical team.

Stockholm, Sweden

Inference & fine-tuning engineer (models, performance & security)

1 open position

Berget AI builds and operates sovereign AI infrastructure on our own GPU hardware. We run large-scale inference in production and are expanding into fine-tuning and post-training for real customer workloads. We’re looking for an inference engineer who masters model bring-up, performance tuning, and secure operation from model release to GPU.

What you’ll work on

You will rapidly evaluate, integrate, and bring new state-of-the-art models into production inference environments. This includes understanding model architectures, dependencies, runtime constraints, and quickly getting models running reliably on our GPU stack.

You’ll optimize inference runtimes and serving stacks (vLLM, Triton, SGLang, CUDA/ROCm), tuning metaparameters such as batching, parallelism, memory layouts, quantization strategies, and scheduling to maximize throughput, minimize latency, and control cost per token.

You will design and operate fine-tuning and post-training workflows: data preparation, training configuration, evaluation, model packaging, and safe rollout into production inference systems.

You’ll work extensively with caching strategies at multiple levels (model, KV/cache, request/result caching) to improve performance, efficiency, and isolation in multi-tenant environments.

You’ll extend Kubernetes-native orchestration for large-scale model serving, profile bottlenecks, benchmark improvements, and continuously harden reliability, observability, and security. Incident response, secure model handling, and controlled rollout are core parts of the role.

What you bring

You closely follow the latest open and commercial model releases and enjoy getting new models running fast in real systems.

You have hands-on experience with inference and ML runtimes such as vLLM, Triton, SGLang, CUDA or ROCm, and understand how model architecture, runtime configuration, and hardware interact.

You’re familiar with fine-tuning or post-training techniques and understand the operational and security implications of shipping trained models into production.

You’re comfortable working in Kubernetes-based environments, think practically about caching and isolation, and treat performance, reliability, and security as inseparable concerns.

Why Berget

You’ll have real ownership over how models are onboarded, optimized, and served in one of Europe’s most ambitious sovereign AI platforms. You’ll work close to hardware, platform, and product, with freedom to push performance and define best practices from day one.

Stockholm, Sweden
Flexible

Platform / SRE / security engineer (Kubernetes & networking)

1 open position

Berget AI builds a sovereign, GPU-backed AI platform running on Kubernetes. We’re looking for a platform engineer who owns the runtime, reliability, and security of our Kubernetes-based systems and makes the infrastructure feel stable and simple for product and inference teams.

What you’ll work on

You will design, operate, and evolve our Kubernetes platform: cluster lifecycle, upgrades, capacity, and multi-tenant isolation. You’ll own core platform components such as GPU operators, storage integrations (CSI), networking, ingress, and observability.

You’ll build and maintain GitOps-based deployment workflows, CI/CD integrations, and internal tooling that supports product, inference, and ML workloads. You’ll participate in on-call and incident response, drive post-incident improvements, and continuously harden reliability and security across the platform.

You’ll work closely with backend, inference, and infrastructure engineers to translate workload needs into stable, scalable platform capabilities.

What you bring

You have hands-on experience running Kubernetes clusters in production and understand how systems fail in practice. You’re comfortable with Linux, containers, networking fundamentals, and debugging distributed systems.

You’ve worked with Kubernetes storage and networking in real environments and care about observability, automation, and security. Experience with GitOps, infrastructure automation, or GPU workloads is a strong plus.

You enjoy creating platforms that other engineers trust and like using.

Why Berget

You’ll own the backbone of a European AI platform, with real autonomy, short feedback loops, and direct influence on reliability, security, and developer experience.

Stockholm, Sweden
Full-Time