Covenant Labs Covenant Labs Platform

Private OSS LLM Platform

Covenant Labs Platform

Covenant Labs Platform is the fastest way to deploy private open‑source LLMs for testing, experimenting, and production.

Private Deployments
OpenAI-compatible API
Usage-based billing
Covenant Labs Logo Mini

Quick Start

Launch a private model

Model

Llama 3.1 8B

Context

128k

Status

Ready

Price/hr

Calculated

Fast Iteration

Experiment quickly find which model works best for your problem.

Transparent pricing

We calculate the exact compute required for your model and let you choose the compute platform that makes the most sense.

Private by default

Deploy inside environments with audit‑friendly logs and no vendor data retention.

Transparent pricing

No black boxes. Enter your model specs — we’ll calculate VRAM and show GPUs that fit, with clear per-hour pricing.

  • Exact compute breakdown
  • Predictable per-hour pricing, terminate any time

GPU

NVIDIA L4 (24GB)

$/hr

$0.60

GPU

NVIDIA A100 (80GB)

$/hr

$2.30

GPU

NVIDIA H100 (80GB)

$/hr

$4.10

Estimates shown for demo purposes. Final pricing is computed from your model and chosen infrastructure.

Experimental end-to-end model encryption

Protect weights at rest and in use through our Model Encryption Protocol and the Covenant Labs Python SDK. Our experimental secure inference keeps your keys under your control while enabling OpenAI-compatible requests.

  • Hold your own secrets — we never see your keys.
  • Compatible with OpenAI-style APIs for seamless swap-in.
  • Deploy in stages: encrypt → deploy → ready.

How experimental secure inference works

  1. Encrypt model weights locally using keys you manage.
  2. Deploy to your chosen compute; keys never leave your boundary.
  3. Send requests via an OpenAI-compatible endpoint; responses are processed securely.
  4. Audit-friendly logs with no vendor data retention.

This flow lets teams experiment with private OSS LLMs while maintaining strict control over secrets and artifacts — ideal for pilots, sensitive evaluations, and staged rollouts.

Experimental feature — availability may vary by model and environment.