Template-backed instances
Launch long-lived SSH, Jupyter, or custom entrypoint environments from reusable templates with versioned defaults.
SciFlow gives labs and AI teams a self-service control plane for template-backed GPU instances, org quotas, image commits, and usage visibility on top of Kubernetes.
Launch instance
PyTorch + Jupyter
Most teams build the same control plane twice — once in scripts, once in tickets. SciFlow replaces both.
Without a real admission layer, the loudest user wins and queues become Slack threads.
Researchers need long-lived SSH or Jupyter workspaces, not short-lived containers tied to a pod spec.
Saving a working environment for the next experiment usually means scripts, registries, and copy-pasted commands.
Org-level quotas, member quotas, and GPU usage rollups end up scattered across spreadsheets and dashboards.
Six product surfaces that together replace the patchwork of scripts, dashboards, and manual approvals most clusters rely on today.
Launch long-lived SSH, Jupyter, or custom entrypoint environments from reusable templates with versioned defaults.
Users pick which org they are charging while admins manage member quotas and integer GPU budgets per type.
Reject invalid requests, queue valid ones when capacity is temporarily unavailable. No empty pods reserving GPUs.
Save a configured workload as a reusable image to seed future templates and reproducible experiments.
Keep SSH keys and API keys attached to accounts and injected at launch time — not buried inside templates.
Usage records, billing summaries, GPU accounting, and scheduled rollups for cluster admins and finance.
A consistent path from template selection to a saved, reusable image — the same loop researchers already follow informally.
Choose a versioned template — image, launch mode, startup script, ports, env vars.
Pick the GPU type and fraction, and select which org membership you are charging.
Admission runs quota math. Start now if capacity exists, otherwise queue fairly.
Commit your running workload back into a reusable image for the next experiment.
Admin · Quotas
Cluster overview
GPUs total
24
In use
11
Queued
3
| Org | H100 | A100 | Queue |
|---|---|---|---|
| Vision Lab | 4 / 6 | 2 / 4 | 1 |
| NLP Group | 1 / 2 | 3 / 3 | 0 |
| Robotics | 0 / 1 | 1 / 2 | 2 |
SciFlow gives platform admins the controls and visibility they need without forcing the underlying cluster into a rigid layout.
Five focused services with clear ownership boundaries. Authentik handles login at the edge; SciFlow handles authorization, quota, and runtime.
Local user projection, quota ledgers, and admission decisions live here.
Templates, instance lifecycle, queue state, and runtime status.
Image metadata, commit operations, registry ownership.
Durable workers, retries, reconciliation, node-local execution.
Usage records, billing summaries, GPU accounting rollups.
Outside SciFlow
Authentik, oauth2-proxy, and cluster infra are managed in a separate FluxCD repository — SciFlow stays application-only.
Shared GPUs across students and projects with sane quotas.
A self-service surface that replaces ticket queues.
Org and member quota math without empty placeholder pods.
Reproducible templates and saved images per experiment.
Org-aware fairness with explicit queueing and leases.
SciFlow helps teams launch, govern, and reuse interactive GPU environments without turning Kubernetes into the user interface.