VLM Run · 7 hours ago
Founding Infrastructure Engineer
VLM Run is building the enterprise infrastructure layer for visual intelligence. They are seeking exceptional engineers to help scale the infrastructure layer for visual intelligence, focusing on optimizing the VLM inference stack and enhancing developer experience.
Artificial Intelligence (AI)Computer VisionEnterprise SoftwareInformation ServicesMachine LearningRoboticsSoftware
Responsibilities
Own and optimize the VLM inference stack end-to-end – from GPU serving and latency/cost to scalable backend systems and reliability
Design clean, ergonomic APIs for multimodal apps – tool/function calling, structured outputs, and workflows developers actually want to build
Take ideas from 0→1, set the bar for quality, and help define what 'production-grade visual intelligence' looks like
Qualification
Required
BS & 4+ YoE
Integrated or built applications with LLMs or VLMs (OpenAI, HuggingFace, Ollama, vLLM), with an understanding of prompt engineering, function calling, and structured outputs
Python, FastAPI, async API design, schema validation, caching, and performance optimization
Docker, Kubernetes, CI/CD, observability (logging, metrics, tracing), GCP or AWS
Postgres, MongoDB, Redis; experience with scalable, reliable data pipelines
Strong testing discipline (TDD), clean code, GitHub workflows (PRs, reviews, CI), and internal tooling mindset
Preferred
Shipped full-stack dev platforms or SaaS products – from landing pages to auth, billing, telemetry, and infra
Benefits
Great healthcare
401K
Company
VLM Run
VLM Run is an unified gateway for enterprise Visual AI that extract structured JSON from images, video, and documents.
Funding
Current Stage
Early StageTotal Funding
unknown2023-01-01Seed
Company data provided by crunchbase