ML Staff Engineer

VLM Run · AI/ML

staff

Salary Range (USD)

Negotiable

Location

Santa Clara, USA

Visa Support

Not mentioned

Funding Stage

Unknown

Job Responsibilities

  • Building the inference and orchestration layer for production Vision-Language Models

Engineering Culture & Tech Stack

RustPython
fast and ergonomic visual inference
reliable structured outputs
observability

Raw Post

Show original text
VLM Run (https://vlm.run) | 1x Product + 1x ML Staff Engineer | Santa Clara, CA (HQ) We're building the inference and orchestration layer for production Vision-Language Models. We care deeply about fast and ergonomic visual inference, reliable structured outputs, and the observability to iterate on them. A few things we've shipped recently you can poke at: 1. Orion: our visual agent that reasons and acts over images, video, and documents. Chat at https://chat.vlm.run. 2. mm-ctx: a Unix-style multimodal CLI (find, cat, grep, wc) that gives coding agents real context over images, video, and PDFs. Rust core, Python devex. 3. vlmbench: single-file CLI for benchmarking VLM inference (TTFT, TPOT, throughput) across vLLM, Ollama, and SGLang. Apply: https://app.dover.com/jobs/vlm-run Email hiring "at" vlm.run with your GitHub + a couple recent projects. [1] https://chat.vlm.run [2] https://pypi.org/project/mm-ctx | https://www.vlm.run/open-source/mm [3] https://github.com/vlm-run/vlmbench | https://www.vlm.run/open-source/vlmbench

AI Risk Insights

No major risk signals detected.

Recent News

No recent updates

Data Source

Content parsed by LLM from Hacker News raw data. Confidence:HIGH

© 2026 Job Signal

Built with ❤️ by Anna4code