Inferact's mission is to grow vLLM as the world's AI inference engine and accelerate AI progress by making inference cheaper and faster. Founded by the creators and core maintainers of vLLM, we sit at the intersection of models and hardware—a position that took years to build.

Member of Technical Staff, Exceptional Generalist (Remote)
Remote · Posted on Jan 22, 2026

Member of Technical Staff, Cloud Orchestration
San Francisco · Posted on Jan 22, 2026

Member of Technical Staff, Kernel Engineering
San Francisco · Posted on Jan 22, 2026

Member of Technical Staff, Performance and Scale
San Francisco · Posted on Jan 22, 2026

Member of Technical Staff, Inference
San Francisco · Posted on Jan 22, 2026














