AI System Engineer
1Mind
Location
Remote - US, San Francisco
Employment Type
Full time
Location Type
Remote
Department
Engineering
About Us
1mind is a platform that deploys multimodal Superhumans for revenue teams. These Superhumans combine a face, a voice, and a GTM brain — equipped with deep technical and product knowledge. They can lead unlimited, simultaneous conversations 24/7, meeting buyers when they’re most active and engaged. Superhumans qualify leads, book meetings, deliver pitches, give interactive demos, handle objections, uncover pain points, build value models, provide support, and onboard customers. They live across websites, inside your product, can join live calls as active participants, and work alongside your team in deal rooms. 1mind Superhumans integrate seamlessly into existing workflows, scale instantly, and drive measurable impact — growing revenue, reducing headcount, accelerating pipeline to closed-won, and creating a more delightful buyer experience.
Job Description
We're looking for an AI Systems Engineer to design and build the technical foundation powering 1mind's Superhumans, real-time, multimodal AI beings that communicate through audio, video, and language. You'll work across pipelines, infrastructure, and orchestration layers to make Superhumans responsive, expressive, and reliable in production environments.
You'll partner closely with AI, infra, and product teams to integrate large language, speech, and rendering models into a unified, low-latency system that can power live human-like interaction at scale.
Key Responsbilities
Design and build real-time multimodal pipelines connecting ASR, LLMs, TTS, and avatar rendering into cohesive experiences
Develop the observability layer for the pipeline, metrics, tracing, and performance instrumentation for end-to-end visibility
Build infrastructure management tools to deploy, version, and scale models powering Superhuman interactions
Develop automated tests and evaluation frameworks to ensure reliability, synchronization, and quality across modalities
Integrate and optimize the full system with a focus on low latency, high reliability, and graceful degradation
-
Collaborate with product and AI teams to enable Superhumans to drive presentations, play media, share information, and be trained on knowledge bases for deeper contextual understanding
Qualifications
Experience in ML systems or applied AI, particularly in LLMs, ASR, or TTS pipelines
Strong background in real-time or streaming systems and concurrent architectures
Proven ability to productionize AI models, serving, scaling, and monitoring them in live environments
Proficiency in Python; experience deploying systems on Kubernetes or similar orchestration frameworks
Familiarity with TypeScript for frontend or orchestration logic
Comfort with automated testing, metrics, and observability as part of core development practice
Excitement about building multimodal AI systems that combine voice, video, and reasoning in real time
Plus: Exposure to multimodal orchestration frameworks such as LangChain, Pipecat, or Daily
-
Plus: Experience with latency optimization, streaming architectures, or real-time synchronization
Why Join Us?
Build the systems behind multimodal Superhumans redefining how AI communicates and collaborates
Work at the intersection of AI research, infrastructure, and real-time human-AI interaction
High-ownership role with deep technical scope and meaningful product impact
-
Global, fast-moving team shipping production-grade AI experiences at scale
Location: San Francisco preferred, open to remote (U.S.)
Employment Type: Full-time