Skip to content

Senior Software Engineer – AI Platform
Company | Humane |
---|
Location | San Francisco, CA, USA |
---|
Salary | $180000 – $230000 |
---|
Type | Full-Time |
---|
Degrees | |
---|
Experience Level | Senior |
---|
Requirements
- 5+ years of hands-on software development experience, including recent expertise in distributed systems, networked devices, or embedded AI.
- Strong background in machine learning model deployment, including LLMs, computer vision, or speech models.
- Experience integrating AI models on edge devices, IoT platforms, or hybrid cloud-edge systems.
- Ability to quickly adapt to new AI advancements and apply them to scalable real-world applications.
- Fast-paced, iterative approach to AI development and deployment.
- 5+ years of experience with Java or C++.
Responsibilities
- Develop and optimize edge inference pipelines to run AI models efficiently on resource-constrained devices while integrating with cloud services.
- Design and implement hybrid cloud-edge AI architectures that balance local inference with cloud scalability.
- Integrate and third-party machine learning APIs, while also training and optimizing open source models for real-time AI applications.
- Build distributed systems and networked AI devices, ensuring seamless connectivity and interoperability between edge and cloud environments.
- Evaluate on-device AI accelerators and emerging hardware solutions to maximize performance, security, and efficiency.
- Drive new features from prototype to production, collaborating closely with AI researchers, infrastructure engineers, and designers.
- Work on privacy-first AI architectures, ensuring data security and compliance in localized AI deployments.
- Develop robust testing and validation pipelines for non-deterministic AI systems, including LLMs and real-time inference models.
Preferred Qualifications
- Expertise in on-device ML optimization techniques (e.g., quantization, pruning, distillation).
- Strong understanding of low-latency computing, edge inference, and privacy-preserving AI.
- Familiarity with AI hardware accelerators (e.g., NPUs, TPUs, GPU-based inference).
- Experience evaluating and integrating third-party inference engines (e.g., TensorFlow Lite, ONNX Runtime, CoreML).
- Knowledge of cloud-edge hybrid architectures, including distributed AI model serving and orchestration.
- Experience developing or working with open-source ML frameworks.
- Prompt engineering and experimentation with AI-powered hardware.