Jobs/San Francisco/Software Engineer, Inference - Multi Modal
San Francisco, California, United States

Software Engineer, Inference - Multi Modal

About the Team OpenAI’s Inference team powers the deployment of our most advanced models - including our GPT models, 4o Image Generation, and Whisper - across a variety of platforms. Our work ensures these models are available, performant, and scalable in production, and we partner closely with Research to bring the ne

Company
OpenAI
Compensation
$295K - $555K
Schedule
Full-Time
Role overview

What this role actually needs.

Software Engineer, Inference - Multi Modal at OpenAI in San Francisco. UpJobz keeps this listing high-signal for applicants targeting serious high-tech roles across the United States, Canada, and Mexico. About the Team OpenAI’s Inference team powers the deployment of our most advanced models - including our GPT models, 4o Image Generation, and Whisper - across a variety of platforms. Our work ensures these models are available, performant, and scalable in production, and we partner closely with Research to bring the ne

Responsibilities

Day-to-day expectations

A clear list of the work this role is designed to cover.

  • Design and implement inference infrastructure for large-scale multimodal models.
  • Optimize systems for high-throughput, low-latency delivery of image and audio inputs and outputs.
  • Enable experimental research workflows to transition into reliable production services.
  • Collaborate closely with researchers, infra teams, and product engineers to deploy state-of-the-art capabilities.
  • Contribute to system-level improvements including GPU utilization, tensor parallelism, and hardware abstraction layers.
  • Have experience building and scaling inference systems for LLMs or multimodal models.
Requirements

What a strong candidate brings

This keeps the job page specific, readable, and easier to match.

    Benefits

    Why people would want this job

    Benefits help searchers understand whether the role is a real fit before they apply.

      Subscriber playbook

      Turn this listing into an application plan.

      This is the first pass at the premium UpJobz layer: a fast brief that helps serious applicants move with more clarity.

      Next moves

      • Tailor your resume around ai and llm instead of sending a generic application.
      • Use the first two bullets of your application to connect your background directly to software engineer, inference - multi modal is a high-signal on-site role in san francisco, and it is most realistic for united states residents.
      • Open the role quickly if it fits and bookmark three similar jobs before you leave the page.

      Interview themes

      Artificial IntelligenceOn-siteaillmmachine-learningresearch

      Watchouts

      • $295K - $555K is visible, so calibrate your application around the posted range.
      • Use united states residents as part of your positioning so the recruiter does not have to infer it.
      • Show concrete examples of succeeding in on-site environments.
      SEO context

      Search intent signals for this listing

      Helpful keyword hooks for serious tech searchers and future programmatic job pages.

      Software Engineer, Inference - Multi ModalOpenAISan FranciscoUSArtificial Intelligenceaillmmachine-learningresearchawssecurityplatformdeveloper-toolsapipythoninfrastructure
      Next step

      Ready to move on this role?

      This page keeps the application flow simple while giving you enough context to decide quickly and move.