×
Register Here to Apply for Jobs or Post Jobs. X

Senior Software Engineer, AI Inference

Remote / Online - Candidates ideally in
Coos Bay, Coos County, Oregon, 97458, USA
Listing for: NVIDIA
Remote/Work from Home position
Listed on 2026-01-13
Job specializations:
  • Software Development
    AI Engineer, Cloud Engineer - Software, Machine Learning/ ML Engineer, Software Engineer
Salary/Wage Range or Industry Benchmark: 356500 USD Yearly USD 356500.00 YEAR
Job Description & How to Apply Below

Employer Industry: Technology and AI Computing

Why consider this job opportunity:
  • Salary up to $356,500
  • Opportunity for equity and comprehensive benefits
  • Remote-friendly work environment, allowing flexibility
  • Work with cutting-edge technology in AI and deep learning
  • Join a diverse and supportive team passionate about innovation
  • Influence the future of computing and make a lasting impact
What to Expect (Job Responsibilities):
  • Build and maintain distributed model management systems for large-scale AI inference workloads
  • Implement inference scheduling and deployment solutions on Kubernetes and Slurm
  • Collaborate with infrastructure engineers and researchers to develop scalable APIs and services
  • Create monitoring, benchmarking, automation, and documentation processes for production-ready inference systems
  • Take ownership of problems end-to-end and proactively acquire necessary knowledge or skills
What is Required (Qualifications):
  • Bachelor’s, Master’s, or PhD in Computer Science, ECE, or related field (or equivalent experience)
  • 6+ years of professional software engineering experience
  • Strong understanding of modern ML architectures and optimization of inference performance
  • Familiarity with vLLM, SGLang, PyTorch, NVIDIA GPUs, and related software stacks
  • Experience in architecting, building, monitoring, and debugging production-grade distributed systems
How to Stand Out (Preferred Qualifications):
  • Experience with inference-serving frameworks (e.g., Dynamo Inference Server, Tensor

    RT, ONNX Runtime)
  • Contributions to large-scale, low-latency distributed systems, preferably in open-source
  • Strong background in GPU inference performance tuning and cloud-native environments (AWS, GCP, Azure)

#Technology #AI #Deep Learning #Remote Work #Career Opportunity

We prioritize candidate privacy and champion equal-opportunity employment. Central to our mission is our partnership with companies that share this commitment. We aim to foster a fair, transparent, and secure hiring environment for all. If you encounter any employer not adhering to these principles, please bring it to our attention immediately.

We are not the EOR (Employer of Record) for this position. Our role in this specific opportunity is to connect outstanding candidates with a top-tier employer.

#J-18808-Ljbffr
Position Requirements
10+ Years work experience
To View & Apply for jobs on this site that accept applications from your location or country, tap the button below to make a Search.
(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary