Jobs
My ads
My job alerts
Sign in
Find a job Career Tips Companies
Find

Ai research engineer (model serving & inference)

London
Tether Operations Limited
Research engineer
Posted: 11h ago
Offer description

Join Tether and Shape the Future of Digital Finance

At Tether, were pioneering a global financial revolution with innovative blockchain solutions that enable seamless digital token transactions worldwide. Our products include the trusted stablecoin USDT, energy-efficient Bitcoin mining solutions, advanced data sharing apps like KEET, and educational initiatives to democratize digital knowledge.

Why join us? Our remote, global team is passionate about fintech innovation. We seek individuals with excellent English communication skills eager to contribute to cutting-edge projects in a fast-growing industry.

About the job:

As part of our AI model team, you will innovate in model serving and inference architectures for advanced AI systems. Your focus will be on optimizing deployment strategies to ensure high responsiveness, efficiency, and scalability across various applications and hardware environments.

Responsibilities:

1. Design and deploy high-performance, resource-efficient model serving architectures adaptable to diverse environments.
2. Establish and track performance metrics like latency, throughput, and memory usage.
3. Develop and monitor inference tests, analyze results, and validate performance improvements.
4. Prepare realistic datasets and scenarios to evaluate model performance in low-resource settings.
5. Identify bottlenecks and optimize serving pipelines for scalability and reliability.
6. Collaborate with teams to integrate optimized frameworks into production, ensuring continuous improvement.

Qualifications:

* Degree in Computer Science or related field; PhD preferred in NLP, Machine Learning, with a strong publication record.
* Proven experience in low-level kernel and inference optimizations on mobile devices, with measurable improvements.
* Deep understanding of model serving architectures, optimization techniques, and memory management in resource-constrained environments.
* Expertise in CPU/GPU kernel development for mobile platforms and deploying inference pipelines on such devices.
* Ability to apply empirical research to overcome latency, bottleneck, and memory challenges, with experience in evaluation frameworks and iterative optimization.


J-18808-Ljbffr

Apply
Create E-mail Alert
Job alert activated
Saved
Save
Similar job
Ai research engineer llm python
London
Permanent
Client Server
Research engineer
£95,000 a year
Similar job
Senior research engineer, deep learning for cancer genomics
London
InstaDeep
Research engineer
Similar job
Research engineer
London
InstaDeep
Research engineer
See more jobs
Similar jobs
Engineering jobs in London
jobs London
jobs Greater London
jobs England
Home > Jobs > Engineering jobs > Research engineer jobs > Research engineer jobs in London > AI Research Engineer (Model Serving & Inference)

About Jobijoba

  • Career Advice
  • Company Reviews

Search for jobs

  • Jobs by Job Title
  • Jobs by Industry
  • Jobs by Company
  • Jobs by Location
  • Jobs by Keywords

Contact / Partnership

  • Contact
  • Publish your job offers on Jobijoba

Legal notice - Terms of Service - Privacy Policy - Manage my cookies - Accessibility: Not compliant

© 2025 Jobijoba - All Rights Reserved

Apply
Create E-mail Alert
Job alert activated
Saved
Save