Staff Researcher, Large Model On-Device Inference (C++)

Lenovo

  • 北京市
  • 长期
  • 全职
  • 13天前
Staff Researcher, Large Model On-Device Inference (C++)General InformationReq #WD00085740Career area:Research/DevelopmentCountry/Region:ChinaState:BeijingCity:北京(Beijing)Date:Friday, August 22, 2025Working time:Full-timeAdditional Locations:
  • China - Beijing - 北京(Beijing)
Why Work at LenovoWe are Lenovo. We do what we say. We own what we do. We WOW our customers.Lenovo is a US$69 billion revenue global technology powerhouse, ranked #196 in the Fortune Global 500, and serving millions of customers every day in 180 markets. Focused on a bold vision to deliver Smarter Technology for All, Lenovo has built on its success as the world's largest PC company with a full-stack portfolio of AI-enabled, AI-ready, and AI-optimized devices (PCs, workstations, smartphones, tablets), infrastructure (server, storage, edge, high performance computing and software defined infrastructure), software, solutions, and services. Lenovo's continued investment in world-changing innovation is building a more equitable, trustworthy, and smarter future for everyone, everywhere. Lenovo is listed on the Hong Kong stock exchange under Lenovo Group Limited (HKSE: 992) (ADR: LNVGY).This transformation together with Lenovo's world-changing innovation is building a more inclusive, trustworthy, and smarter future for everyone, everywhere. To find out more visit , and read about the latest news via our .Description and RequirementsStaff Researcher, Large Model On-Device Inference (C++)Job Description:
  • Responsible for the frontier on device inference algorithm investigation and implementation.
  • Responsible for LLM inference optimization on device
Requirements:
  • Master's degree or higher in Computer Science, Networking, Communications, or related fields
  • Familiar with operating system principles, with product development experience
  • Familiar in C/C++ programming and knowledgeable in common scripting languages such as Shell and Python
  • Familiar with large models, including the model components, the acceleration principles, the graphs and so on
  • Prior experience op optimization such as CUDA or SYCL is a plus
  • Hands-on development experience with on device inference engine is a plus
Additional Locations:
  • China - Beijing - 北京(Beijing)
  • China
  • China - Beijing
  • China - Beijing - 北京(Beijing)

Lenovo