
Staff Researcher, Large Model On-Device Inference (C++)
- Beijing
- Permanent
- Full-time
- China - Beijing - 北京(Beijing)
- Responsible for the frontier on device inference algorithm investigation and implementation.
- Responsible for LLM inference optimization on device
- Master's degree or higher in Computer Science, Networking, Communications, or related fields
- Familiar with operating system principles, with product development experience
- Familiar in C/C++ programming and knowledgeable in common scripting languages such as Shell and Python
- Familiar with large models, including the model components, the acceleration principles, the graphs and so on
- Prior experience op optimization such as CUDA or SYCL is a plus
- Hands-on development experience with on device inference engine is a plus
- China - Beijing - 北京(Beijing)
- China
- China - Beijing
- China - Beijing - 北京(Beijing)