We are seeking a highly motivated and talented Applied Researcher Intern to join our team in Shanghai and contribute to exciting projects involving the integration of Large Language Models (LLMs) with multimodal data. This is an opportunity to work alongside experienced researchers and engineers on cutting-edge challenges in the e-commerce domain, applying the latest research to real-world problems and potentially contributing to publications or product features.
As an Applied Researcher Intern focusing on LLMs and Multimodality, you will delve into novel techniques, architectures, and applications in areas such as multimodal understanding of product listings (combining text and images), leveraging LLMs for enhanced search and discovery experiences, natural language generation grounded in visual information, and exploring novel multimodal reasoning tasks relevant to e-commerce. You will be responsible for researching, implementing, experimenting, and analyzing the performance of state-of-the-art LLM and multimodal techniques.
Responsibilities:
1. Research and stay up-to-date with the latest advancements in LLMs, Multimodal AI, NLP, Computer Vision, and related fields.
2. Design and implement novel LLM or multimodal architectures, training techniques, or applications for improving multimodal search and retrieval, enhancing product understanding from images and text, or generating rich, multimodal content.
3. Conduct experiments to evaluate the performance of developed models and techniques on relevant datasets and tasks.
4. Analyze experimental results, identify insights, and propose improvements.
5. Write clean, efficient, and well-documented code using relevant programming languages and frameworks.
6. Collaborate closely with team members to brainstorm ideas, share findings, and integrate work.
7. Document research findings and progress in reports and presentations.
8. Potentially contribute to internal or external publications, patents, or product development.