About me
Wei Han is now a research scientist. His research interests now focus on a broad range of topics regarding multimodal large language models (LLMs), mainly focus on the following topics with specific research questions:
- Efficient Data-centric Alignment of LLMs
- What makes reasoning data βhigh-qualityβ?
- Can we develop automatic methods to curate or distill alignment data?
- How can we balance correctness and efficiency in reasoning output?
- Large Reasoning Models
- How to controll the Depth and Style of Reasoning?
- How to preserve reasoning consistency and faithfulness and avoid reasoning path that are logically flawed or disconnected from the final answer?
- Test-Time Control over Reasoning Behavior
- Agentic LLMs
- Long-Horizon planning and decomposition
- Memory and context management
- Tool Use and Decision Making
Other interested directions or those I have worked on include retrive-augmented generation (RAG), long-context modeling, video understanding, multimodal representation learning and semi-supervised learning (SSL). For more information, you can check my CV here.
Collaborations are always welcome! If you are looking forward to research experience or trying academic projects with me, feel free to drop me an email (better to attach your CV or personal website so that I can get to know you well).
I am actively seeking research scientist/engineer positions starting in 2025 based in Singapore/China/U.S (no particular preference). Feel free to reach out to me if you are interested in my profile and my skill set meets the position requirements!
News
π₯π₯π₯ [2025.01] Our work on multimodal recommendation has been accepeted to NAACL 2025. π π» code β π paper
π₯π₯π₯ [2024.10] We release the latest work on long-context modeling, in which we propose a new architecture for efficient post-training and finetuning. π π» code β π paper
[2024.03] Our work on efficient sampling in video question answering has been accepeted to NAACL 2024. π π» code β π paper
[2024.02] We have released the code of Auto Scaling, the first framework for automatic instruction scaling in (multimodal) instruction fine-tuning. π π» code β π paper
[2022.10] Two papers are accpeted to EMNLP and Findings of EMNLP 2022!