Menu

Internship Opportunities Available

I am currently seeking motivated PhD students for Summer 2026 internships at NEC Labs America San Jose. If you are working on Vision-Language Models (VLMs) or Vision-Language-Action Models (VLAs), I would love to hear from you. Please contact me at zhenyue.qin at nec-labs dot com.

Contact Me

Set-of-Vision Prompting

This paper introduces Set-of-Vision (SoV) prompting, which enhances emotion recognition in Vision Large Language Models by using spatial visual cues like bounding boxes, numbers, and facial landmarks to precisely identify and analyze facial expressions while preserving image context.

Read More
Scroll

A Survey Paper Accepted by EACL 2026

Our survey paper on plane geometry problem solving with multi-modal reasoning has been accepted by EACL Findings 2026. I'm grateful to have collaborated with an amazing team: Seunghyuk, Yang, Youngbin, Seungbeom, and Dongwoo.

Read More

DermEVAL Accepted by WACV 2026

Our paper on evaluating multimodal large language models for dermatology has been accepted by WACV 2026. I'm grateful to have collaborated with an amazing team: Hongjin, Weihao, Gepeng, Yang, Tom, and Nick.

Read More

GeoDANO Accepted by EMNLP 2025

Our paper on geometry problem solving with large vision-language models has been accepted by EMNLP Findings 2025. I'm grateful to have collaborated with an amazing team: Seunghyuk, Yang, Youngbin, Seungbeom, and Dongwoo.

Read More

LMOD Accepted by NAACL 2025

Our paper on multimodal large models and ophthalmology benchmarks has been accepted by NAACL Findings 2025. I'm grateful to have collaborated with an amazing team: Yu, Dylan, Xuansheng, Ke, Yih-Chung, Ninghao, Xiuzhen, and Qingyu.

Read More