My name is Yuxiang Lin, pronounced as “You-shee-ahng Lin”. I am an MS student at Georgia Institute of Technology . I earned my B.S. degree from Shenzhen Technology University under supervision from Prof. Xiaojiang Peng. In 2023, I was a visiting student at the Shenzhen Institute of Advanced Technology (SIAT), CAS and a research intern with Prof. Chen Chen at UCF remotely. In 2024, I interned at Baidu Inc in the Group of Multimodal Retrieval, gaining experience in representation learning and big data.
Additionally, I volunteered as a Teaching Assistant for a Large Language Models/Computer Vision tutorial hosted by the Shanghai AI Laboratory .
You can find me at yuxiang.lin@gatech.edu or lin.yuxiang.contact@gmail.com.
My research interest mainly includes:
- Foundation model: Representation Learning, Post-Pretraining, Contrastive Learning.
- Multi-modal LLM: LLM Reasoning, LLM Application.
- AI4Science: LLM with Medical Analysis.
🔥 News
- 2024.07: One co-first author paper about invisible gas detection is accepted by CVIU (JCR Q1, CCF rank B). 🎉
- 2024.03: One paper about Conversational Emotion-Cause Pair Analysis with LLM is accepted by SemEval 2024, NAACL.
- 2024.01: I was awarded the First Prize of Research and Innovation Award (3000 CNY) and Star of Craftsmanship (3000 CNY).
- 2023.08: My instance segmentation tutorial has been featured in MMYOLO v0.6.0 highlight! Check out the tutorial here to master the essentials of instance segmentation.
- 2023.07: One paper on multimodal emotion recognition is accepted by ACM MM! 🎉
- 2023.07: We are the runner up in the Grand Challenge (MER 2023) of ACM MM! 🥈
📝 Publications
📌 Pinned
Invisible Gas Detection: An RGB-Thermal Cross Attention Network and A New Benchmark
Jue Wang*, Yuxiang Lin*, Qi Zhao, Dong Luo, Shuaibao Chen, Wei Chen, Xiaojiang Peng (* denotes equal contribution)
CVIU (JCR Q1, CCF-B) | [Paper] [Code]
- We design the RGB-Thermal Cross Attention Network for invisible gas detection, by effectively integrating texture information from RGB images and gas information from thermal images.
- We propose the RGB-assisted Cross Attention Module and the Global Textual Attention Module for cross modality feature fusion and diverse contextual information extraction.
- We introduce Gas-DB, the first comprehensive open-source gas detection database, including approximately 1.3K well-annotated RGB-thermal images for gas detection algorithms.
Semi-Supervised Multimodal Emotion Recognition with Expression MAE
Zebang Cheng, Yuxiang Lin, Zhaoru Chen, Xiang Li, Shuyi Mao, Fan Zhang, Daijun Ding, Bowen Zhang, Xiaojiang Peng
ACMMM 2023 (CCF-A) | [Paper] [Slides]
- We propose expMAE (combined with MAE and VideoMAE) to build an impressive emotion recognition classifier.
- We utilize multi-modal foundation model such as CLIP, MacBERT, and HuBERT to boost the ability of visual features from expMAE.
- We use the semi-supervised method of pseudo-labeling to solve the skewed distribution of the training set and finally achieve rank 2th among all the participants.
👨💻 Experience
- 2022.10 - 2024.06 Student Research Fellow, MIPS-Lab
- 2024.01 - 2024.04 Research Intern, Baidu Inc
- 2024.01 - 2024.01 Teaching Assistant, InternLM
- 2023.10 - 2024.01 Research Intern, University of Central Florida (UCF)
- 2023.06 - 2023.07 Teaching Assistant, OpenMMLab
- 2023.02 - 2023.07 Visiting Student, Shenzhen Institute of Advanced Technology, CAS
🏅 Selected Awards
- 2020 Second Prize of SZTU Freshman Scholarship (6000 CNY)
- 2022 China Undergraduate Mathematical Contest in Modeling, National Second Prize (top 2%)
- 2023 Dahua Outstanding Scholarship (4000 CNY)
- 2023 OpenMMLab MMSTAR I
- 2024 First Prize of Research and Innovation Award (3000 CNY)
- 2024 Star of Craftsmanship (3000 CNY)