Research
My research interests lie in Computer Vision and Machine Learning. I'm particularly interested in hand-object interaction under the First-Person perspective and aim to leverage Egocentric Vision to model daily activities and social behaviors to help understand the interaction between humans and their surrounding world.
|
News
π½π½π½π½π½π½π½
04.2024   Passed my Qualifier!π§ I will transfer to UIUC in Fall 24. π½π½π½π§π½π½π½
π½π½π½π½π½π½π½
02.2024   One paper accepted by CVPR '24! Check out the AV-CONV!! ππ€ΉββοΈ
01.2023   I'll be interning at Meta Reality Labs in summer 2023! βΎοΈπ©βπ»
08.2022   One journal paper accepted by TNNLS! π
07.2022   One paper accepted by ECCV '22!! πβ‘οΈ
04.2022   I've accepted my CS PhD offer at Georgia Tech. πππ
03.2022   Our Ego4D paper has been accepted by CVPR '22 as an oral paper! π₯
|
|
|
Listen to Look into the Future: Audio-Visual Egocentric Gaze Anticipation
Bolin Lai,
Fiona Ryan,
Wenqi Jia,
Miao Liu†,
James M. Rehg†
ECCV, 2024
project /
paper /
code /
data Split /
supplementary /
video /
poster
|
|
[New!] The Audio-Visual Conversational Graph: From an Egocentric-Exocentric Perspective
Wenqi Jia,
Miao Liu,
Hao Jiang,
Ishwarya Ananthabhotla,
James Rehg†,
Vamsi Krishna Ithapu†,
Ruohan Gao†
CVPR, 2024
project /
paper /
code /
dataset (under internal review) /
bibtex
(⇐ Move your cursor on the image for a short demo video!)
|
|
Werewolf Among Us: A Multimodal Dataset for Modeling Persuasion Behaviors in Social Deduction Games
Bolin Lai,
Hongxin Zhang,
Miao Liu,
Aryan Pariani,
Fiona Ryan,
Wenqi Jia,
James Rehg†,
Diyi Yang†
ACL Findings, 2023
project /
paper /
bibtex
|
|
Generative Adversarial Network for Future Hand Segmentation from Egocentric Video
Wenqi Jia,
Miao Liu,
James Rehg†
ECCV, 2022
project /
paper / 
code / 
supplement /
poster /
video / 
bibtex
(⇐ Move your cursor on the image for a short demo video!)
|
|
Ego4D: Around the World in 3,000 Hours of Egocentric Video
Kristen Grauman, et al.
CVPR, 2022   (Oral)
project /
paper /
code /
dataset /
video /
bibtex
|
|
Paying More Attention to Motion:
Attention Distillation for Learning Video Representations
Miao Liu,
Wenqi Jia,
Xin Chen,
Yun Zhang,
Yin Li,
James Rehg†
IJCV, 2021   (Special Issue) (under review)
arXiv /
code /
bibtex
|
|
Holistic-Guided Disentangled Learning With Cross-Video Semantics Mining for Concurrent
First-Person and Third-Person Activity Recognition
Tianshan Liu,
Wenqi Jia,
Rui Zhao,
Kin-Man Lam†,
Jun Kong
TNNLS, 2022
paper /
bibtex
|
|
Feature Redundancy Mining: Deep Light-Weight Image Super-Resolution Model
Jun Xiao,
Wenqi Jia,
Kin-Man Lam†
ICASSP, 2021
paper /
bibtex
|
|
Deep progressive convolutional neural network for blind super-resolution with multiple degradations
Jun Xiao,
Rui Zhao,
Shun-Cheung Lai,
Wenqi Jia,
Kin-Man Lam†
ICIP, 2019
paper /
bibtex
View extrapolation with multiplane images works better if you reason about disocclusions and disparity sampling frequencies.
|
Academic Service and Honor
- Reviewer π, CVPR '22, '23, '24, ECCV '22, '24, BMVC '24
- First Runner Up π₯, IEEE Video and Image Processing Cup (VIP Cup) 2017
|
Volunteer Work
- Certified First Responder π, Hong Kong Red Cross, Jun 2015 - Jun 2019
- General Secretary π€ΉββοΈ, Exploring Hong Kong Community (EHKC), Jan 2017 - Jan 2018
- International Volunteer π€, Service Learning Project in Cambodia, Jun 2016
|
|