- 👩 I’m Sheng, a PhD student from China, currently studying as a visiting student at the National University of Singapore.
- 🧐 My focus is multimedia learning, especially VQA, and I’m currently exploring multimodal LLMs.
- 💬 As an ENFJ-A, I thrive on meaningful collaboration and communication.
- 📫 You can reach me at hzgn97@gmail.com—let’s connect!
🐢
Focusing
VQA & MLLM.
-
Hefei University of Technology
- China
-
11:41
(UTC +08:00) - https://zhousheng97.github.io/
Pinned Loading
-
EgoTextVQA
EgoTextVQA Public[CVPR 2025] 🌟🌟 EgoTextVQA: Towards Egocentric Scene-Text Aware Video Question Answering
Python 24
-
Awesome-MLLM-TextVQA
Awesome-MLLM-TextVQA Public✨✨Latest Research on Multimodal Large Language Models on Scene-Text VQA Tasks
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.