Hi, I'm

Zhishuo

Multimodal AI researcher focused on emotion understanding, LLM systems, and practical intelligent products.

Multimodal Affective Computing LLM + MoE Systems Applied AI Productization

About

I am a Ph.D. candidate at the School of Computer Science, Sichuan University. My research centers on multimodal learning, affective computing, agent systems, large models, and robust speech understanding under real-world conditions. I work on multimodal sentiment analysis, cross-modal contrastive optimization, noise-resilient speech recognition, and emotion-aware agent workflows, with an emphasis on interpretability and generalization.

Research Focus

Core Research

Multimodal Emotion Understanding

Modeling emotion from speech, vision, and text with depth-aware representations.

Systems

LLM + MoE Systems

Task-adaptive routing and efficient expert collaboration for better generalization.

Impact

Applied AI Products

Bridging research and deployment through reliable workflows and automation.

Selected Projects

Research

HEME

Hierarchical emotion modeling with adaptive multi-level mixture-of-experts.

Platform

Emotion Agent Stack

End-to-end pipeline for multimodal emotion analysis and conversational AI.

Workflow

AV-RISE

Hierarchical cross-modal denoising for robust audio-visual speech representation under noisy real-world conditions.

Tooling

Paper-to-Product Toolkit

Templates and scripts to accelerate turning academic ideas into usable demos.

News

Resume

Open CV

Blog

Open blog

Contact

Available for collaboration

Email

zhiishuo@163.com

Usually within 24–48 hours.