Re-se-arch
Our re-se-arch has been generously supported by ARO, NSF, ARFL, IARPA, BlueHalo and Salesforce.
Our current re-se-arch interests mainly focus on:
(i) Grammar-Guided Interpretable and Robust Representation Learning. This line of research is motivated by “the belief that thinking of all kinds requires grammars” and “Grammar in language is merely a recent extension of much older grammars that are built into the brains of all intelligent animals to analyze sensory input, to structure their actions and even formulate their thoughts.” — Professor David Mumford.
(ii) Deep Consensus Lifelong Learning for Joint Discriminative and Generative Modeling. The world is highly structural with complex compositional regularities. To facilitate developing a unified AI ALTERing (Ask, Learn, Test, Explain and Refine) framework, on top of the research in (i), this line of research is to address one grand challenge in computer vision and machine (deep) learning, that is to model and learn the joint distribution of Grammar-like structures and raw data, p(structures, data), in a principled way. It typically consists of two tasks: structured output prediction that aims to learn p(structures | data) (e.g. image semantic segmentation or image parsing), and structured input synthesis that aims to learn p(data | structures), i.e., controllable and reconfigurable conditional generative learning (e.g., text/layout-to-image synthesis), or AIGC emerged more recently. Deep consensus lifelong learning aims to integrate them in a closed loop for AI ALTERing and AIGCGT (AI Generated Content and Ground-Truth).
2022
Learning Patch-to-Cluster Attention in Vision Transformer Working paper
arXiv preprint, 2022.
2016
arXiv preprint, 2016.