Liangzhi Li concurrently serves as Chief Scientist at Xiamen Meet You Co., Ltd. (美柚), Co-Founder & CTO of Climind, and Professor in the Department of Computer Science at Qufu Normal University (曲阜师范大学). Previously, he was an Assistant Professor at Osaka University (大阪大学) from 2021 to 2023, following a postdoctoral tenure there from 2019 to 2021.
He earned his Ph.D. in Engineering from the Muroran Institute of Technology (室蘭工業大学), Japan, in 2019, after completing both his B.S. and M.S. degrees in Computer Science at South China University of Technology (华南理工大学, SCUT) in 2012 and 2016, respectively.
His research interests encompass computer vision, explainable AI, large language models (LLMs), and medical imaging. His academic accomplishments have garnered recognition through programs such as the Shandong Taishan Scholar Young Expert (泰山学者青年专家) and the Shandong Provincial Overseas Excellent Young Scholars (山东省海外优青).
By bridging industry and academia, he remains committed to fostering technological innovation, advancing artificial intelligence research, and contributing to the educational sphere.
Interpreting and explaining the behavior of deep neural networks is critical for many tasks. Explainable AI provides a way to address this challenge, mostly by providing per-pixel relevance to the decision. Yet, interpreting such explanations may require expert knowledge. Some recent attempts toward interpretability adopt a concept-based framework, giving a higher-level relationship between some concepts and model decisions. This paper proposes Bottleneck Concept Learner (BotCL), which represents an image solely by the presence/absence of concepts learned through training over the target task without explicit supervision over the concepts. It uses self-supervision and tailored regularizers so that learned concepts can be human-understandable. Using some image classification tasks as our testbed, we demonstrate BotCL’s potential to rebuild neural networks for better interpretability.
Explainable artificial intelligence has been gaining attention in the past few years. However, most existing methods are based on gradients or intermediate features, which are not directly involved in the decision-making process of the classifier. In this paper, we propose a slot attention-based classifier called SCOUTER for transparent yet accurate classification. Two major differences from other attention-based methods include: (a) SCOUTER’s explanation is involved in the final confidence for each category, offering more intuitive interpretation, and (b) all the categories have their corresponding positive or negative explanation, which tells “why the image is of a certain category” or “why the image is not of a certain category.” We design a new loss tailored for SCOUTER that controls the model’s behavior to switch between positive and negative explanations, as well as the size of explanatory regions. Experimental results show that SCOUTER can give better visual explanations while keeping good accuracy on small and medium-sized datasets.
Retinal vessel segmentation is of great interest for diagnosis of retinal vascular diseases. To further improve the performance of vessel segmentation, we propose IterNet, a new model based on UNet, with the ability to find obscured details of the vessel from the segmented vessel image itself, rather than the raw input image. IterNet consists of multiple iterations of a mini-UNet, which can be 4× deeper than the common UNet. IterNet also adopts the weight-sharing and skip-connection features to facilitate training; therefore, even with such a large architecture, IterNet can still learn from merely 10∼20 labeled images, without pre-training or any prior knowledge. IterNet achieves AUCs of 0.9816, 0.9851, and 0.9881 on three mainstream datasets, namely DRIVE, CHASE-DB1, and STARE, respectively, which currently are the best scores in the literature. The source code is available.