TY - JOUR
T1 - Global-Mapping-Consistency-Constrained Visual-Semantic Embedding for Interpreting Autonomous Perception Models
AU - Zhang, Chi
AU - Yuan, Meng
AU - Ma, Xiaoning
AU - Wei, Ping
AU - Su, Yuanqi
AU - Li, Li
AU - Liu, Yuehu
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2024
Y1 - 2024
N2 - From the perspective of artificial intelligence evaluation, the need to discover and explain the potential shortness of the evaluated intelligent algorithms/systems as well as the need to evaluate the intelligence level of such testees are of equal importance. In this paper, we propose a possible solution to these challenges: Explainable Evaluation for visual intelligence. Specifically, we focus on the problem setting where the internal mechanisms of AI algorithms are sophisticated, heterogeneous or unreachable. In this case, a latent attribute dictionary learning method with constrained by mapping consistency is proposed to explain the performance variation patterns of visual perception intelligence under different test samples. By jointly iteratively solving the learning of latent concept representation for test samples and the regression of latent concept-generalization performance, the mapping relationship between deep representation, semantic attribute annotation, and generalization performance of test samples is established to predict the degree of influence of semantic attributes on visual perception generalization performance. The optimal solution of proposed method could be reached via an alternating optimization process. Through quantitative experiments, we find that global mapping consistency constraints can make the learned latent concept representation strictly consistent with deep representation, thereby improving the accuracy of semantic attribute-perception performance correlation calculation.
AB - From the perspective of artificial intelligence evaluation, the need to discover and explain the potential shortness of the evaluated intelligent algorithms/systems as well as the need to evaluate the intelligence level of such testees are of equal importance. In this paper, we propose a possible solution to these challenges: Explainable Evaluation for visual intelligence. Specifically, we focus on the problem setting where the internal mechanisms of AI algorithms are sophisticated, heterogeneous or unreachable. In this case, a latent attribute dictionary learning method with constrained by mapping consistency is proposed to explain the performance variation patterns of visual perception intelligence under different test samples. By jointly iteratively solving the learning of latent concept representation for test samples and the regression of latent concept-generalization performance, the mapping relationship between deep representation, semantic attribute annotation, and generalization performance of test samples is established to predict the degree of influence of semantic attributes on visual perception generalization performance. The optimal solution of proposed method could be reached via an alternating optimization process. Through quantitative experiments, we find that global mapping consistency constraints can make the learned latent concept representation strictly consistent with deep representation, thereby improving the accuracy of semantic attribute-perception performance correlation calculation.
KW - Explainable AI evaluation
KW - dictionary learning
KW - latent knowledge representation
UR - https://www.scopus.com/pages/publications/85197074239
U2 - 10.1109/OJITS.2024.3418552
DO - 10.1109/OJITS.2024.3418552
M3 - 文章
AN - SCOPUS:85197074239
SN - 2687-7813
VL - 5
SP - 393
EP - 408
JO - IEEE Open Journal of Intelligent Transportation Systems
JF - IEEE Open Journal of Intelligent Transportation Systems
ER -