TY - GEN
T1 - Representation interpretation with spatial encoding and multimodal analytics
AU - Liu, Ninghao
AU - Du, Mengnan
AU - Hu, Xia
N1 - Publisher Copyright:
© 2019 Association for Computing Machinery.
PY - 2019/1/30
Y1 - 2019/1/30
N2 - Representation learning models map data instances into a low-dimensional vector space, thus facilitating the deployment of subsequent models such as classification and clustering models, or the implementation of downstream applications such as recommendation and anomaly detection. However, the outcome of representation learning is difficult to be directly understood by users, since each dimension of the latent space may not have any specific meaning. Understanding representation learning could be beneficial to many applications. For example, in recommender systems, knowing why a user instance is mapped to a certain position in the latent space may unveil the user's interests and profile. In this paper, we propose an interpretation framework to understand and describe how representation vectors distribute in the latent space. Specifically, we design a coding scheme to transform representation instances into spatial codes to indicate their locations in the latent space. Following that, a multimodal autoencoder is built for generating the description of a representation instance given its spatial codes. The coding scheme enables indication of position with different granularity. The incorporation of autoencoder makes the framework capable of dealing with different types of data. Several metrics are designed to evaluate interpretation results. Experiments under various application scenarios and different representation learning models are conducted to demonstrate the flexibility and effectiveness of the proposed framework.
AB - Representation learning models map data instances into a low-dimensional vector space, thus facilitating the deployment of subsequent models such as classification and clustering models, or the implementation of downstream applications such as recommendation and anomaly detection. However, the outcome of representation learning is difficult to be directly understood by users, since each dimension of the latent space may not have any specific meaning. Understanding representation learning could be beneficial to many applications. For example, in recommender systems, knowing why a user instance is mapped to a certain position in the latent space may unveil the user's interests and profile. In this paper, we propose an interpretation framework to understand and describe how representation vectors distribute in the latent space. Specifically, we design a coding scheme to transform representation instances into spatial codes to indicate their locations in the latent space. Following that, a multimodal autoencoder is built for generating the description of a representation instance given its spatial codes. The coding scheme enables indication of position with different granularity. The incorporation of autoencoder makes the framework capable of dealing with different types of data. Several metrics are designed to evaluate interpretation results. Experiments under various application scenarios and different representation learning models are conducted to demonstrate the flexibility and effectiveness of the proposed framework.
KW - Interpretation
KW - Recommender systems
KW - Representation learning
UR - http://www.scopus.com/inward/record.url?scp=85061735902&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85061735902&partnerID=8YFLogxK
U2 - 10.1145/3289600.3290960
DO - 10.1145/3289600.3290960
M3 - Conference contribution
AN - SCOPUS:85061735902
T3 - WSDM 2019 - Proceedings of the 12th ACM International Conference on Web Search and Data Mining
SP - 60
EP - 68
BT - WSDM 2019 - Proceedings of the 12th ACM International Conference on Web Search and Data Mining
PB - Association for Computing Machinery, Inc
T2 - 12th ACM International Conference on Web Search and Data Mining, WSDM 2019
Y2 - 11 February 2019 through 15 February 2019
ER -