TY - JOUR
T1 - Learning Smooth Representation for Unsupervised Domain Adaptation
AU - Cai, Guanyu
AU - He, Lianghua
AU - Zhou, Mengchu
AU - Alhumade, Hesham
AU - Hu, Die
N1 - Funding Information:
This work was supported in part by the National Key Research and Development Program of China under Grant 2020YFA0711400; in part by the Joint Funds of the National Science Foundation of China under Grant U18092006; in part by the Shanghai Municipal Science and Technology Committee of Shanghai Outstanding Academic Leaders Plan under Grant 19XD1434000; in part by the Projects of International Cooperation of Shanghai Municipal Science and Technology Committee under Grant 19490712800; in part by the National Natural Science Foundation of China under Grant 61772369, Grant 61773166, and Grant 61771144; in part by the Shanghai Municipal Science and Technology Major Project under Grant 2021SHZDZX0100; in part by the Shanghai Municipal Commission of Science and Technology Project under Grant 19511132101; in part by the Changjiang Scholars Program of China; in part by the Fundamental Research Funds for the Central Universities; and in part by the Institutional Fund Projects from the Ministry of Education and King Abdulaziz University, Jeddah, Saudi Arabia, under Grant IFPNC-001-135-2020.
Publisher Copyright:
© 2021 IEEE.
PY - 2023/8/1
Y1 - 2023/8/1
N2 - Typical adversarial-training-based unsupervised domain adaptation (UDA) methods are vulnerable when the source and target datasets are highly complex or exhibit a large discrepancy between their data distributions. Recently, several Lipschitz-constraint-based methods have been explored. The satisfaction of Lipschitz continuity guarantees a remarkable performance on a target domain. However, they lack a mathematical analysis of why a Lipschitz constraint is beneficial to UDA and usually perform poorly on large-scale datasets. In this article, we take the principle of utilizing a Lipschitz constraint further by discussing how it affects the error bound of UDA. A connection between them is built, and an illustration of how Lipschitzness reduces the error bound is presented. A local smooth discrepancy is defined to measure the Lipschitzness of a target distribution in a pointwise way. When constructing a deep end-to-end model, to ensure the effectiveness and stability of UDA, three critical factors are considered in our proposed optimization strategy, i.e., the sample amount of a target domain, dimension, and batchsize of samples. Experimental results demonstrate that our model performs well on several standard benchmarks. Our ablation study shows that the sample amount of a target domain, the dimension, and batchsize of samples, indeed, greatly impact Lipschitz-constraint-based methods' ability to handle large-scale datasets.
AB - Typical adversarial-training-based unsupervised domain adaptation (UDA) methods are vulnerable when the source and target datasets are highly complex or exhibit a large discrepancy between their data distributions. Recently, several Lipschitz-constraint-based methods have been explored. The satisfaction of Lipschitz continuity guarantees a remarkable performance on a target domain. However, they lack a mathematical analysis of why a Lipschitz constraint is beneficial to UDA and usually perform poorly on large-scale datasets. In this article, we take the principle of utilizing a Lipschitz constraint further by discussing how it affects the error bound of UDA. A connection between them is built, and an illustration of how Lipschitzness reduces the error bound is presented. A local smooth discrepancy is defined to measure the Lipschitzness of a target distribution in a pointwise way. When constructing a deep end-to-end model, to ensure the effectiveness and stability of UDA, three critical factors are considered in our proposed optimization strategy, i.e., the sample amount of a target domain, dimension, and batchsize of samples. Experimental results demonstrate that our model performs well on several standard benchmarks. Our ablation study shows that the sample amount of a target domain, the dimension, and batchsize of samples, indeed, greatly impact Lipschitz-constraint-based methods' ability to handle large-scale datasets.
KW - Lipschitz constraint
KW - local smooth discrepancy
KW - transfer learning
KW - unsupervised domain adaptation (UDA)
UR - http://www.scopus.com/inward/record.url?scp=85147549805&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85147549805&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2021.3119889
DO - 10.1109/TNNLS.2021.3119889
M3 - Article
C2 - 34788221
AN - SCOPUS:85147549805
SN - 2162-237X
VL - 34
SP - 4181
EP - 4195
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 8
ER -