A journal of IEEE and CAA , publishes high-quality papers in English on original theoretical/experimental research and development in all areas of automation
Volume 11 Issue 7
Jul.  2024

IEEE/CAA Journal of Automatica Sinica

  • JCR Impact Factor: 15.3, Top 1 (SCI Q1)
    CiteScore: 23.5, Top 2% (Q1)
    Google Scholar h5-index: 77, TOP 5
Turn off MathJax
Article Contents
B. Xu, J. Yin, C. Lian, Y. Su, and  Z. Zeng,  “Low-rank optimal transport for robust domain adaptation,” IEEE/CAA J. Autom. Sinica, vol. 11, no. 7, pp. 1667–1680, Jul. 2024. doi: 10.1109/JAS.2024.124344
Citation: B. Xu, J. Yin, C. Lian, Y. Su, and  Z. Zeng,  “Low-rank optimal transport for robust domain adaptation,” IEEE/CAA J. Autom. Sinica, vol. 11, no. 7, pp. 1667–1680, Jul. 2024. doi: 10.1109/JAS.2024.124344

Low-Rank Optimal Transport for Robust Domain Adaptation

doi: 10.1109/JAS.2024.124344
Funds:  This work was supported by the National Natural Science Foundation of China (62206204, 62176193), the Natural Science Foundation of Hubei Province, China (2023AFB705), and the Natural Science Foundation of Chongqing, China (CSTB2023NSCQ-MSX0932)
More Information
  • When encountering the distribution shift between the source (training) and target (test) domains, domain adaptation attempts to adjust the classifiers to be capable of dealing with different domains. Previous domain adaptation research has achieved a lot of success both in theory and practice under the assumption that all the examples in the source domain are well-labeled and of high quality. However, the methods consistently lose robustness in noisy settings where data from the source domain have corrupted labels or features which is common in reality. Therefore, robust domain adaptation has been introduced to deal with such problems. In this paper, we attempt to solve two interrelated problems with robust domain adaptation: distribution shift across domains and sample noises of the source domain. To disentangle these challenges, an optimal transport approach with low-rank constraints is applied to guide the domain adaptation model training process to avoid noisy information influence. For the domain shift problem, the optimal transport mechanism can learn the joint data representations between the source and target domains using a measurement of discrepancy and preserve the discriminative information. The rank constraint on the transport matrix can help recover the corrupted subspace structures and eliminate the noise to some extent when dealing with corrupted source data. The solution to this relaxed and regularized optimal transport framework is a convex optimization problem that can be solved using the Augmented Lagrange Multiplier method, whose convergence can be mathematically proved. The effectiveness of the proposed method is evaluated through extensive experiments on both synthetic and real-world datasets.

     

  • loading
  • [1]
    D. Wu and X. Luo, “Robust latent factor analysis for precise representation of high-dimensional and sparse data,” IEEE/CAA J. Autom. Sinica, vol. 8, no. 4, pp. 796–805, Apr. 2021. doi: 10.1109/JAS.2020.1003533
    [2]
    B. Xu, Q. Liu, and T. Huang, “A discrete-time projection neural network for sparse signal reconstruction with application to face recognition,” IEEE Trans. Neural Netw. Learn. Syst., vol. 30, no. 1, pp. 151–162, Jan. 2019. doi: 10.1109/TNNLS.2018.2836933
    [3]
    J. Yin, D. Shen, X. Du, and L. Li, “Distributed stochastic model predictive control with Taguchi’s robustness for vehicle platooning,” IEEE Trans. Intell. Transp. Syst., vol. 23, no. 9, pp. 15967–15979, Sep. 2022. doi: 10.1109/TITS.2022.3146715
    [4]
    E. Tzeng, J. Hoffman, T. Darrell, and K. Saenko, “Simultaneous deep transfer across domains and tasks,” in Proc. IEEE Int. Conf. Computer Vision, Santiago, Chile, 2015, pp. 4068–4076.
    [5]
    C. Peng and J. Ma, “Domain adaptive semantic segmentation via entropy-ranking and uncertain learning-based self-training,” IEEE/CAA J. Autom. Sinica, vol. 9, no. 8, pp. 1524–1527, Aug. 2022. doi: 10.1109/JAS.2022.105767
    [6]
    Y. Wang, S. Qiu, D. Li, C. Du, B.-L. Lu, and H. He, “Multi-modal domain adaptation variational autoencoder for EEG-based emotion recognition,” IEEE/CAA J. Autom. Sinica, vol. 9, no. 9, pp. 1612–1626, Sep. 2022. doi: 10.1109/JAS.2022.105515
    [7]
    H. Guan and M. Liu, “Domain adaptation for medical image analysis: A survey,” IEEE Trans. Biomed. Eng., vol. 69, no. 3, pp. 1173–1185, Mar. 2022. doi: 10.1109/TBME.2021.3117407
    [8]
    J. Yin, Z. Hu, Z. P. Mourelatos, D. Gorsich, A. Singh, and S. Tau, “Efficient reliability-based path planning of off-road autonomous ground vehicles through the coupling of surrogate modeling and RRT,” IEEE Trans. Intell. Transp. Syst., vol. 24, no. 12, pp. 15035–15050, Dec. 2023. doi: 10.1109/TITS.2023.3296651
    [9]
    J. Yin, L. Li, Z. P. Mourelatos, Y. Liu, D. Gorsich, A. Singh, S. Tau, and Z. Hu, “Reliable global path planning of off-road autonomous ground vehicles under uncertain terrain conditions,” IEEE Trans. Intell. Veh., vol. 9, no. 1, pp. 1161–1174, Jan. 2024. doi: 10.1109/TIV.2023.3317833
    [10]
    Y. Shao, L. Li, W. Ren, C. Gao, and N. Sang, “Domain adaptation for image dehazing,” in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, Seattle, USA, 2020, pp. 2805–2814.
    [11]
    B. Xu, Z. Zeng, C. Lian, and Z. Ding, “Few-shot domain adaptation via mixup optimal transport,” IEEE Trans. Image Process., vol. 31, pp. 2518–2528, Mar. 2022. doi: 10.1109/TIP.2022.3157139
    [12]
    K. Zhang, V. W. Zheng, Q. Wang, J. T.-Y. Kwok, Q. Yang, and I. Marsic, “Covariate shift in Hilbert space: A solution via sorrogate kernels,” in Proc. 30th Int. Conf. Machine Learning, Atlanta, USA, 2013, pp. 388–395.
    [13]
    C. Villani, Optimal Transport: Old and New. Berlin, Heidelberg, Germany: Springer, 2009.
    [14]
    A. Akbari, M. Awais, S. Fatemifar, S. S. Khalid, and J. Kittler, “A novel ground metric for optimal transport-based chronological age estimation,” IEEE Trans. Cybern., vol. 52, no. 10, pp. 9986–9999, Oct. 2022. doi: 10.1109/TCYB.2021.3083245
    [15]
    J. Qian, W. K. Wong, H. Zhang, J. Xie, and J. Yang, “Joint optimal transport with convex regularization for robust image classification,” IEEE Trans. Cybern., vol. 52, no. 3, pp. 1553–1564, Mar. 2022. doi: 10.1109/TCYB.2020.2991219
    [16]
    N. Courty, R. Flamary, A. Habrard, and A. Rakotomamonjy, “Joint distribution optimal transportation for domain adaptation,” in Proc. 31st Int. Conf. Neural Information Processing Systems, Long Beach, USA, 2017, pp. 3733–3742.
    [17]
    B. B. Damodaran, B. Kellenberger, R. Flamary, D. Tuia, and N. Courty, “DeepJDOT: Deep joint distribution optimal transport for unsupervised domain adaptation,” in Proc. 15th European Conf. Computer Vision, Munich, Germany, 2018, pp. 467–483.
    [18]
    J. Zhuo, S. Wang, and Q. Huang, “Uncertainty modeling for robust domain adaptation under noisy environments,” IEEE Trans. Multimedia, vol. 25, pp. 6157–6170, 2023. doi: 10.1109/TMM.2022.3205457
    [19]
    Y. Shu, Z. Cao, M. Long, and J. Wang, “Transferable curriculum for weakly-supervised domain adaptation,” in Proc. 33rd AAAI Conf. Artificial Intelligence, Honolulu, USA, 2019, pp. 4951–4958.
    [20]
    C. Zhang, S. Bengio, M. Hardt, B. Recht, and O. Vinyals, “Understanding deep learning (still) requires rethinking generalization,” Commun. ACM, vol. 64, no. 3, pp. 107–115, Mar. 2021. doi: 10.1145/3446776
    [21]
    F. Liu, J. Lu, B. Han, G. Niu, G. Zhang, and M. Sugiyama, “Butterfly: A panacea for all difficulties in wildly unsupervised domain adaptation,” arXiv preprint arXiv: 1905.07720, 2019.
    [22]
    X. Yu, T. Liu, M. Gong, K. Zhang, K. Batmanghelich, and D. Tao, “Label-noise robust domain adaptation,” in Proc. 37th Int. Conf. Machine Learning, 2020, pp. 10913–10924.
    [23]
    A. Chaddad, Q. Lu, J. Li, Y. Katib, R. Kateb, C. Tanougast, A. Bouridane, and A. Abdulkadir, “Explainable, domain-adaptive, and federated artificial intelligence in medicine,” IEEE/CAA J. Autom. Sinica, vol. 10, no. 4, pp. 859–876, Apr. 2023. doi: 10.1109/JAS.2023.123123
    [24]
    M. Long, Y. Cao, J. Wang, and M. I. Jordan, “Learning transferable features with deep adaptation networks,” in Proc. 32nd Int. Conf. Machine Learning, Lille, France, 2015, pp. 97–105.
    [25]
    M. Long, H. Zhu, J. Wang, and M. I. Jordan, “Deep transfer learning with joint adaptation networks,” in Proc. 34th Int. Conf. Machine Learning, Sydney, Australia, 2017, pp. 2208–2217.
    [26]
    I.-H. Jhuo, D. Liu, D. T. Lee, and S.-F. Chang, “Robust visual domain adaptation with low-rank reconstruction,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Providence, USA, 2012, pp. 2168–2175.
    [27]
    W. Zhang, L. Deng, L. Zhang, and D. Wu, “A survey on negative transfer,” IEEE/CAA J. Autom. Sinica, vol. 10, no. 2, pp. 305–329, Feb. 2023. doi: 10.1109/JAS.2022.106004
    [28]
    H. Yan, Y. Ding, P. Li, Q. Wang, Y. Xu, and W. Zuo, “Mind the class weight bias: Weighted maximum mean discrepancy for unsupervised domain adaptation,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Honolulu, USA, 2017, pp. 945–954.
    [29]
    P. Haeusser, T. Frerix, A. Mordvintsev, and D. Cremers, “Associative domain adaptation,” in Proc. IEEE Int. Conf. Computer Vision, Venice, Italy, 2017, pp. 2784–2792.
    [30]
    Z. Zheng, L. Teng, W. Zhang, N. Wu, and S. Teng, “Knowledge transfer learning via dual density sampling for resource-limited domain adaptation,” IEEE/CAA J. Autom. Sinica, vol. 10, no. 12, pp. 2269–2291, Dec. 2023. doi: 10.1109/JAS.2023.123342
    [31]
    I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” Commun. ACM, vol. 63, no. 11, pp. 139–144, Nov. 2020. doi: 10.1145/3422622
    [32]
    E. Tzeng, C. Devin, J. Hoffman, C. Finn, P. Abbeel, S. Levine, K. Saenko, and T. Darrell, “Adapting deep visuomotor representations with weak pairwise constraints,” in Algorithmic Foundations of Robotics XⅡ: Proceedings of the Twelfth Workshop on the Algorithmic Foundations of Robotics, K. Goldberg, P. Abbeel, K. Bekris, and L. Miller, Eds. Cham, Germany: Springer, 2020, pp. 688–703.
    [33]
    K.-C. Peng, Z. Wu, and J. Ernst, “Zero-shot deep domain adaptation,” in Proc. 15th European Conf. Computer Vision, Munich, Germany, 2018, pp. 793–810.
    [34]
    Z. Murez, S. Kolouri, D. Kriegman, R. Ramamoorthi, and K. Kim, “Image to image translation for domain adaptation,” in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018, pp. 4500–4509.
    [35]
    M.-Y. Liu and O. Tuzel, “Coupled generative adversarial networks,” in Proc. 30th Int. Conf. Neural Information Processing Systems, Barcelona, Spain, 2016, pp. 469–477.
    [36]
    E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Honolulu, USA, 2017, pp. 2962–2971.
    [37]
    R. Wang, Z. Wu, Z. Weng, J. Chen, G.-J. Qi, and Y.-G. Jiang, “Cross-domain contrastive learning for unsupervised domain adaptation,” IEEE Trans. Multimedia, vol. 25, pp. 1665–1673, 2023. doi: 10.1109/TMM.2022.3146744
    [38]
    M. Thota and G. Leontidis, “Contrastive domain adaptation,” in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, Nashville, USA, 2021, pp. 2209–2218.
    [39]
    J. Yang, J. Liu, N. Xu, and J. Huang, “TVT: Transferable vision transformer for unsupervised domain adaptation,” in Proc. IEEE/CVF Winter Conf. Applications of Computer Vision, Waikoloa, USA, 2023, pp. 520–530.
    [40]
    M. Wang, J. Chen, Y. Wang, Z. Gong, K. Wu, and V. C. M. Leung, “TFC: Transformer fused convolution for adversarial domain adaptation,” IEEE Trans. Comput. Soc. Syst., vol. 11, no. 1, pp. 697–706, Feb. 2024. doi: 10.1109/TCSS.2022.3229693
    [41]
    T. Xiao, T. Xia, Y. Yang, C. Huang, and X. Wang, “Learning from massive noisy labeled data for image classification,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, Boston, USA, 2015, pp. 2691–2699.
    [42]
    X. Zeng, W. Ouyang, M. Wang, and X. Wang, “Deep learning of scene-specific classifier for pedestrian detection,” in Proc. 13th European Conf. Computer Vision, Zurich, Switzerland, 2014, pp. 472–487.
    [43]
    Z. Han, X.-J. Gui, C. Cui, and Y. Yin, “Towards accurate and robust domain adaptation under noisy environments,” in Proc. 29th Int. Joint Conf. Artificial Intelligence, 2020, pp. 2269–2276.
    [44]
    C. Villani, Topics in Optimal Transportation. Providence, USA: American Mathematical Society, 2021.
    [45]
    G. Monge, “Mémoire sur la théorie des déblais et des remblais,” Histoire de l’Académie Royale des Sciences, pp. 666–704, 1781.
    [46]
    L. V. Kantorovich, “On the translocation of masses,” J. Math. Sci., vol. 133, no. 4, pp. 1381–1382, Mar.s 2006. doi: 10.1007/s10958-006-0049-2
    [47]
    Y. Guo, X. Wang, C. Li, and S. Ying, “Domain adaptive semantic segmentation by optimal transport,” Fundam. Res., p. , 2023. doi: 10.1016/j.fmre.2023.06.006
    [48]
    C.-Y. Chuang, S. Jegelka, and D. Alvarez-Melis, “InfoOT: Information maximizing optimal transport,” in Proc. 40th Int. Conf. Machine Learning, Honolulu, USA, 2023, pp. 6228–6242.
    [49]
    L. Chapel, R. Flamary, H. Wu, C. Févotte, and G. Gasso, “Unbalanced optimal transport through non-negative penalized linear regression,” in Proc. 35th Conf. Neural Information Processing Systems, 2021, pp. 23270–23282.
    [50]
    I. Redko, N. Courty, R. Flamary, and D. Tuia, “Optimal transport for multi-source domain adaptation under target shift,” in Proc. 22nd Int. Conf. Artificial Intelligence and Statistics, Okinawa, Japan, 2019, pp. 849–858.
    [51]
    Y. Balaji, R. Chellappa, and S. Feizi, “Robust optimal transport with applications in generative modeling and domain adaptation,” in Proc. 34th Int. Conf. Neural Information Processing Systems, Vancouver, Canada, 2020, pp. 12934–12944.
    [52]
    S. Ferradans, N. Papadakis, G. Peyré, and J.-F. Aujol, “Regularized discrete optimal transport,” SIAM J. Imaging Sci., vol. 7, no. 3, pp. 1853–1882, Jan. 2014. doi: 10.1137/130929886
    [53]
    M. Blondel, V. Seguy, and A. Rolet, “Smooth and sparse optimal transport,” in Proc. 21st Int. Conf. Artificial Intelligence and Statistics, Playa Blanca, Lanzarote, Canary Islands, Spain, 2018, pp. 880–889.
    [54]
    N. Courty, R. Flamary, and D. Tuia, “Domain adaptation with regularized optimal transport,” in Proc. European Conf. Machine Learning and Knowledge Discovery in Databases, Nancy, France, 2014, pp. 274–289.
    [55]
    W. Lu, Y. Chen, J. Wang, and X. Qin, “Cross-domain activity recognition via substructural optimal transport,” Neurocomputing, vol. 454, pp. 65–75, Sep. 2021. doi: 10.1016/j.neucom.2021.04.124
    [56]
    M. Scetbon, M. Cuturi, and G. Peyré, “Low-rank sinkhorn factorization,” in Proc. 38th Int. Conf. Machine Learning, 2021, pp. 9344–9354.
    [57]
    C. Strössner and D. Kressner, “Low-rank tensor approximations for solving multi-marginal optimal transport problems,” arXiv preprint arXiv: 2202.07340, 2022.
    [58]
    K. Fatras, T. Séjourné, R. Flamary, and N. Courty, “Unbalanced minibatch optimal transport; applications to domain adaptation,” in Proc. 38th Int. Conf. Machine Learning, 2021, pp. 3186–3197.
    [59]
    W. Chang, Y. Shi, H. D. Tuan, and J. Wang, “Unified optimal transport framework for universal domain adaptation,” in Proc. 36th Int. Conf. Neural Information Processing Systems, New Orleans, USA, 2022, pp. 29512–29524.
    [60]
    M. Cuturi, “Sinkhorn distances: Lightspeed computation of optimal transport,” in Proc. 26th Int. Conf. Neural Information Processing Systems, Lake Tahoe, USA, 2013, pp. 2292–2300.
    [61]
    J.-F. Cai, E. J. Candès, and Z. Shen, “A singular value thresholding algorithm for matrix completion,” SIAM J. Optim., vol. 20, no. 4, pp. 1956–1982, Jan. 2010. doi: 10.1137/080738970
    [62]
    A. Beck and M. Teboulle, “A fast iterative shrinkage-thresholding algorithm for linear inverse problems,” SIAM J. Imaging Sci., vol. 2, no. 1, pp. 183–202, Jan. 2009. doi: 10.1137/080716542
    [63]
    Z. Lin, M. Chen, and Y. Ma, “The augmented Lagrange multiplier method for exact recovery of corrupted low-rank matrices,” arXiv preprint arXiv: 1009.5055, 2010.
    [64]
    Z. Lin, A. Ganesh, J. Wright, L. Wu, M. Chen, and Y. Ma, “Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix,” Coordinated Science Lab., Urbana, IL, USA, Tech. Rep. UILU-ENG-09-2214, Aug. 2009.
    [65]
    W. Li, L. Wang, W. Li, E. Agustsson, and L. Van Gool, “WebVision database: Visual learning and understanding from web data,” arXiv preprint arXiv: 1708.02862, 2017.
    [66]
    Z. Han, X.-J. Gui, H. Sun, Y. Yin, and S. Li, “Towards accurate and robust domain adaptation under multiple noisy environments,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 5, pp. 6460–6479, May 2023.
    [67]
    M. P. Kumar, B. Packer, and D. Koller, “Self-paced learning for latent variable models,” in Proc. 23rd Int. Conf. Neural Information Processing Systems, Vancouver, Canada, 2010, pp. 1189–1197.
    [68]
    Y. Ganin and V. S. Lempitsky, “Unsupervised domain adaptation by backpropagation,” in Proc. 32nd Int. Conf. Machine Learning, Lille, France, 2015, pp. 1180–1189.
    [69]
    M. Long, H. Zhu, J. Wang, and M. I. Jordan, “Unsupervised domain adaptation with residual transfer networks,” in Proc. 30th Int. Conf. Neural Information Processing Systems, Barcelona, Spain, 2016, pp. 136–144.
    [70]
    L. Jiang, Z. Zhou, T. Leung, L.-J. Li, and L. Fei-Fei, “MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels,” in Proc. 35th Int. Conf. Machine Learning, Stockholm, Sweden, 2018, pp. 2309–2318.
    [71]
    Y. Zhang, T. Liu, M. Long, and M. I. Jordan, “Bridging theory and algorithm for domain adaptation,” in Proc. 36th Int. Conf. Machine Learning, Long Beach, USA, 2019, pp. 7404–7413.

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(9)  / Tables(6)

    Article Metrics

    Article views (300) PDF downloads(35) Cited by()

    Highlights

    • A low-rank optimal transport algorithm is presented for the robust domain adaptation problem
    • Discrete formulation of optimal transport with low-rank constraints is solved by the Augmented Lagrange Multiplier method
    • The rank constraint on the transport matrix recovers the corrupted subspace structures and extracts the class structure information
    • The convergence of the low-rank regularized optimal transport algorithm is mathematically proved

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return