skip to main content
10.1145/3488560.3502211acmconferencesArticle/Chapter ViewAbstractPublication PageswsdmConference Proceedingsconference-collections
extended-abstract
Public Access

Trustworthy Machine Learning: Fairness and Robustness

Published:15 February 2022Publication History

ABSTRACT

In recent years, machine learning (ML) technologies have experienced swift developments and attracted extensive attention from both academia and industry. The applications of ML are extended to multiple domains, from computer vision, text processing, to recommendations, etc. However, recent studies have uncovered the untrustworthy side of ML applications. For example, ML algorithms could show human-like discrimination against certain individuals or groups, or make unreliable decisions in safety-critical scenarios, which implies the absence of fairness and robustness, respectively. Consequently, building trustworthy machine learning systems has become an urgent need. My research strives to help meet this demand. In particular, my research focuses on designing trustworthy ML models and spans across three main areas: (1) fairness in ML, where we aim to detect, eliminate bias and ensure fairness in various ML applications; (2) robustness in ML, where we seek to ensure the robustness of certain ML applications towards adversarial attacks; (3) specific applications of ML, where my research involves the development of ML-based natural language processing (NLP) models and recommendation systems.

Skip Supplemental Material Section

Supplemental Material

WSDM22-dc01.mp4

mp4

88.6 MB

References

  1. Yonatan Belinkov and Yonatan Bisk. 2018. Synthetic and Natural Noise Both Break Neural Machine Translation. In ICLR.Google ScholarGoogle Scholar
  2. Tolga Bolukbasi, Kai-Wei Chang, James Y Zou, Venkatesh Saligrama, and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In NeurIPS.Google ScholarGoogle Scholar
  3. Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In FAccT.Google ScholarGoogle Scholar
  4. Alvin Chan, Lei Ma, Felix Juefei-Xu, Xiaofei Xie, Yang Liu, and Yew Soon Ong. 2018. Metamorphic relation based adversarial attacks on differentiable neural computer. arXiv preprint (2018).Google ScholarGoogle Scholar
  5. Faisal Kamiran and Toon Calders. 2009. Classifying without discriminating. In International Conference on Computer, Control and Communication.Google ScholarGoogle ScholarCross RefCross Ref
  6. Qi Lei, Lingfei Wu, Pin-Yu Chen, Alexandros G Dimakis, Inderjit S Dhillon, and Michael Witbrock. 2018. Discrete attacks and submodular optimization with applications to text classification. arXiv preprint (2018).Google ScholarGoogle Scholar
  7. Paul Pu Liang, Chiyu Wu, Louis-Philippe Morency, and Ruslan Salakhutdinov. 2021. Towards understanding and mitigating social biases in language models. In ICML.Google ScholarGoogle Scholar
  8. Haochen Liu, Jamell Dacon, Wenqi Fan, Hui Liu, Zitao Liu, and Jiliang Tang. 2020 a. Does Gender Matter? Towards Fairness in Dialogue Systems. In COLING.Google ScholarGoogle Scholar
  9. Haochen Liu, Tyler Derr, Zitao Liu, and Jiliang Tang. 2019. Say what i want: Towards the dark side of neural dialogue models. arXiv preprint (2019).Google ScholarGoogle Scholar
  10. Haochen Liu, Wei Jin, Hamid Karimi, Zitao Liu, and Jiliang Tang. 2021a. The Authors Matter: Understanding and Mitigating Implicit Bias in Deep Text Classification. In Findings of ACL-IJCNLP.Google ScholarGoogle Scholar
  11. Haochen Liu, Zitao Liu, Zhongqin Wu, and Jiliang Tang. 2020b. Personalized Multimodal Feedback Generation in Education. In COLING.Google ScholarGoogle Scholar
  12. Haochen Liu, Da Tang, Ji Yang, Xiangyu Zhao, Jiliang Tang, and Youlong Cheng. 2021b. Self-supervised Learning for Alleviating Selection Bias in Recommendation Systems. (2021).Google ScholarGoogle Scholar
  13. Haochen Liu, Joseph Thekinen, Sinem Mollaoglu, Da Tang, Ji Yang, Youlong Cheng, Hui Liu, and Jiliang Tang. 2021 c. Toward Annotator Group Bias in Crowdsourcing. arXiv preprint (2021).Google ScholarGoogle Scholar
  14. Haochen Liu, Wentao Wang, Yiqi Wang, Hui Liu, Zitao Liu, and Jiliang Tang. 2020 d. Mitigating Gender Bias for Neural Dialogue Generation with Adversarial Learning. In EMNLP .Google ScholarGoogle Scholar
  15. Haochen Liu, Yiqi Wang, Wenqi Fan, Xiaorui Liu, Yaxin Li, Shaili Jain, Anil K Jain, and Jiliang Tang. 2021 d. Trustworthy ai: A computational perspective. arXiv preprint (2021).Google ScholarGoogle Scholar
  16. Haochen Liu, Zhiwei Wang, Tyler Derr, and Jiliang Tang. 2020c. Chat as expected: Learning to manipulate black-box neural dialogue models. arXiv preprint (2020).Google ScholarGoogle Scholar
  17. Haochen Liu, Xiangyu Zhao, Chong Wang, Xiaobing Liu, and Jiliang Tang. 2020e. Automated embedding size search in deep recommender systems. In SIGIR.Google ScholarGoogle Scholar
  18. Marcelo O.R. Prates, Pedro H. Avelar, and Luis C. Lamb. 2019. Assessing gender bias in machine translation: a case study with google translate. Neural Computing and Applications (2019).Google ScholarGoogle Scholar
  19. Hee Jung Ryu, Margaret Mitchell, and Hartwig Adam. 2017. Improving smiling detection with race and gender diversity. arXiv preprint (2017).Google ScholarGoogle Scholar
  20. Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2014. Intriguing properties of neural networks.Google ScholarGoogle Scholar
  21. Kush R Varshney. 2019. Trustworthy machine learning and artificial intelligence. XRDS: Crossroads, The ACM Magazine for Students (2019).Google ScholarGoogle Scholar
  22. Marty J. Wolf, Keith W. Miller, and Frances S. Grodzinsky. 2017. Why we should have seen that coming: comments on microsoft's tay "experiment," and wider implications. The ORBIT Journal (2017).Google ScholarGoogle Scholar
  23. Han Xu, Yao Ma, Hao-Chen Liu, Debayan Deb, Hui Liu, Ji-Liang Tang, and Anil K Jain. 2020. Adversarial attacks and defenses in images, graphs and text: A review. International Journal of Automation and Computing (2020).Google ScholarGoogle ScholarCross RefCross Ref
  24. Wei Emma Zhang, Quan Z Sheng, Ahoud Abdulrahmn F Alhazmi, and Chenliang Li. 2019. Generating textual adversarial examples for deep learning models: A survey. arXiv preprint (2019).Google ScholarGoogle Scholar
  25. Xiangyu Zhao, Haochen Liu, Wenqi Fan, Hui Liu, Jiliang Tang, and Chong Wang. 2021a. AutoLoss: Automated Loss Function Search in Recommendations. arXiv preprint (2021).Google ScholarGoogle Scholar
  26. Xiangyu Zhao, Haochen Liu, Hui Liu, Jiliang Tang, Weiwei Guo, Jun Shi, Sida Wang, Huiji Gao, and Bo Long. 2021b. AutoDim: Field-aware Embedding Dimension Searchin Recommender Systems. In Web Conference.Google ScholarGoogle Scholar

Index Terms

  1. Trustworthy Machine Learning: Fairness and Robustness

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      WSDM '22: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining
      February 2022
      1690 pages
      ISBN:9781450391320
      DOI:10.1145/3488560

      Copyright © 2022 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 15 February 2022

      Check for updates

      Qualifiers

      • extended-abstract

      Acceptance Rates

      Overall Acceptance Rate498of2,863submissions,17%

      Upcoming Conference

    • Article Metrics

      • Downloads (Last 12 months)194
      • Downloads (Last 6 weeks)16

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader