skip to main content
10.1145/3595916.3626351acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
demonstration

FinGuard: A Multimodal AIGC Guardrail in Financial Scenarios

Published: 01 January 2024 Publication History

Abstract

Recently, the development of foundation models has led to significant advances in the ability of artificial intelligence (AI) to generate multimodal content such as text and images. However, specialized industrial scenarios such as finance, which require high levels of security and compliance, pose challenges for the application of generative AI due to its uncontrollability. To address this issue, we propose FinGuard, a multimodal AI-generated content (AIGC) guardrail specifically designed for financial scenarios. We provide detailed definitions of the general quality, financial compliance, and security dimensions of AIGC, and implement the evaluation and inspection of multimodal AIGC including text and images. Our proposed FinGuard has been applied to a financial marketing application serving hundreds of millions of users.

Supplementary Material

MP4 File (FinGuard_Demo_Video.mp4)
Demo video

References

[1]
Rishi Bommasani, Drew A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, 2021. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258 (2021). https://arxiv.org/abs/2108.07258
[2]
Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Shijin Wang, and Guoping Hu. 2020. Revisiting Pre-Trained Models for Chinese Natural Language Processing. In Findings of EMNLP. 657–668. https://doi.org/10.18653/v1/2020.findings-emnlp.58
[3]
Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, and Ziqing Yang. 2021. Pre-training with whole word masking for chinese bert. IEEE/ACM Transactions on Audio, Speech, and Language Processing 29 (2021), 3504–3514. https://doi.org/10.1109/TASLP.2021.3124365
[4]
Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of ICLR. https://arxiv.org/abs/2010.11929
[5]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of CVPR. 770–778. https://doi.org/10.1109/CVPR.2016.90
[6]
Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C Berg, Wan-Yen Lo, 2023. Segment anything. arXiv preprint arXiv:2304.02643 (2023). https://arxiv.org/abs/2304.02643
[7]
Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. 2022. BLIP: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In Proceedings of ICML. 12888–12900. https://proceedings.mlr.press/v162/li22n.html
[8]
OpenAI. 2023. GPT-4 Technical Report. arXiv preprint arXiv:2303.08774 (2023). https://arxiv.org/abs/2303.08774
[9]
Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, 2019. Language models are unsupervised multitask learners. OpenAI blog (2019). https://insightcivic.s3.us-east-1.amazonaws.com/language-models.pdf
[10]
Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125 (2022). https://arxiv.org/abs/2204.06125
[11]
Sidi Yang, Tianhe Wu, Shuwei Shi, Shanshan Lao, Yuan Gong, Mingdeng Cao, Jiahao Wang, and Yujiu Yang. 2022. MANIQA: Multi-dimension attention network for no-reference image quality assessment. In Proceedings of CVPR. 1191–1200. https://doi.org/10.1109/CVPRW56347.2022.00126
[12]
Yuheng Zha, Yichi Yang, Ruichen Li, and Zhiting Hu. 2023. AlignScore: Evaluating Factual Consistency with a Unified Alignment Function. arXiv preprint arXiv:2305.16739 (2023). https://arxiv.org/abs/2305.16739

Cited By

View all
  • (2024)Detoxifying Large Language Models via Kahneman-Tversky OptimizationNatural Language Processing and Chinese Computing10.1007/978-981-97-9443-0_36(409-417)Online publication date: 1-Nov-2024
  • (2024)Human- and AI-Generated Marketing Content Comparison Corpus, Evaluation, and DetectionPRICAI 2024: Trends in Artificial Intelligence10.1007/978-981-96-0119-6_18(177-183)Online publication date: 12-Nov-2024

Index Terms

  1. FinGuard: A Multimodal AIGC Guardrail in Financial Scenarios

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      MMAsia '23: Proceedings of the 5th ACM International Conference on Multimedia in Asia
      December 2023
      745 pages
      ISBN:9798400702051
      DOI:10.1145/3595916
      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 01 January 2024

      Check for updates

      Author Tags

      1. AIGC
      2. finance
      3. guardrail
      4. multimodal

      Qualifiers

      • Demonstration
      • Research
      • Refereed limited

      Conference

      MMAsia '23
      Sponsor:
      MMAsia '23: ACM Multimedia Asia
      December 6 - 8, 2023
      Tainan, Taiwan

      Acceptance Rates

      Overall Acceptance Rate 59 of 204 submissions, 29%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)96
      • Downloads (Last 6 weeks)8
      Reflects downloads up to 28 Feb 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Detoxifying Large Language Models via Kahneman-Tversky OptimizationNatural Language Processing and Chinese Computing10.1007/978-981-97-9443-0_36(409-417)Online publication date: 1-Nov-2024
      • (2024)Human- and AI-Generated Marketing Content Comparison Corpus, Evaluation, and DetectionPRICAI 2024: Trends in Artificial Intelligence10.1007/978-981-96-0119-6_18(177-183)Online publication date: 12-Nov-2024

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format.

      HTML Format

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media