skip to main content
10.1145/3477495.3536325acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

Scalable User Interface Optimization Using Combinatorial Bandits

Published: 07 July 2022 Publication History

Abstract

The mission of major e-commerce platforms is to enable their customers to find the best products for their needs. In the common case of large inventories, complex User Interfaces (UIs) are required to allow a seamless navigation. However, as UIs often contain many widgets of different relevance, the task of constructing an optimal layout arises in order to improve the customer's experience. This is a challenging task, especially in the typical industrial setup where multiple independent teams conflict by adding and modifying UI widgets. It becomes even more challenging due to the customer preferences evolving over time, bringing the need for adaptive solutions. In a previous work [6], we addressed this task by introducing a UI governance framework powered by Machine Learning (ML) algorithms that automatically and continuously search for the optimal layout. Nevertheless, we highlighted that naive algorithmic choices exhibit several issues when implemented in the industry, such as widget dependency, combinatorial solution space and cold start problem. In this work, we demonstrate how we deal with these issues using Combinatorial Bandits, an extension of Multi-Armed Bandits (MAB) where the agent selects not only one but multiple arms at the same time. We develop two novel approaches to model combinatorial bandits, inspired by the Natural Language Processing (NLP) and the Evolutionary Algorithms (EA) fields and present their ability to enable scalable UI optimization.

References

[1]
Peter Auer, Nicolo Cesa-Bianchi, and Paul Fisher. 2002. Finitetime Analysis of the Multiarmed Bandit Problem. Machine Learning 47 (2002), 235--256. https://doi.org/10.1162/evco.1996.4.4.361 arXiv: https://link.springer.com/content/pdf/10.1023/A:1013689704352.pdf
[2]
Nicolò Cesa-Bianchi and GáBor Lugosi. 2012. Combinatorial Bandits. J. Comput. Syst. Sci. 78, 5 (Sept. 2012), 1404--1422. https://doi.org/10.1016/j.jcss.2012.01.001
[3]
Dmitri Goldenberg, Javier Albert, Lucas Bernardi, and Pablo Estevez. 2020. Free Lunch! Retrospective Uplift Modeling for Dynamic Promotions Recommendation within ROI Constraints. CoRR abs/2008.06293 (2020). arXiv:2008.06293 https://arxiv.org/abs/2008.06293
[4]
Seokyung Hahn. 2012. Understanding noninferiority trials. Korean journal of pediatrics 55, 11 (2012), 403.
[5]
Daniel Jurafsky and James H. Martin. 2021. Speech and Language Processing (Draft of December 29, 2021). Stanford University, Chapter 3. https://web.stanford.edu/~jurafsky/slp3/3.pdf
[6]
Ioannis Kangas, Maud Schwoerer, and Lucas Bernardi. 2021. Recommender Systems for Personalized User Experience: Lessons learned at Booking.com. Fifteenth ACM Conference on Recommender Systems (RecSys '21), 583--586. https://doi.org/10.1145/3460231.3474611
[7]
Tor Lattimore and Szepesvári Csaba. 2020. Bandit algorithms. Cambridge University Press, Chapter 32, 386--387. https://tor-lattimore.com/downloads/book/book.pdf
[8]
Risto Miikkulainen, Neil Iscoe, Aaron Shagrin, Ron Cordell, Sam Nazari, Cory Schoolland, Myles Brundage, Jonathan Epstein, Randy Dean, and Gurmeet Lamba. 2017. Conversion Rate Optimization through Evolutionary Computation. In Proceedings of the Genetic and Evolutionary Computation Conference (Berlin, Germany) (GECCO '17). Association for Computing Machinery, New York, NY, USA, 1193--1199. https://doi.org/10.1145/3071178.3071312
[9]
Kumara Sastry, David Goldberg, and Graham Kendall. 2005. Genetic algorithms. In Search methodologies. Springer, 97--125.
[10]
Timothy Starkweather, S. McDaniel, Keith E. Mathias, L. Darrell Whitley, and C. Whitley. 1991. A Comparison of Genetic Sequencing Operators. In ICGA. 69--76.

Cited By

View all
  • (2024)MARLUI: Multi-Agent Reinforcement Learning for Adaptive Point-and-Click UIsProceedings of the ACM on Human-Computer Interaction10.1145/36611478:EICS(1-27)Online publication date: 17-Jun-2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SIGIR '22: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval
July 2022
3569 pages
ISBN:9781450387323
DOI:10.1145/3477495
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 07 July 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. combinatorial bandits
  2. multi-armed bandits
  3. user interface

Qualifiers

  • Short-paper

Conference

SIGIR '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 792 of 3,983 submissions, 20%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)17
  • Downloads (Last 6 weeks)1
Reflects downloads up to 28 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)MARLUI: Multi-Agent Reinforcement Learning for Adaptive Point-and-Click UIsProceedings of the ACM on Human-Computer Interaction10.1145/36611478:EICS(1-27)Online publication date: 17-Jun-2024

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media