Skip to main content

Neural Maximum Independent Set

  • Conference paper
  • First Online:
Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2021)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1524))

  • 2261 Accesses

Abstract

The emergence of deep learning brought solutions to many difficult problems and has recently motivated new studies that try to solve hard combinatorial optimization problems with machine learning approaches. We propose a framework based on Expert Iteration, an imitation learning method that we apply to solve combinatorial optimization problems on graphs, in particular the Maximum Independent Set problem. Our method relies on training GNNs to recognize how to complete a solution, given a partial solution of the problem as an input. This paper emphasizes some interesting findings such as the introduction of learned nodes features helping the neural network to give relevant solutions. Moreover, we represent the space of good solutions and discuss the ability of GNN’s to solve the problem on a graph without training on it.

The authors acknowledge the support of the ANR as part of the “Investissements d’avenir” program (ANR-19-P3IA-0001, PRAIRIE 3IA Institute) and through the project DELCO (ANR-19-CE23-0016).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    We kept the same name used in [1] but it is actually an instance of BHOSLIB [33].

References

  1. Abe, K., Xu, Z., Sato, I., Sugiyama, M.: Solving NP-hard problems on graphs with extended AlphaGo zero (2020)

    Google Scholar 

  2. Agarwal, P.K., van Kreveld, M., Suri, S.: Label placement by maximum independent set in rectangles. Comput. Geom. 11(3), 209–218 (1998)

    Article  MathSciNet  Google Scholar 

  3. Akiba, T., Sano, S., Yanase, T., Ohta, T., Koyama, M.: Optuna: a next-generation hyperparameter optimization framework. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, pp. 2623–2631. Association for Computing Machinery, New York (2019)

    Google Scholar 

  4. Alexander, J., Mink, T.: A new method for enumerating independent sets of a fixed size in general graphs. J. Graph Theory 81(1), 57–72 (2016)

    Article  MathSciNet  Google Scholar 

  5. Anthony, T., Tian, Z., Barber, D.: Thinking fast and slow with deep learning and tree search. In: Advances in Neural Information Processing Systems, pp. 5360–5370 (2017)

    Google Scholar 

  6. Barabási, A.L., Albert, R.: Emergence of Scaling in Random Networks. Science 286(5439), 509–512 (1999)

    Article  MathSciNet  Google Scholar 

  7. Bengio, Y., Lodi, A., Prouvost, A.: Machine learning for combinatorial optimization: a methodological tour d’horizon. Eur. J. Oper. Res. 290(2), 405–421 (2021)

    Article  MathSciNet  Google Scholar 

  8. Bourjolly, J.M., Laporte, G., Mercure, H.: A combinatorial column generation algorithm for the maximum stable set problem. Oper. Res. Lett. 20(1), 21–29 (1997)

    Article  MathSciNet  Google Scholar 

  9. Byskov, J.M.: Enumerating maximal independent sets with applications to graph colouring. Oper. Res. Lett. 32(6), 547–556 (2004)

    Article  MathSciNet  Google Scholar 

  10. Chen, D., Lin, Y., Li, W., Li, P., Zhou, J., Sun, X.: Measuring and relieving the over-smoothing problem for graph neural networks from the topological view. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, February 7–12, 2020, pp. 3438–3445. AAAI Press, New York (2020)

    Google Scholar 

  11. Chen, M., Wei, Z., Huang, Z., Ding, B., Li, Y.: Simple and deep graph convolutional networks. In: III, H.D., Singh, A. (eds.) Proceedings of the 37th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 119, pp. 1725–1735. PMLR, 13–18 July 2020

    Google Scholar 

  12. Chen, X., Tian, Y.: Learning to perform local rewriting for combinatorial optimization. In: Wallach, H., Larochelle, H., Beygelzimer, A., d’ Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32. Curran Associates, Inc. (2019)

    Google Scholar 

  13. Chen, Z., et al.: Bridging the gap between spatial and spectral domains: a survey on graph neural networks (2020)

    Google Scholar 

  14. Das, K.N., Chaudhuri, B.: Heuristics to find maximum independent set: An overview. In: Deep, K., Nagar, A., Pant, M., Bansal, J.C. (eds.) Proceedings of the International Conference on Soft Computing for Problem Solving (SocProS 2011) 20–22 December, 2011, pp. 881–892. Springer, India (2012)

    Google Scholar 

  15. Dwivedi, V.P., Joshi, C.K., Laurent, T., Bengio, Y., Bresson, X.: Benchmarking Graph Neural Networks. arXiv e-prints (2020)

    Google Scholar 

  16. Gasse, M., Chételat, D., Ferroni, N., Charlin, L., Lodi, A.: Exact combinatorial optimization with graph convolutional neural networks. In: Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R. (eds.) Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, 8–14 December, 2019, Vancouver, BC, Canada, pp. 15554–15566 (2019)

    Google Scholar 

  17. Gurski, F., Rehs, C.: Counting and enumerating independent sets with applications to combinatorial optimization problems. Math. Methods Oper. Res. 91(3), 439–463 (2020)

    Article  MathSciNet  Google Scholar 

  18. Hu, W., et al.: Open graph benchmark: Datasets for machine learning on graphs (2021)

    Google Scholar 

  19. Karalias, N., Loukas, A.: Erdos goes neural: an unsupervised learning framework for combinatorial optimization on graphs. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 6659–6672. Curran Associates, Inc. (2020)

    Google Scholar 

  20. Karp, R.M.: Reducibility among Combinatorial Problems, pp. 85–103. Springer, US (1972)

    Google Scholar 

  21. Khalil, E., Dai, H., Zhang, Y., Dilkina, B., Song, L.: Learning combinatorial optimization algorithms over graphs. In: Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc. (2017)

    Google Scholar 

  22. Lamm, S., Sanders, P., Schulz, C., Strash, D., Werneck, R.F.: Finding near-optimal independent sets at scale. J. Heuristics 23(4), 207–229 (2017)

    Article  Google Scholar 

  23. Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11) (2008)

    Google Scholar 

  24. Mazyavkina, N., Sviridov, S., Ivanov, S., Burnaev, E.: Reinforcement learning for combinatorial optimization: a survey (2020)

    Google Scholar 

  25. Piotr, B., Marek, K.: On some tighter inapproximability results. Technical report (1999)

    Google Scholar 

  26. Rossi, F., Smriglio, S.: A branch-and-cut algorithm for the maximum cardinality stable set problem. Oper. Res. Lett. 28(2), 63–74 (2001)

    Article  MathSciNet  Google Scholar 

  27. Rossi, R.A., Ahmed, N.K.: The Network Data Repository with Interactive Graph Analytics and Visualization. In: Bonet, B., Koenig, S. (eds.) Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, January 25–30, 2015, Austin, Texas, USA, pp. 4292–4293. AAAI Press (2015)

    Google Scholar 

  28. Sato, R., Yamada, M., Kashima, H.: Approximation ratios of graph neural networks for combinatorial problems. In: Wallach, H., Larochelle, H., Beygelzimer, A., d’ Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32. Curran Associates, Inc. (2019)

    Google Scholar 

  29. Sato, R., Yamada, M., Kashima, H.: Random Features Strengthen Graph Neural Networks. CoRR abs/2002.03155 (2020). https://arxiv.org/abs/2002.03155

  30. Silver, D., et al.: A general reinforcement learning algorithm that masters chess, shogi, and go through self-play. Science 362(6419), 1140–1144 (2018)

    Article  MathSciNet  Google Scholar 

  31. Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., Bengio, Y.: Graph attention networks. In: 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30–3 May 2018, Conference Track Proceedings (2018)

    Google Scholar 

  32. Xiao, M., Nagamochi, H.: Exact algorithms for maximum independent set. Inf. Comput. 255, 126–146 (2017)

    Article  MathSciNet  Google Scholar 

  33. Xu, K.: BHOSLIB: Benchmarks with Hidden Optimum Solutions for Graph Problems (Maximum Clique, Maximum Independent Set, Minimum Vertex Cover and Vertex Coloring). http://sites.nlsde.buaa.edu.cn/~kexu/benchmarks/graph-benchmarks.htm

  34. Xu, K., Hu, W., Leskovec, J., Jegelka, S.: How Powerful are Graph Neural Networks? In: 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019 (2019)

    Google Scholar 

  35. Yehuda, G., Gabel, M., Schuster, A.: It’s not what machines can learn, it’s what we cannot teach. In: III, H.D., Singh, A. (eds.) Proceedings of the 37th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 119, pp. 10831–10841 (2020)

    Google Scholar 

  36. Zhou, J., et al.: Graph neural networks: a review of methods and applications (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Thomas Pontoizeau .

Editor information

Editors and Affiliations

Appendices

A Results on other instances

Fig. 3.
figure 3

The score of the argmax playout each epoch of the learning on each instance. Learning curves with embeddings are highlighted by a plain curve.

Fig. 4.
figure 4

Distribution of scores of 500 solutions from the training set and an exploration of 500 rollouts with the GNN for each instance, first with only the state, and then with all the features.

B Stochastic exploration method allows to find various good solutions

By giving a set of good solutions, the stochastic exploration method gives us some interesting insights about the distribution of the solutions in a given instance. After several exploration methods, we saved all good solutions of the three best found scores in order to observe the clusters they form (i.e. in the sense of hamming distance between sets). Note that the pool of best solutions for dimacs-frb30-15-1 is the only one that necessitated 5 iterations of labeling/learning phase and the solutions were found in the labeling phase, the stochastic exploration method giving poor solutions as discussed previously.

We sum up the information about the set of solutions we obtained:

  • ba200_5: 1400 solutions with score 80, 700 solutions with score 81, 123 solutions with score 82.

  • er200_10: 1400 solutions with score 39, 700 solutions with score 40, 94 solutions with score 41.

  • dimacs-frb30-15-1: 499 solutions with score 26, 77 solutions with score 27, 4 solutions with score 28.

  • bio-SC-LC: 21 solutions with score 966, 5 solutions with score 967, 1 solution with score 968.

In order to observe how the solutions are organized, we first computed all hamming distances between each pair of solution to see how far they are from each other. In Fig. 5, we represented the distribution of all possible hamming distances for the three best found scores for each instance. For er200_10 and dimacs-frb30-15-1, we also represented the graphic in log scale for more readability.

Fig. 5.
figure 5

The distribution of hamming distances between best found solutions for each instance of the dataset. Sometimes we also give the same graphic in log scale for more readability.

For ba200_5, the solutions are well organized along a Gaussian curve. For er200_10, we observe almost the same phenomenon except that the best solutions (with score 41) are split into three clusters: one big cluster of close solutions, one sparse cluster of partially close solutions, and a small sparse cluster of solutions. For dimacs-frb30-15-1, all best found solutions look very sparse and have very few nodes in common. For bio-SC-LC, the solutions seem to be well organized and not so far from each other. Note that there is no hamming distance for the score 968 since we only obtained one solution for this score.

For each instance, we embedded our pool of solutions into a three dimensional space with a t-SNE [23] in Fig. 6, that allows to observe our previous remarks on the distribution of hamming distances. We can notice that the scatterplots are always nested around the best solutions.

Fig. 6.
figure 6

The projection of the set of best found solutions for each instance with the t-SNE method. (Color figure online)

Note that for the instances ba200_5 and er200_10, we had to sample the second and third best found solutions (in green and orange) in order to highlight the pool of best solutions (red).

With the instance dimacs-frb30-15-1, we observe that all best found solutions look very sparse in the sense that the solutions have very few nodes in common. This could explain why GNN cannot converge properly into a good solution. The latter remark could explain why our neural network has so much difficulty to learn how to solve the problem on this instance and converge to good solutions.

Since enumerating maximal or maximum independent sets has also been studied in the literature [4, 9, 17], our method provides some interesting tools to obtain good various solutions with the help of a GNN.

C GNNs can export expertise on Max Independent Set from a small graph to a larger graph

The last observation we want to highlight is that once a neural network has learned on a small graph, it performs a little better on other bigger instances it never saw.

In order to do so, we had to remove the footprint features and the embedding features during the learning phase since they had sense only in a particular instance. Thus, the quality of the learning is quite less effective than with all features.

In our experiment, we compare training on ba200_5 and er200_10 and see how much the GNN performs on the other instances (including two other instances: ba100_5 and ba1000_5 constructed in the same way than ba200_5).

After a labeling phase, we make a learning phase and compute the score of the argmax sequence for each other instance using the trained model at each epoch. We then smoothed the curves by computing the average score of the last 5 scores at each epoch, and then represented the approximation ratio between the score of the argmax sequence and the best known result on the instance for each epoch. The resulted curves are represented in Fig. 7.

Fig. 7.
figure 7

Evolution of approximation ratio (compared to the best known score) on the argmax playout on other graph training on ba200_5 and er200_10.

When training on ba200_5, the quality of the argmax sequence improves quickly at the beginning and then stagnates. When training on er200_10, the qualitywork of the argmax sequence increases slower than with ba200_5 but increases all along the learning. On the other hand, not surprisingly, note that the learning phase did not make any improvement on dimacs-frb30-15-1. Those observations indicate that our GNN was able to transfer some knowledge about Maximum Independent Set on brand new graphs, and is promising for future work.

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Pontoizeau, T., Sikora, F., Yger, F., Cazenave, T. (2021). Neural Maximum Independent Set. In: Kamp, M., et al. Machine Learning and Principles and Practice of Knowledge Discovery in Databases. ECML PKDD 2021. Communications in Computer and Information Science, vol 1524. Springer, Cham. https://doi.org/10.1007/978-3-030-93736-2_18

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-93736-2_18

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-93735-5

  • Online ISBN: 978-3-030-93736-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics