Restricted Boltzmann machines (RBM) are generative models capable to learn data with a rich underlying structure. We study the teacher–student setting where a student RBM learns structured data generated by a teacher RBM. The amount of structure in the data is controlled by adjusting the number of hidden units of the teacher and the correlations in the rows of the weights, a.k.a. patterns. In the absence of correlations, we validate the conjecture that the performance is independent of the number of teacher patterns and hidden units of the student RBMs, and we argue that the teacher–student setting can be used as a toy model for studying the lottery ticket hypothesis. Beyond this regime, we find that the critical amount of data required to learn the teacher patterns decreases with both their number and correlations. In both regimes, we find that, even with a relatively large dataset, it becomes impossible to learn the teacher patterns if the inference temperature used for regularization is kept too low. In our framework, the student can learn teacher patterns one-to-one or many-to-one, generalizing previous findings about the teacher–student setting with two hidden units to any arbitrary finite number of hidden units.

Thériault, R., Tosello, F., Tantari, D. (2025). Modeling structured data learning with Restricted Boltzmann machines in the teacher–student setting. NEURAL NETWORKS, 189, 1-24 [10.1016/j.neunet.2025.107542].

Modeling structured data learning with Restricted Boltzmann machines in the teacher–student setting

Tosello, Francesco;Tantari, Daniele
2025

Abstract

Restricted Boltzmann machines (RBM) are generative models capable to learn data with a rich underlying structure. We study the teacher–student setting where a student RBM learns structured data generated by a teacher RBM. The amount of structure in the data is controlled by adjusting the number of hidden units of the teacher and the correlations in the rows of the weights, a.k.a. patterns. In the absence of correlations, we validate the conjecture that the performance is independent of the number of teacher patterns and hidden units of the student RBMs, and we argue that the teacher–student setting can be used as a toy model for studying the lottery ticket hypothesis. Beyond this regime, we find that the critical amount of data required to learn the teacher patterns decreases with both their number and correlations. In both regimes, we find that, even with a relatively large dataset, it becomes impossible to learn the teacher patterns if the inference temperature used for regularization is kept too low. In our framework, the student can learn teacher patterns one-to-one or many-to-one, generalizing previous findings about the teacher–student setting with two hidden units to any arbitrary finite number of hidden units.
2025
Thériault, R., Tosello, F., Tantari, D. (2025). Modeling structured data learning with Restricted Boltzmann machines in the teacher–student setting. NEURAL NETWORKS, 189, 1-24 [10.1016/j.neunet.2025.107542].
Thériault, Robin; Tosello, Francesco; Tantari, Daniele
File in questo prodotto:
File Dimensione Formato  
1-s2.0-S0893608025004216-main.pdf

accesso aperto

Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione (CCBY)
Dimensione 6.44 MB
Formato Adobe PDF
6.44 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1016220
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact