This research aims to enhance the combinatorial generation of assemblages made of spatial units with adaptive control over climate-related factors (such as daylight hours access on exposed surfaces) and topological aspects using Reinforcement Learning (RL). The approach uses RL-trained agents to adapt to lighting conditions in the process of selection and aggregation of parts. The agents move within a voxel space, and place parts (single voxels) aiming to ensure topological consistency and a target daylight hours access on exposed surfaces in the final assemblage. Agents' behavior is shaped by both intermediate and final rewards, providing them with feedback on the quality of their individual and overall choices. The research introduces condition-based adaptivity into a combinatorial process by means of RL training, moving beyond both simple random choice and predetermined heuristics sets; although both policies can relate to boundary conditions, they are respectively non-controllable and tied to a specific environmental scenario. Through the agent's trained policy, the system learns a state-action-reward relationship in a continuous feedback process between space, environment and climate data that generalizes to any environmental configuration that can be coded in the system's terms.
Massafra, G., Erioli, A. (2025). ReLighting Spaces: Training Daylight Access Cognition in Combinatorial Spatial Assemblages Using Reinforcement Learning. Cham : Springer Nature [10.1007/978-3-032-02782-5_21].
ReLighting Spaces: Training Daylight Access Cognition in Combinatorial Spatial Assemblages Using Reinforcement Learning
Massafra, Giuseppe
;Erioli, Alessio
2025
Abstract
This research aims to enhance the combinatorial generation of assemblages made of spatial units with adaptive control over climate-related factors (such as daylight hours access on exposed surfaces) and topological aspects using Reinforcement Learning (RL). The approach uses RL-trained agents to adapt to lighting conditions in the process of selection and aggregation of parts. The agents move within a voxel space, and place parts (single voxels) aiming to ensure topological consistency and a target daylight hours access on exposed surfaces in the final assemblage. Agents' behavior is shaped by both intermediate and final rewards, providing them with feedback on the quality of their individual and overall choices. The research introduces condition-based adaptivity into a combinatorial process by means of RL training, moving beyond both simple random choice and predetermined heuristics sets; although both policies can relate to boundary conditions, they are respectively non-controllable and tied to a specific environmental scenario. Through the agent's trained policy, the system learns a state-action-reward relationship in a continuous feedback process between space, environment and climate data that generalizes to any environmental configuration that can be coded in the system's terms.| File | Dimensione | Formato | |
|---|---|---|---|
|
P116_Massafra Giuseppe Erioli Alessio_Relighting spaces_acceptedmanuscript.pdf
embargo fino al 19/10/2026
Tipo:
Postprint / Author's Accepted Manuscript (AAM) - versione accettata per la pubblicazione dopo la peer-review
Licenza:
Licenza per accesso libero gratuito
Dimensione
789.7 kB
Formato
Adobe PDF
|
789.7 kB | Adobe PDF | Visualizza/Apri Contatta l'autore |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


