{"id":296905,"date":"2025-07-27T21:59:25","date_gmt":"2025-07-27T21:59:25","guid":{"rendered":"https:\/\/www.europesays.com\/uk\/296905\/"},"modified":"2025-07-27T21:59:25","modified_gmt":"2025-07-27T21:59:25","slug":"the-artificial-intelligence-technology-for-immersion-experience-and-space-design-in-museum-exhibition","status":"publish","type":"post","link":"https:\/\/www.europesays.com\/uk\/296905\/","title":{"rendered":"The artificial intelligence technology for immersion experience and space design in museum exhibition"},"content":{"rendered":"<p>Optimization model of Spatial layout driven by AI<\/p>\n<p>Figure <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig1\" target=\"_blank\" rel=\"noopener\">1<\/a> presents the design of the spatial optimization model, illustrating the entire process from data input to dynamic optimization. At the core of the system is the RL model, which drives layout adjustments based on real-time reward feedback. Each step is supported by segmented descriptions, preprocessing techniques, feature threshold settings, and relevant literature, highlighting the scientific rationale behind the model\u2019s design.<\/p>\n<p><b id=\"Fig1\" class=\"c-article-section__figure-caption\" data-test=\"figure-caption-text\">Fig. 1<\/b><a class=\"c-article-section__figure-link\" data-test=\"img-link\" data-track=\"click\" data-track-label=\"image\" data-track-action=\"view figure\" href=\"https:\/\/www.nature.com\/articles\/s41598-025-13408-2\/figures\/1\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" aria-describedby=\"Fig1\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/07\/41598_2025_13408_Fig1_HTML.png\" alt=\"figure 1\" loading=\"lazy\" width=\"685\" height=\"399\"\/><\/a><\/p>\n<p>Design of spatial optimization model.<\/p>\n<p>Design of Spatial optimization model<\/p>\n<p>To enhance both the exhibition effect and the audience experience, this study proposes a spatial layout optimization model based on AI technology. The model integrates RL and CV to enable automated and dynamic adjustments of exhibition layouts. The optimization process begins by abstracting the exhibition space as a dynamic environment. Visitor behavior data\u2014such as dwell time, navigation paths, and interaction frequency\u2014are treated as inputs representing the state of the environment. The model\u2019s actions include adjusting exhibit positions, reordering content, and redefining exhibition zones. To train the model in selecting optimal layout strategies, a reward function is constructed based on visitor behavior. Layout changes that lead to longer dwell times, more efficient navigation, or increased interaction receive higher rewards. This incentivizes the model to continuously improve the exhibition layout. The DQN algorithm is employed to implement this strategy, using a Q-value function to evaluate state-action pairs and updating its parameters through temporal difference learning. By interacting with the environment over time, the RL model refines its decision-making based on real-time feedback. This approach makes the layout more adaptive and significantly enhances visitor engagement and satisfaction. Within the RL framework, the museum environment is modeled as the agent\u2019s environment, audience behavior as the state input, and layout adjustments as the output actions. The reward mechanism is designed to prioritize configurations that maximize audience interaction and experiential satisfaction. Therefore, this study focuses on key audience behaviors and constructs a reward function accordingly.<\/p>\n<p>1) Dwell Time: Longer visitor dwell time in front of an exhibit typically indicated higher interest and engagement. Therefore, the model assigned positive rewards for increased dwell time, encouraging the system to adjust exhibit positions in a way that captured and sustained audience attention.<\/p>\n<p>2) Visiting Path: A shorter or more coherent visiting path suggested that the spatial layout effectively guided the audience through the exhibition. The model rewarded rational, efficient navigation paths to reduce unnecessary detours and enhance the overall flow and user experience.<\/p>\n<p>3) Interaction Frequency: The frequency of interactions\u2014such as touchscreen use, AR\/VR engagement, or physical exhibit interaction\u2014served as a key metric of exhibit attractiveness. Higher interaction frequency received greater rewards, prompting the model to favor more interactive and engaging layout configurations.<\/p>\n<p>4) Emotional Feedback: Using CV and affective computing, the system analyzed facial expressions to detect positive emotional responses such as joy and surprise. Areas that consistently elicited positive emotional reactions were given higher rewards, guiding the layout optimization toward emotionally resonant spatial arrangements.<\/p>\n<p>5) Audience Distribution Balance: Uneven visitor distribution\u2014where some areas were overcrowded while others were underutilized\u2014was considered detrimental to visitor experience. The model rewarded more balanced audience distribution across exhibition zones to enhance space utilization and visitor comfort.<\/p>\n<p>Dwell time is a key indicator of visitor interest and has been widely validated in museum behavior studies<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 40\" title=\"Grajal, A. et al. Conservation psychology and the legacy of Carol Saunders for zoo and aquarium programs, networks, practices, and exhibits. Zoo Biol. 41, 379&#x2013;385 (2022).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR40\" id=\"ref-link-section-d272762662e605\" target=\"_blank\" rel=\"noopener\">40<\/a>. This study tracked the duration of time visitors spent in front of exhibits in real time. Following the threshold method proposed by Jouibari\u2014where a stay of \u2265\u20093\u00a0s indicates valid interest\u2014the attractiveness of exhibits was quantitatively assessed<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 41\" title=\"Jouibari, F. R. et al. Navigation behaviour of visitors in museums based on visibility analysis and neural network simulation. Museum Manage. Curatorship. 36, 30&#x2013;47 (2021).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR41\" id=\"ref-link-section-d272762662e609\" target=\"_blank\" rel=\"noopener\">41<\/a>. For visiting path optimization, this study referred to the path-planning framework developed by Ntakolia<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 42\" title=\"Ntakolia, C. &amp; Iakovidis, D. K. A route planning framework for smart wearable assistive navigation systems. SN Appl. Sci. 3, 104 (2021).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR42\" id=\"ref-link-section-d272762662e613\" target=\"_blank\" rel=\"noopener\">42<\/a>which identified path coherence and redundancy as critical factors influencing the visitor experience. Interaction frequency was measured using the standards outlined in Wu\u2019s interaction design theory<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 43\" title=\"Wu, Y. et al. What drives users to adopt a digital museum? A case of virtual exhibition hall of National costume museum. Sage Open. 12, 21582440221082105 (2022).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR43\" id=\"ref-link-section-d272762662e617\" target=\"_blank\" rel=\"noopener\">43<\/a>. A high interaction rate (\u2265\u200980%) was used as a benchmark for participatory and user-friendly exhibit design. Audience emotional feedback was analyzed based on EC theory. Multimodal data\u2014including facial expressions, voice tone, and body posture\u2014were integrated to classify emotional states in real time<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 44\" title=\"Ojha, S., Vitale, J. &amp; Williams, M. A. Computational emotion models: A thematic review. Int. J. Social Robot. 13, 1253&#x2013;1279 (2021).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR44\" id=\"ref-link-section-d272762662e621\" target=\"_blank\" rel=\"noopener\">44<\/a>. Audience distribution balance was evaluated using the social force model, with the Gini coefficient applied to measure the evenness of crowd density across the space. This approach aimed to prevent the negative impact of local overcrowding on the overall visitor experience<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 45\" title=\"Darsena, D. et al. Sensing technologies for crowd management, adaptation, and information dissemination in public transportation systems: A review. IEEE Sens. J. 23, 68&#x2013;87 (2022).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR45\" id=\"ref-link-section-d272762662e626\" target=\"_blank\" rel=\"noopener\">45<\/a>. The goal was not to eliminate natural crowd clustering around popular exhibits, but rather to distinguish it from inefficient congestion using dynamic thresholds. Following Easson\u2019s visitor interest-driven theory<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 46\" title=\"Easson, H. &amp; Leask, A. After-hours events at the National museum of scotland: A product for attracting, engaging and retaining new museum audiences? Curr. Issues Tourism. 23, 1343&#x2013;1356 (2020).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR46\" id=\"ref-link-section-d272762662e630\" target=\"_blank\" rel=\"noopener\">46<\/a>a density threshold (\\(\\:{\\rho\\:}_{max}=5people\/{m}^{2}\\)) was set for exhibit areas. When local density remained at or below this threshold (\\(\\:\\rho\\:\\le\\:{\\rho\\:}_{max}\\)), the system interpreted it as reasonable clustering and applied a base-level reward. However, if density exceeded the threshold (\\(\\:\\rho\\:&gt;{\\rho\\:}_{max}\\)), a congestion optimization mechanism was triggered\u2014such as path redirection or minor exhibit repositioning\u2014and a negative reward was assigned to the excess density (\\(\\:\\rho\\:-{\\rho\\:}_{max}\\)). This mechanism preserved natural interest clusters (e.g., crowds around iconic pieces like the Mona Lisa) while preventing disruptive congestion. Furthermore, \\(\\:{\\rho\\:}_{max}\\) was adaptively adjusted based on the type of exhibit. Interactive zones (e.g., VR areas), which required more space for safe and effective engagement, were assigned a lower threshold (3 people\/m\u00b2), whereas static display zones were allowed a higher limit (6 people\/m\u00b2).<\/p>\n<p>Realization of DQN algorithm<\/p>\n<p>The DQN was employed as the primary RL algorithm<a data-track=\"click\" data-track-action=\"reference anchor\" data-track-label=\"link\" data-test=\"citation-ref\" aria-label=\"Reference 47\" title=\"Zhu, Z. et al. Effective defense strategies in network security using improved double dueling deep Q-network. Computers Secur. 136, 103578 (2024).\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#ref-CR47\" id=\"ref-link-section-d272762662e698\" target=\"_blank\" rel=\"noopener\">47<\/a>. In this framework, each spatial layout decision\u2014such as adjusting exhibit positions or redefining exhibition zones\u2014was treated as a distinct state. The reward signal was generated based on real-time audience behavior data, including dwell time and path selection. Through iterative training, the model progressively learned the optimal spatial layout strategy. The overall optimization objective is expressed in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ1\" target=\"_blank\" rel=\"noopener\">1<\/a>):<\/p>\n<p>$$\\:{R}_{t}=\\sum\\:_{i=1}^{n}{\\text{{\\rm\\:Y}}}^{i}\\cdot\\:{r}_{i}$$<\/p>\n<p>\n                    (1)\n                <\/p>\n<p>\\(\\:{R}_{t}\\) represents the total reward at time step t, where \u03b3 is the discount factor, \\(\\:{r}_{i}\\) is the immediate reward for the ith behavior, and n is the length of the behavior sequence. This equation indicates that the model\u2019s reward depends not only on the current action but also on the expected future rewards. As a result, the model optimizes the museum\u2019s spatial layout with a long-term perspective rather than focusing solely on immediate feedback.<\/p>\n<p>To more accurately analyze visitor behavior within the exhibition space, CV technology is integrated into the system. Sensors and cameras are deployed to capture real-time behavioral data, including movement trajectories, dwell times, and gaze points. These data not only assist in evaluating the current layout\u2019s effectiveness but also provide critical feedback for ongoing optimization. Figure\u00a0<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig2\" target=\"_blank\" rel=\"noopener\">2<\/a> illustrates a sample spatial layout used in some museums.<\/p>\n<p><b id=\"Fig2\" class=\"c-article-section__figure-caption\" data-test=\"figure-caption-text\">Fig. 2<\/b><a class=\"c-article-section__figure-link\" data-test=\"img-link\" data-track=\"click\" data-track-label=\"image\" data-track-action=\"view figure\" href=\"https:\/\/www.nature.com\/articles\/s41598-025-13408-2\/figures\/2\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" aria-describedby=\"Fig2\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/07\/41598_2025_13408_Fig2_HTML.png\" alt=\"figure 2\" loading=\"lazy\" width=\"001\" height=\"001\"\/><\/a><\/p>\n<p>Some examples of exhibition space layout in museums.<\/p>\n<p>In processing image data, CV technology is used to track visitor behavior through image recognition and analysis. Cameras installed throughout the exhibition space capture facial expressions, eye movements, body posture, and motion. These image inputs are processed using CNNs to extract key behavioral features.<\/p>\n<p>Facial expression recognition, in particular, relies on a CNN-based deep learning model to identify and classify emotional states from captured facial images. High-resolution cameras record these facial images in real time within the museum environment. The raw data then undergoes several preprocessing steps\u2014including image denoising, grayscale conversion, and face detection\u2014to ensure input quality and accuracy.<\/p>\n<p>The core of facial expression recognition involves feature extraction using a CNN. The convolutional layers identify critical facial features, such as the shape, position, and dynamic changes of the eyebrows, eyes, and mouth. Pooling layers reduce data dimensionality and enhance feature robustness. Finally, fully connected layers classify the extracted features into specific emotional categories such as joy, surprise, anger, sadness, and interest.<\/p>\n<p>Dynamic layout optimization<\/p>\n<p>To enhance the accuracy and adaptability of facial expression recognition, a pre-trained neural network model is fine-tuned using an open-access facial expression dataset. This allows the model to better accommodate the diversity and complexity of real-world facial expressions in a museum setting. Data augmentation techniques\u2014such as rotation, translation, scaling, and image flipping\u2014are also applied to improve the model\u2019s robustness under varying environmental conditions.<\/p>\n<p>The results of facial expression recognition are then integrated with other behavioral data. For instance, by analyzing changes in facial expressions, the system can infer visitor interest. Strong positive emotions (e.g., joy or surprise) observed while viewing an exhibit suggest high engagement, prompting the system to prioritize or enhance that exhibit\u2019s location and interactivity. Conversely, negative emotions (e.g., boredom or confusion) may trigger adjustments to the exhibit\u2019s content or presentation to improve visitor engagement and satisfaction.<\/p>\n<p>Facial expression data can also be fused with eye-tracking and body posture information. Eye movement analysis helps identify which exhibits draw the most attention, offering valuable insights for spatial layout decisions. Meanwhile, body posture cues\u2014such as lingering, movement, or hesitation\u2014can reflect the visitor\u2019s level of engagement and intent to interact. This enables the system to dynamically tailor display modes or interactive content to suit different visitor preferences.<\/p>\n<p>Once processed, these image-based behavior features are quantified and fed into the RL model to guide spatial layout optimization. For example, if a visitor lingers in front of an exhibit while displaying high interest, the system may increase that exhibit\u2019s visibility or adjust its location to maximize impact. The CV processing flow is formalized in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ2\" target=\"_blank\" rel=\"noopener\">2<\/a>):<\/p>\n<p>$$\\:{D}_{feedback}=f({I}_{input},{W}_{model})$$<\/p>\n<p>\n                    (2)\n                <\/p>\n<p>\\(\\:{D}_{feedback}\\) represents the audience behavior data captured from the camera. \\(\\:{I}_{input}\\) is the input image data. \\(\\:{W}_{model}\\) is the weight of the CV model. The spatial layout optimization process is dynamic, with the model continuously adjusting based on real-time interactions with the museum environment and audience feedback. For instance, if the system detects excessive crowd density in a particular exhibition area, it can proactively modify the number of exhibits or adjust the spatial arrangement to alleviate congestion. These changes aim to enhance visitor comfort and engagement. The algorithmic structure of the spatial layout optimization model is illustrated in Fig.\u00a0<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig3\" target=\"_blank\" rel=\"noopener\">3<\/a>:<\/p>\n<p><b id=\"Fig3\" class=\"c-article-section__figure-caption\" data-test=\"figure-caption-text\">Fig. 3<\/b><a class=\"c-article-section__figure-link\" data-test=\"img-link\" data-track=\"click\" data-track-label=\"image\" data-track-action=\"view figure\" href=\"https:\/\/www.nature.com\/articles\/s41598-025-13408-2\/figures\/3\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" aria-describedby=\"Fig3\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/07\/41598_2025_13408_Fig3_HTML.png\" alt=\"figure 3\" loading=\"lazy\" width=\"001\" height=\"001\"\/><\/a><\/p>\n<p>AI-driven spatial layout optimization model structure.<\/p>\n<p>As shown in Fig.\u00a0<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig3\" target=\"_blank\" rel=\"noopener\">3<\/a>, the process begins with analyzing audience behavior data using CV to extract key features. These features are then fed into the RL model, which makes optimal decisions for spatial layout. The optimized layout is subsequently applied to the museum space, enabling dynamic adjustments. By combining RL and CV, the model can adapt to varying museum environments. Since exhibit locations, exhibition area mobility, and audience preferences constantly change, the model continuously learns and adjusts to enhance the visitor experience.<\/p>\n<p>Exhibition Liquidity is defined as the proportion of exhibits that visitors can efficiently access within a given time. Its calculation follows Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ3\" target=\"_blank\" rel=\"noopener\">3<\/a>):<\/p>\n<p>$$\\:L=\\frac{{N}_{visited}}{{N}_{total}}\\times\\:\\left(1-\\frac{{T}_{av{g}_{detour}}}{{T}_{shortest}}\\right)\\times\\:100\\%$$<\/p>\n<p>\n                    (3)\n                <\/p>\n<p>\\(\\:{N}_{visited}\\) is the number of exhibits actually visited by the audience. \\(\\:{N}_{total}\\) is the total number of exhibits in the exhibition hall. \\(\\:{T}_{av{g}_{detour}}\\) is the average time difference between the actual path of the audience and the theoretical shortest path, and \\(\\:{T}_{shortest}\\) is the total time of the theoretical shortest path. Equation\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ4\" target=\"_blank\" rel=\"noopener\">4<\/a>) shows the calculation of Path Optimization Rate:<\/p>\n<p>$$\\:P=\\left(1-\\frac{\\sum\\:_{i=1}^{n}\\left({D}_{i}-{D}_{min}\\right)}{\\sum\\:_{i=1}^{n}{D}_{min}}\\right)\\times\\:100\\%$$<\/p>\n<p>\n                    (4)\n                <\/p>\n<p>\\(\\:{D}_{i}\\) is the actual path length of the \\(\\:i\\)-th audience. \\(\\:{D}_{min}\\) is the theoretical shortest path length of the corresponding exhibit sequence. The balance of crowd density distribution is measured by Gini Coefficient, as shown in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ5\" target=\"_blank\" rel=\"noopener\">5<\/a>):<\/p>\n<p>$$\\:G=\\frac{\\sum\\:_{i=1}^{k}\\sum\\:_{j=1}^{k}|{x}_{i}-{x}_{j}|}{2k\\sum\\:_{i=1}^{k}{x}_{i}}$$<\/p>\n<p>\n                    (5)\n                <\/p>\n<p>\\(\\:{x}_{i}\\) is the density of people in the \\(\\:i\\)-th exhibition area, and \\(\\:k\\) is the total number of exhibition areas. The Frequency of Congested Areas is defined as the proportion of exhibition areas where the daily pedestrian density exceeds the threshold (5 people\/m2), as shown in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ6\" target=\"_blank\" rel=\"noopener\">6<\/a>):<\/p>\n<p>$$\\:F=\\frac{\\sum\\:_{t=1}^{T}{C}_{t}}{T\\times\\:k}\\times\\:100\\%$$<\/p>\n<p>\n                    (6)\n                <\/p>\n<p>\\(\\:{C}_{t}\\) is the number of congested exhibition areas in the \\(\\:t\\) hour. \\(\\:T\\) is the total observation time (hours), and \\(\\:k\\) is the total number of exhibition areas.<\/p>\n<p>This study referenced standard museum industry guidelines on crowd density, which recommended limiting exhibition areas to a maximum of five people per square meter to ensure visitor comfort. These guidelines were further adjusted to reflect the specific spatial layout and exhibit types of the target museum. A two-week on-site test was carried out to collect visitor flow data across different time periods. Simulations then assessed space utilization and visitor experience under varying density thresholds. The analysis showed that maintaining a density of five people per square meter effectively prevented congestion, maximized space use, and improved the overall visitor experience.<\/p>\n<p>                           <b>Interactive experience optimization model<\/b><br \/>\n                        Design of personalized recommendation model<\/p>\n<p>To further enhance immersion in the museum space and increase audience engagement, this study also developed an AI-based interactive experience optimization model, as illustrated in Fig.\u00a0<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig4\" target=\"_blank\" rel=\"noopener\">4<\/a>, alongside the spatial layout optimization.<\/p>\n<p><b id=\"Fig4\" class=\"c-article-section__figure-caption\" data-test=\"figure-caption-text\">Fig. 4<\/b><a class=\"c-article-section__figure-link\" data-test=\"img-link\" data-track=\"click\" data-track-label=\"image\" data-track-action=\"view figure\" href=\"https:\/\/www.nature.com\/articles\/s41598-025-13408-2\/figures\/4\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" aria-describedby=\"Fig4\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/07\/41598_2025_13408_Fig4_HTML.png\" alt=\"figure 4\" loading=\"lazy\" width=\"001\" height=\"001\"\/><\/a><\/p>\n<p>Optimization model of museum interactive experience.<\/p>\n<p>A key component of the interactive experience optimization model shown in Fig.\u00a0<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig4\" target=\"_blank\" rel=\"noopener\">4<\/a> is an intelligent recommendation system driven by audience behavior data. By analyzing real-time data such as interest points, dwell time, and interaction frequency, the model infers individual preferences and automatically adjusts the displayed content to provide a personalized experience. For example, if the model detects that a visitor shows strong interest in a particular exhibit\u2014indicated by prolonged viewing or frequent interactions\u2014the system dynamically recommends additional information about that exhibit or other exhibits with similar themes. This recommendation process is implemented using a Collaborative Filtering approach.<\/p>\n<p>Hybrid recommendation strategy and cold start processing<\/p>\n<p>This study adopts an item-based collaborative filtering algorithm, a method widely used in product recommendation systems. It analyzes user behavior data\u2014such as points of interest, dwell time, and interaction frequency\u2014to calculate similarities between exhibits and recommend others with similar themes or styles. Unlike traditional product recommendations, museum exhibition recommendations prioritize enhancing visitor immersion and interactive participation. Building on the proven success of collaborative filtering in e-commerce, this study adapts the approach to the museum context by using visitor behavior as implicit feedback. This enables the creation of a personalized recommendation model that not only captures a visitor\u2019s interest in specific exhibits but also uncovers potential preferences across different exhibition content by integrating various behavior data. The model adjusts display content in real time to match these inferred preferences. Moreover, the collaborative filtering algorithm operates in tandem with RL, CV, and emotion computing technologies. This integration enables dynamic optimization throughout the entire process\u2014from data collection to exhibit recommendation and interaction mode adjustment. As a result, the algorithm retains its strength in accurately capturing user preferences while effectively meeting the museum\u2019s higher demands for personalization, interactivity, and immersive experience. By analyzing audience behavior data, including interest points, dwell time, and interaction frequency, the system infers visitor preferences and automatically tailors display content to provide a personalized experience. The algorithm calculates exhibit similarity based on this behavior data and recommends exhibits related to those in which visitors show interest. This approach effectively delivers personalized exhibition recommendations, enhancing the overall visitor experience. To further improve recommendation accuracy and address data sparsity issues, the model incorporates implicit feedback such as browsing history and dwell time. These additional data points help capture visitor preferences more comprehensively, thereby boosting the recommendation system\u2019s performance.<\/p>\n<p>To address the cold start problem in collaborative filtering, this study implements a hybrid recommendation strategy. For newly added exhibits, the system begins by extracting content-based features such as exhibit type, historical period, material, and thematic tags. It then uses cosine similarity to compare these features with those of existing exhibits, generating an initial recommendation list. As interaction data for the new exhibit accumulates, the system gradually transitions to a collaborative filtering-based recommendation approach. For first-time visitors, the system assigns them to predefined audience groups based on demographic information, such as age and cultural background. It then recommends exhibits that have historically been favored by that group. As visitor behavior data becomes available, the model dynamically updates its parameters and shifts toward a personalized recommendation based on collaborative filtering. This hybrid approach effectively balances recommendation accuracy and data availability, ensuring reliable performance during the cold start phase for both new users and new exhibits.<\/p>\n<p>Adaptive adjustment of interactive mode<\/p>\n<p>In the museum\u2019s interactive experience optimization model, adaptive adjustment of interaction modes is incorporated alongside content recommendation. Interactive experiences may include touchscreen displays, AR\/VR environments, audio feedback, and other formats. By analyzing real-time audience behavior, the model can intelligently modify the interaction mode to enhance engagement and participation. For instance, if the system detects low user engagement with touchscreen displays\u2014such as infrequent touch activity\u2014it can automatically switch to alternative modes like VR experiences or AR visualizations to improve interactivity.<\/p>\n<p>This adaptive process relies on behavioral analysis and ML algorithms. The system continuously monitors audience behavior, including dwell time, viewing frequency, and interaction rates. If the interaction frequency within a particular exhibit area drops below a threshold, the system dynamically adjusts the display format to better capture attention and increase involvement. For example, introducing AR features to provide richer, more immersive content can re-engage visitors with low initial interest. The decision to change the interaction mode is guided by real-time data analysis, ensuring that the interactive experience remains responsive and tailored to visitor behavior. This optimization process is formally represented in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ7\" target=\"_blank\" rel=\"noopener\">7<\/a>).<\/p>\n<p>$$\\:{I}_{t}=f({B}_{t},{A}_{t})$$<\/p>\n<p>\n                    (7)\n                <\/p>\n<p>\\(\\:{I}_{t}\\) represents the interaction mode provided by the model for the audience at time t. \\(\\:{B}_{t}\\) represents the behavior data of the audience at time t (such as interaction frequency, residence time, etc.). \\(\\:{A}_{t}\\) represents the display characteristics of the current exhibit (such as exhibit type, interaction mode, etc.). Through this equation, the model can adjust the content and mode of interaction in real time.<\/p>\n<p>Audiences are grouped based on demographic attributes such as age and cultural background. Age categories may include adolescents, young and middle-aged adults, and seniors, while cultural backgrounds can be classified into local and international cultures. These groupings can be established through clustering analysis of historical museum visitor data or by referencing relevant academic research and the specific context of the museum. Before the system is officially deployed, a small group of visitors with known demographic information is invited for testing. This allows observation of whether the initial recommendations based on demographic profiles align with the actual interests of these visitors. Based on the test results, the predefined audience groupings are refined and optimized. As more data is collected, online learning algorithms are employed to continuously update and improve the recommendation model. For visitors who were initially misclassified, the system automatically adjusts future recommendations based on their behavioral data, gradually increasing recommendation accuracy. Additionally, visitors can provide direct feedback on recommendations\u2014such as through likes or bookmarks\u2014which is also incorporated into the system to further enhance recommendation strategies.<\/p>\n<p>When switching interactive modes, the system relies on real-time data to ensure a seamless transition between touchscreen and VR\/AR experiences. If the interaction frequency in touchscreen mode is low, the system automatically prompts the audience to switch to VR or AR, offering a more immersive experience. Intelligent algorithms and sensors support this transition, ensuring it occurs smoothly and without disrupting the visitor\u2019s engagement. When a user shows limited responsiveness to touchscreen interaction, the system analyzes their behavior history and current activity to recommend alternative display modes. User identification and behavior tracking technologies are employed to preserve interaction history during mode switches, maintaining a continuous and uninterrupted experience. For instance, when transitioning from touchscreen to VR or AR, the system ensures that previously viewed content is carried over and displayed in the new mode.<\/p>\n<p>This adaptive optimization enables the model to dynamically adjust interactive experiences and flexibly switch between different display modes. As a result, the system delivers a more personalized and immersive museum experience that aligns closely with visitors\u2019 needs and preferences.<\/p>\n<p>Emotional computing module<\/p>\n<p>To further enhance audience immersion, the interactive experience optimization model integrates EC and situational awareness technologies to detect the audience\u2019s emotional state in real time. Based on these emotional changes, the system dynamically adjusts both the exhibit content and the interaction mode. By analyzing physiological and behavioral cues\u2014such as facial expressions, vocal tone, and body posture\u2014the model can assess the visitor\u2019s emotional response and make appropriate adjustments. For instance, if a visitor appears engaged or curious, the system may display more detailed content or introduce interactive elements. Conversely, if signs of fatigue or boredom are detected, the system may shift to more entertaining or stimulating content to sustain interest and participation.<\/p>\n<p>The objective of EC is to predict and interpret the audience\u2019s emotional states using multi-dimensional input signals. This process is modeled as a multi-input, multi-output function, formally expressed in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ8\" target=\"_blank\" rel=\"noopener\">8<\/a>):<\/p>\n<p>$$\\:{E}_{t}=g({F}_{t},{V}_{t},{A}_{t})$$<\/p>\n<p>\n                    (8)\n                <\/p>\n<p>\\(\\:{E}_{t}\\) represents the emotional state of the audience at time t, and it is a numerical emotional score (such as positive, negative, excited, calm, etc.). \\(\\:{F}_{t}\\) is the audience\u2019s facial expression data, which is extracted by facial expression recognition algorithm, including facial features such as smile, frown and wide eyes. In this study, a CNN is employed to extract facial features and learn the mapping between these features and emotional states using a large set of labeled data. The model recognizes expressions such as smiles and frowns through a deep neural network and maps them to emotional categories (e.g., happiness, anger, sadness). These features serve as indicators of audience emotional changes. When multiple viewers are present simultaneously, the system analyzes each individual\u2019s facial expression data independently. It identifies and classifies each viewer\u2019s facial expressions, determining their emotional state within a specific time frame. The emotional data from all viewers are then aggregated to calculate the overall emotional distribution during that period. This approach avoids the complexity of merging raw facial data and ensures accurate emotional analysis.<\/p>\n<p>\\(\\:{V}_{t}\\) represents the audience\u2019s voice tone data. Voice emotion is analyzed using speech recognition technology, focusing on features such as intonation, speech rate, and pitch. These vocal cues often convey subtle emotional states like happiness, surprise, or anxiety. \\(\\:{A}_{t}\\) refers to body posture data, which captures audience movements and postures (e.g., standing, sitting, waving) through sensors or CV. These physical behaviors can indicate emotional and physiological responses. After collecting data from facial expressions, voice tone, and body posture, a fusion model integrates these multimodal signals to derive a comprehensive emotional assessment. The final emotional score is calculated using a weighted approach, as defined in Eq.\u00a0(<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"equation anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Equ9\" target=\"_blank\" rel=\"noopener\">9<\/a>):<\/p>\n<p>$$\\:{E}_{t}={w}_{F}\\cdot\\:{F}_{t}+{w}_{V}\\cdot\\:{V}_{t}+{w}_{A}\\cdot\\:{A}_{t}$$<\/p>\n<p>\n                    (9)\n                <\/p>\n<p>\\(\\:{w}_{F}\\), \\(\\:{w}_{V}\\) and \\(\\:{w}_{A}\\) are the weights of facial expression, voice tone and body posture, respectively, indicating the importance of different signals to EC. These weights can be optimized through the training process to learn the contribution of different types of signals to emotional judgment.<\/p>\n<p>In the EC module, the weights for facial expressions, voice tone, and body posture (\\(\\:{w}_{F},{w}_{V},{w}_{A}\\)) are learned through a supervised learning framework. A multimodal dataset comprising 1,000 samples of audience behavior\u2014including facial expressions, audio recordings, and posture videos\u2014is annotated by experts with emotion labels (positive, neutral, or negative). A fully connected neural network integrates these multimodal features, and the weights are optimized using gradient descent to minimize the cross-entropy loss in emotion classification.<\/p>\n<p>The experiment collected emotional data from audiences with diverse cultural backgrounds, including facial expressions, voice tone, and body posture. These data were analyzed using ML algorithms. By comparing the emotion classification accuracy under different weight combinations, the current weight allocation was determined. Additionally, cross-cultural validation experiments were conducted, revealing that the weights exhibit good adaptability and stability across various cultural contexts. However, slight adjustments may be necessary for specific cultures. These findings indicate that the model possesses a certain degree of generalizability while allowing for personalized adaptation based on cultural differences. The final weights, determined through 5-fold cross-validation, are \\(\\:{w}_{F}=0.52\\pm\\:0.03\\), \\(\\:{w}_{V}=0.28\\pm\\:0.02\\) and \\(\\:{w}_{A}=0.20\\pm\\:0.02\\) (mean\u2009\u00b1\u2009standard deviation). Experiments show that this weight combination achieves an emotion classification accuracy of 87.6%, significantly higher than uniform weights (76.2%) and single-modality models (facial: 79.3%, voice: 68.5%, posture: 62.1%).<\/p>\n<p>A Multi-technology collaborative optimization framework<\/p>\n<p>This study establishes a dynamic optimization feedback loop through a collaborative multi-technology framework, as illustrated in Fig.\u00a0<a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig5\" target=\"_blank\" rel=\"noopener\">5<\/a>. CV continuously tracks audience behavior, including movement patterns and dwell time. EC analyzes facial expressions, voice tone, and body posture to generate emotional scores\u2014such as interest or confusion. VR devices capture interactions in virtual environments, including hotspot clicks and navigation paths. Data from these three sources are integrated into a RL model, which dynamically adjusts exhibit layouts and interaction logic based on a multi-dimensional reward function. This function assigns weights of 0.4 to dwell time, 0.3 to path efficiency, and 0.3 to emotional feedback. For instance, if EC detects persistently low emotional engagement in a particular area, the RL model can trigger a transition to AR mode, presenting dynamic content such as 3D animations to recapture audience attention. The optimized layouts and interaction modes are implemented via VR devices and physical space sensors. Their effectiveness is evaluated through audience surveys measuring immersion and knowledge acquisition. The results are continuously fed back into the AI processing layer, enabling ongoing optimization. This integrated system is the first to achieve a multi-modal, closed-loop feedback cycle of \u201cbehavior\u2013emotion\u2013space,\u201d overcoming the limitations of traditional single-technology approaches.<\/p>\n<p><b id=\"Fig5\" class=\"c-article-section__figure-caption\" data-test=\"figure-caption-text\">Fig. 5<\/b><a class=\"c-article-section__figure-link\" data-test=\"img-link\" data-track=\"click\" data-track-label=\"image\" data-track-action=\"view figure\" href=\"https:\/\/www.nature.com\/articles\/s41598-025-13408-2\/figures\/5\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" aria-describedby=\"Fig5\" src=\"https:\/\/www.europesays.com\/uk\/wp-content\/uploads\/2025\/07\/41598_2025_13408_Fig5_HTML.png\" alt=\"figure 5\" loading=\"lazy\" width=\"001\" height=\"001\"\/><\/a><\/p>\n<p>Multi-technology collaborative optimization framework.<\/p>\n<p>Figure <a data-track=\"click\" data-track-label=\"link\" data-track-action=\"figure anchor\" href=\"http:\/\/www.nature.com\/articles\/s41598-025-13408-2#Fig5\" target=\"_blank\" rel=\"noopener\">5<\/a> illustrates a three-layer multi-technology collaborative optimization framework. The data acquisition layer collects real-time information on audience behavior, emotions, and interactions using CV sensors, EC analysis devices, and VR equipment. The AI processing layer integrates these data\u2014such as movement paths, emotional scores, and VR interactions\u2014through a RL model. This model dynamically optimizes spatial layouts by adjusting exhibit positions and refines interaction modes, including switching between AR and VR. The feedback execution layer applies these optimized results to both physical and virtual environments in real time and assesses audience experience through surveys. Together, these layers create a closed-loop system for continuous improvement.<\/p>\n","protected":false},"excerpt":{"rendered":"Optimization model of Spatial layout driven by AI Figure 1 presents the design of the spatial optimization model,&hellip;\n","protected":false},"author":2,"featured_media":296906,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3164],"tags":[1942,3284,110021,8173,3965,110020,7462,3966,31953,222,70,110019,53,16,15],"class_list":{"0":"post-296905","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-artificial-intelligence","9":"tag-computing","10":"tag-emotional-computing","11":"tag-engineering","12":"tag-humanities-and-social-sciences","13":"tag-interactive-experience","14":"tag-mathematics-and-computing","15":"tag-multidisciplinary","16":"tag-museum","17":"tag-psychology","18":"tag-science","19":"tag-spatial-layout-optimization","20":"tag-technology","21":"tag-uk","22":"tag-united-kingdom"},"share_on_mastodon":{"url":"https:\/\/pubeurope.com\/@uk\/114927446687798908","error":""},"_links":{"self":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/296905","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/comments?post=296905"}],"version-history":[{"count":0,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/posts\/296905\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media\/296906"}],"wp:attachment":[{"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/media?parent=296905"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/categories?post=296905"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.europesays.com\/uk\/wp-json\/wp\/v2\/tags?post=296905"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}