In the realm of disaster management and public safety, the accurate monitoring of crowds during emergencies has become increasingly vital. Natural calamities, such as earthquakes and tsunamis, demand swift and precise evacuation strategies that hinge upon real-time data regarding crowd movement and congestion points. Technological innovations are at the forefront of such efforts, and among these, LiDAR (Light Detection and Ranging) technology stands out for its ability to create three-dimensional representations of human targets by sending out laser pulses and measuring their reflections. However, traditional LiDAR systems with high-density scanning capabilities remain prohibitively expensive and limited in their field of view, restricting their deployment in broad public environments.
Addressing these limitations is non-repetitive scanning LiDAR—an emerging affordable technology that provides a wider field of view suited for large-area monitoring. Despite this advantage, it suffers from inherent sparsity and irregularity in its depth measurements. The scans often produce sparse, discontinuous point cloud images that fail to deliver the comprehensive depth information required for effective crowd assessment. Moreover, the irregular sampling patterns intrinsic to non-repetitive LiDAR complicate the reconstruction of depth maps, curtailing their operational feasibility in fast-evolving scenarios such as emergency evacuations.
A groundbreaking approach to surmount this challenge has been developed by Zixuan Zhang, a doctoral researcher from Doshisha University’s Graduate School of Science and Engineering, alongside distinguished collaborators including Professors Nobutaka Tsujiuchi and Akihito Ito from Doshisha University, and Professor Hirosuke Horii from Kokushikan University. Their innovative research, recently published in the prestigious volume 36 of Transportation Research Interdisciplinary Perspectives, introduces a novel RGB-guided depth completion framework tailored specifically for the unique sampling profiles of non-repetitive scanning LiDAR.
The core innovation lies in leveraging RGB (red, green, and blue) color data to guide the depth recovery process. This method integrates advanced computational techniques such as confidence-aware bilateral filtering and masked reconstruction, combined with a self-consistent parameter optimization algorithm that obviates the need for dense ground-truth depth data. By doing so, the framework reconstructs dense, continuous depth maps from initially sparse and disjointed LiDAR readings, maintaining structural integrity without succumbing to excessive smoothing or erroneous gap-filling.
A significant hurdle in developing and validating such algorithms has been the lack of publicly available datasets that authentically replicate the sparse, irregular sampling patterns characteristic of non-repetitive LiDAR systems. Addressing this gap, the research team ingeniously simulated non-repetitive LiDAR scans by employing a 3D human motion dataset to generate dense ground-truth depth maps and paired color images. By mimicking rotational scanning dynamics analogous to LiDAR’s operational principles, they produced over 30,000 synthetic samples that authentically represent the sparse measurement conditions encountered in practice.
Testing their RGB-guided depth completion method on this extensive synthetic dataset yielded compelling results. The reconstructed depth maps demonstrated markedly enhanced structural coherence and accuracy compared to conventional filtering techniques. Notably, the method adeptly preserved human silhouette definition and salient depth gradients, key for accurate crowd monitoring. Furthermore, its computational lightness makes it eminently suitable for real-time deployment in crisis response systems, where rapid processing and reliability are paramount.
However, the researchers also caution that the reliance on RGB imagery introduces sensitivity to varying environmental conditions, including fluctuations in lighting, saturation, motion blur, and potential misalignment between color and depth data streams. These factors can influence the precision of depth reconstruction and necessitate rigorous real-world testing. The authors underscore that their framework must be validated against authentic field-collected data before it can be fully integrated into operational dynamic evacuation guidance systems.
Yet, the implications of this technology extend beyond immediate disaster contexts. As cities evolve into complex smart infrastructures, the integration of LiDAR sensors for situational awareness will become ubiquitous. Robust depth reconstruction algorithms, exemplified by this research, will form the backbone of sophisticated public safety networks, enabling enhanced real-time monitoring of urban dynamics and rapid hazard response.
Zhang’s research heralds a new paradigm in computational depth recovery from sparse LiDAR data, offering a promising pathway for more reliable, large-scale human crowd analysis. By enhancing the fidelity of depth perception through sophisticated RGB-guided processing techniques, this work directly contributes to safer, more resilient urban environments during disasters.
The multidisciplinary nature of this project—merging computational modeling, optical sensing, and intelligent system design—showcases the transformative potential of collaborative scientific inquiry. With continued refinement and eventual real-world validation, this technology stands to play a vital role in saving lives and mitigating risks in future emergency scenarios worldwide.
As the research continues to evolve, deploying these depth completion methods in diverse environments and integrating them with other sensor modalities will be critical next steps. These efforts promise to realize fully automated, scalable crowd monitoring platforms that can dynamically adjust to real-time conditions and provide actionable intelligence for first responders and city planners alike.
In summary, the work by Zhang et al. represents a pivotal advancement in the intersection of remote sensing and disaster management technologies. By addressing the intricate challenges posed by non-repetitive LiDAR scanning through inventive computational strategies, this research sets the foundation for next-generation situational awareness systems, empowering safer urban evacuations and vibrant smart city functionalities.
Subject of Research: Not applicable
Article Title: RGB-guided sparse depth completion for non-repetitive scanning LiDAR in crowd dynamics analysis
News Publication Date: 1-Mar-2026
Web References:
https://doi.org/10.1016/j.trip.2026.101879
Image Credits: Zixuan Zhang from Doshisha University, Japan
Keywords
LiDAR, Disaster management, Remote sensing, Depth completion, Crowd dynamics, Non-repetitive scanning, RGB-guided reconstruction, Computational modeling, Emergency evacuation, Smart city, Situational awareness
Tags: 3D human target detection technologyadvanced crowd mapping technologyaffordable wide field LiDAR solutionschallenges in LiDAR depth map reconstructiondisaster management crowd controlevacuation strategy optimization toolsinnovative public safety technologieslaser pulse crowd detection methodslimitations of traditional LiDAR in public safetynon-repetitive scanning LiDAR applicationsreal-time crowd monitoring systemssmart cameras for emergency evacuations



