Abstract:3D scene reconstruction is a critical research topic in autonomous driving, robotics, and related fields, with extensive applications in navigation mapping, environmental interaction, and virtual/augmented reality tasks. Current deep learning-based reconstruction methods can be primarily categorized into five groups from the perspectives of scene representation and core modeling techniques: cost volume-based depth estimation methods, truncated signed distance function (TSDF)-based voxel approaches, transformer architecture-based large-scale feedforward methods, multilayer perceptron (MLP)-based neural radiance fields (NeRF), and 3D Gaussian Splatting (3DGS). Each category exhibits unique strengths and limitations. The emerging 3DGS method distinguishes itself by explicitly representing scenes through Gaussian functions while achieving rapid scene rendering and novel view synthesis via efficient rasterization operations. Its most significant advantage lies in diverging from NeRF''s MLP-based scene representation paradigm - 3DGS ensures both efficient rendering and interpretable editable scene modeling, thereby paving the way for accurate 3D scene reconstruction. However, 3DGS still faces numerous challenges in practical scene reconstruction applications. Based on this analysis, this paper first provides a concise introduction to 3DGS fundamentals and conducts comparative analysis with the aforementioned four categories. Following a systematic survey of existing 3DGS reconstruction algorithms, we summarize the key challenges addressed by these methods and review current research progress on core technical difficulties through representative case studies. Finally, we prospect potential future research directions worthy of exploration.