Large-scale 3D environmental modelling and visualisation for flood hazard warning.
View/ Open
Chen+WANG_2009PhDThesis.pdf (9.558Mb)
Download
Publication date
2009-08-24T11:15:12ZAuthor
Wang, ChenSupervisor
Wan, Tao RuanPalmer, Ian J.
Keyword
Large-scaleDigital Terrain Model
Active Contour Based Image Segmentation
Intelligent Shape Match Method
3D Environmental reconstruction
Flood Spreading ModelFlood Spreading Model
Finite Volume Hydrodynamic Model
3D flood simulation
Rights
The University of Bradford theses are licenced under a Creative Commons Licence.
Institution
University of BradfordDepartment
Department of Creative Technology. School of Computing, Informatics and Media.Awarded
2009
Metadata
Show full item recordAbstract
3D environment reconstruction has received great interest in recent years in areas such as city planning, virtual tourism and flood hazard warning. With the rapid development of computer technologies, it has become possible and necessary to develop new methodologies and techniques for real time simulation for virtual environments applications. This thesis proposes a novel dynamic simulation scheme for flood hazard warning. The work consists of three main parts: digital terrain modelling; 3D environmental reconstruction and system development; flood simulation models. The digital terrain model is constructed using real world measurement data of GIS, in terms of digital elevation data and satellite image data. An NTSP algorithm is proposed for very large data assessing, terrain modelling and visualisation. A pyramidal data arrangement structure is used for dealing with the requirements of terrain details with different resolutions. The 3D environmental reconstruction system is made up of environmental image segmentation for object identification, a new shape match method and an intelligent reconstruction system. The active contours-based multi-resolution vector-valued framework and the multi-seed region growing method are both used for extracting necessary objects from images. The shape match method is used with a template in the spatial domain for a 3D detailed small scale urban environment reconstruction. The intelligent reconstruction system is designed to recreate the whole model based on specific features of objects for large scale environment reconstruction. This study then proposes a new flood simulation scheme which is an important application of the 3D environmental reconstruction system. Two new flooding models have been developed. The first one is flood spreading model which is useful for large scale flood simulation. It consists of flooding image spatial segmentation, a water level calculation process, a standard gradient descent method for energy minimization, a flood region search and a merge process. The finite volume hydrodynamic model is built from shallow water equations which is useful for urban area flood simulation. The proposed 3D urban environment reconstruction system was tested on our simulation platform. The experiment results indicate that this method is capable of dealing with complicated and high resolution region reconstruction which is useful for many applications. When testing the 3D flood simulation system, the simulation results are very close to the real flood situation, and this method has faster speed and greater accuracy of simulating the inundation area in comparison to the conventional flood simulation modelsType
ThesisQualification name
PhDCollections
Related items
Showing items related by title, author, creator and subject.
-
Interpretation, Identification and Reuse of Models. Theory and algorithms with applications in predictive toxicology.Neagu, Daniel; Ridley, Mick J.; Travis, Kim; Palczewska, Anna Maria (University of BradfordSchool of Electrical Engineering and Computer Science, 2015-07-15)This thesis is concerned with developing methodologies that enable existing models to be effectively reused. Results of this thesis are presented in the framework of Quantitative Structural-Activity Relationship (QSAR) models, but their application is much more general. QSAR models relate chemical structures with their biological, chemical or environmental activity. There are many applications that offer an environment to build and store predictive models. Unfortunately, they do not provide advanced functionalities that allow for efficient model selection and for interpretation of model predictions for new data. This thesis aims to address these issues and proposes methodologies for dealing with three research problems: model governance (management), model identification (selection), and interpretation of model predictions. The combination of these methodologies can be employed to build more efficient systems for model reuse in QSAR modelling and other areas. The first part of this study investigates toxicity data and model formats and reviews some of the existing toxicity systems in the context of model development and reuse. Based on the findings of this review and the principles of data governance, a novel concept of model governance is defined. Model governance comprises model representation and model governance processes. These processes are designed and presented in the context of model management. As an application, minimum information requirements and an XML representation for QSAR models are proposed. Once a collection of validated, accepted and well annotated models is available within a model governance framework, they can be applied for new data. It may happen that there is more than one model available for the same endpoint. Which one to chose? The second part of this thesis proposes a theoretical framework and algorithms that enable automated identification of the most reliable model for new data from the collection of existing models. The main idea is based on partitioning of the search space into groups and assigning a single model to each group. The construction of this partitioning is difficult because it is a bi-criteria problem. The main contribution in this part is the application of Pareto points for the search space partition. The proposed methodology is applied to three endpoints in chemoinformatics and predictive toxicology. After having identified a model for the new data, we would like to know how the model obtained its prediction and how trustworthy it is. An interpretation of model predictions is straightforward for linear models thanks to the availability of model parameters and their statistical significance. For non linear models this information can be hidden inside the model structure. This thesis proposes an approach for interpretation of a random forest classification model. This approach allows for the determination of the influence (called feature contribution) of each variable on the model prediction for an individual data. In this part, there are three methods proposed that allow analysis of feature contributions. Such analysis might lead to the discovery of new patterns that represent a standard behaviour of the model and allow additional assessment of the model reliability for new data. The application of these methods to two standard benchmark datasets from the UCI machine learning repository shows a great potential of this methodology. The algorithm for calculating feature contributions has been implemented and is available as an R package called rfFC.
-
Interpreting random forest models using a feature contribution methodPalczewska, Anna Maria; Palczewski, J.; Marchese-Robinson, R.M.; Neagu, Daniel (2013)
-
Hybrid Dynamic Modelling of Engine Emissions on Multi-Physics Simulation Platform. A Framework Combining Dynamic and Statistical Modelling to Develop Surrogate Models of System of Internal Combustion Engine for Emission ModellingCampean, Felician; Neagu, Daniel; Pant, Gaurav (University of BradfordFaculty of Engineering and Informatics, 2018)The data-driven models used for the design of powertrain controllers are typically based on the data obtained from steady-state experiments. However, they are only valid under stable conditions and do not provide any information on the dynamic behaviour of the system. In order to capture this behaviour, dynamic modelling techniques are intensively studied to generate alternative solutions for engine mapping and calibration problem, aiming to address the need to increase productivity (reduce development time) and to develop better models for the actual behaviour of the engine under real-world conditions. In this thesis, a dynamic modelling approach is presented undertaken for the prediction of NOx emissions for a 2.0 litre Diesel engine, based on a coupled pre-validated virtual Diesel engine model (GT- Suite ® 1-D air path model) and in-cylinder combustion model (CMCL ® Stochastic Reactor Model Engine Suite). In the context of the considered Engine Simulation Framework, GT Suite + Stochastic Reactor Model (SRM), one fundamental problem is to establish a real time stochastic simulation capability. This problem can be addressed by replacing the slow combustion chemistry solver (SRM) with an appropriate NOx surrogate model. The approach taken in this research for the development of this surrogate model was based on a combination of design of dynamic experiments run on the virtual diesel engine model (GT- Suite), with a dynamic model fitted for the parameters required as input to the SRM, with a zonal design of experiments (DoEs), using Optimal Latin Hypercubes (OLH), run on the SRM model. A response surface model was fitted on the predicted NOx from the SRM OLH DoE data. This surrogate NOx model was then used to replace the computationally expensive SRM simulation, enabling real-time simulations of transient drive cycles to be executed. The performance of the approach was validated on a simulated NEDC drive cycle, against experimental data collected for the engine case study. The capability of methodology to capture the transient trends of the system shows promising results and will be used for the development of global surrogate prediction models for engine-out emissions.