Search

found 23 results

Research papers, University of Canterbury Library

The 2015 New Zealand strong-motion database provides a wealth of new strong motion data for engineering applications. An important component of this database is the compilation of new site metadata, describing the soil conditions and site response at GeoNet strong motion stations. We have assessed and compiled four key site parameters for the ~460 GeoNet stations that recorded significant historical ground motions. Parameters include: site classification (NZS1170.5), Vs30, fundamental site period (Tsite) and depth to bedrock (Z1.0, i.e. depth to material with Vs > 1000 m/s). In addition, we have assigned a quality estimate (Quality 1 – 3) to these parameters to provide a qualitative estimate of the uncertainty. New highquality Tsite estimates have largely been obtained from newly available HVSR amplification curves and spectral ratios from inversion of regional strong motion data that has been reconciled with available geological information. Good quality Vs30 estimates, typically in urban centres, have also been incorporated following recent studies. Where site-specific measurements of Vs30 are not available, Vs30 is estimated based on surface geology following national Vs30 maps. New Z1.0 values have been provided from 3D subsurface models for Canterbury and Wellington. This database will be used in efforts to guide development and testing of new and existing ground motion prediction models in New Zealand. In particular, it will allow reexamination of the most important site parameters that control and predict site response in a New Zealand setting. Furthermore, it can be used to provide information about suitable rock reference sites for seismological research, and as a guide to site-specific references in the literature. We discuss compilation of the database, preliminary insights so far, and future directions.

Research papers, University of Canterbury Library

Advanced seismic effective-stress analysis is used to scrutinize the liquefaction performance of 55 well-documented case-history sites from Christchurch. The performance of these sites during the 2010-2011 Canterbury earthquake sequence varied significantly, from no liquefaction manifestation at the ground surface (in any of the major events) to severe liquefaction manifestation in multiple events. For the majority of the 55 sites, the simplified liquefaction evaluation procedures, which are conventionally used in engineering practice, could not explain these dramatic differences in the manifestation. Detailed geotechnical characterization and subsequent examination of the soil profile characteristics of the 55 sites identified some similarities but also important differences between sites that manifested liquefaction in the two major events of the sequence (YY-sites) and sites that did not manifest liquefaction in either event (NN-sites). In particular, while the YY-sites and NN-sites are shown to have practically identical critical layer characteristics, they have significant differences with regard to their deposit characteristics including the thickness and vertical continuity of their critical zones and liquefiable materials. A CPT-based effective stress analysis procedure is developed and implemented for the analyses of the 55 case history sites. Key features of this procedure are that, on the one hand, it can be fully automated in a programming environment and, on the other hand, it is directly equivalent (in the definition of cyclic resistance and required input data) to the CPT-based simplified liquefaction evaluation procedures. These features facilitate significantly the application of effective-stress analysis for simple 1D free-field soil-column problems and also provide a basis for rigorous comparisons of the outcomes of effective-stress analyses and simplified procedures. Input motions for the analyses are derived using selected (reference) recordings from the two major events of the 2010-2011 Canterbury earthquake sequence. A step-by-step procedure for the selection of representative reference motions for each site and their subsequent treatment (i.e. deconvolution and scaling) is presented. The focus of the proposed procedure is to address key aspects of spatial variability of ground motion in the near-source region of an earthquake including extended-source effects, path effects, and variation in the deeper regional geology.

Research papers, University of Canterbury Library

Post-earthquake cordons have been used after seismic events around the world. However, there is limited understanding of cordons and how contextual information of place such as geography, socio-cultural characteristics, economy, institutional and governance structure etc. affect decisions, operational procedures as well as spatial and temporal attributes of cordon establishment. This research aims to fill that gap through a qualitative comparative case study of two cities: Christchurch, New Zealand (Mw 6.2 earthquake, February 2011) and L’Aquila, Italy (Mw 6.3 earthquake, 2009). Both cities suffered comprehensive damage to its city centre and had cordons established for extended period. Data collection was done through purposive and snowball sampling methods whereby 23 key informants were interviewed in total. The interviewee varied in their roles and responsibilities i.e. council members, emergency managers, politicians, business/insurance representatives etc. We found that cordons were established to ensure safety of people and to maintain security of place in both the sites. In both cities, the extended cordon was met with resistance and protests. The extent and duration of establishment of cordon was affected by recovery approach taken in the two cities i.e. in Christchurch demolition was widely done to support recovery allowing for faster removal of cordons where as in L’Aquila, due to its historical importance, the approach to recovery was based on saving all the buildings which extended the duration of cordon. Thus, cordons are affected by site specific needs. It should be removed as soon as practicable which could be made easier with preplanning of cordons.

Research papers, Lincoln University

Globally, the maximum elevations at which treelines are observed to occur coincide with a 6.4 °C soil isotherm. However, when observed at finer scales, treelines display a considerable degree of spatial complexity in their patterns across the landscape and are often found occurring at lower elevations than expected relative to the global-scale pattern. There is still a lack of understanding of how the abiotic environment imposes constraints on treeline patterns, the scales at which different effects are acting, and how these effects vary over large spatial extents. In this thesis, I examined abrupt Nothofagus treelines across seven degrees of latitude in New Zealand in order to investigate two broad questions: (1) What is the nature and extent of spatial variability in Nothofagus treelines across the country? (2) How is this variation associated with abiotic variation at different spatial scales? A range of GIS, statistical, and atmospheric modelling methods were applied to address these two questions. First, I characterised Nothofagus treeline patterns at a 15x15km scale across New Zealand using a set of seven, GIS-derived, quantitative metrics that describe different aspects of treeline position, shape, spatial configuration, and relationships with adjacent vegetation. Multivariate clustering of these metrics revealed distinct treeline types that showed strong spatial aggregation across the country. This suggests a strong spatial structuring of the abiotic environment which, in turn, drives treeline patterns. About half of the multivariate treeline metric variation was explained by patterns of climate, substrate, topographic and disturbance variability; on the whole, climatic and disturbance factors were most influential. Second, I developed a conceptual model that describes how treeline elevation may vary at different scales according to three categories of effects: thermal modifying effects, physiological stressors, and disturbance effects. I tested the relevance of this model for Nothofagus treelines by investigating treeline elevation variation at five nested scales (regional to local) using a hierarchical design based on nested river catchments. Hierarchical linear modelling revealed that the majority of the variation in treeline elevation resided at the broadest, regional scale, which was best explained by the thermal modifying effects of solar radiation, mountain mass, and differences in the potential for cold air ponding. Nonetheless, at finer scales, physiological and disturbance effects were important and acted to modify the regional trend at these scales. These results suggest that variation in abrupt treeline elevations are due to both broad-scale temperature-based growth limitation processes and finer-scale stress- and disturbance-related effects on seedling establishment. Third, I explored the applicability of a meso-scale atmospheric model, The Air Pollution Model (TAPM), for generating 200 m resolution, hourly topoclimatic data for temperature, incoming and outgoing radiation, relative humidity, and wind speeds. Initial assessments of TAPM outputs against data from two climate station locations over seven years showed that the model could generate predictions with a consistent level of accuracy for both sites, and which agreed with other evaluations in the literature. TAPM was then used to generate data at 28, 7x7 km Nothofagus treeline zones across New Zealand for January (summer) and July (winter) 2002. Using mixed-effects linear models, I determined that both site-level factors (mean growing season temperature, mountain mass, precipitation, earthquake intensity) and local-level landform (slope and convexity) and topoclimatic factors (solar radiation, photoinhibition index, frost index, desiccation index) were influential in explaining variation in treeline elevation within and among these sites. Treelines were generally closer to their site-level maxima in regions with higher mean growing season temperatures, larger mountains, and lower levels of precipitation. Within sites, higher treelines were associated with higher solar radiation, and lower photoinhibition and desiccation index values, in January, and lower desiccation index values in July. Higher treelines were also significantly associated with steeper, more convex landforms. Overall, this thesis shows that investigating treelines across extensive areas at multiple study scales enables the development of a more comprehensive understanding of treeline variability and underlying environmental constraints. These results can be used to formulate new hypotheses regarding the mechanisms driving treeline formation and to guide the optimal choice of field sites at which to test these hypotheses.

Research papers, University of Canterbury Library

This paper concerns the explicit consideration of near-fault directivity in conventional ground motion prediction models, and its implication for probabilistic seismic hazard analysis (PSHA) in New Zealand. The proposed approach utilises recently developed models by Shahi & Baker (2011), which account for both the 'narrowband' nature of the directivity pulse on spectral ordinates, and the probability of pulse occurrence at the site of interest. Furthermore, in order to correctly consider directivity, distributed seismicity sources are considered as finite-faults, as opposed to their (incorrect) conventional treatment as point-sources. The significance of directivity on hazard analysis results is illustrated for various vibration periods at generic sites located in Christchurch and Otira, two locations whose seismic hazard is comprised of notably different seismic sources. When compared to the PSHA results considering directivity and distributed seismicity as finite faults, it is shown that the NZS1170.5:2004 directivity factor is notably unconservative for all vibration periods in Otira (i.e. high seismic hazard region); and unconservative for Christchurch at short-to-moderate vibration periods ( < 3s); but conservative at long periods ( > 4s).

Research papers, University of Canterbury Library

This paper provides an overview of the salient aspects of the dense array of ground motions observed in the 4 September 2010 Darfield and 22 February 2011 Christchurch earthquakes. Particular attention is given to inferred physical reasons for the observed ground motions, which include: (i) source features such as forward directivity effects; (ii) The effects of the Canterbury Plains sedimentary basin on basin-generated surface waves, and waveguide effects through the region; and (iii) the importance of local site response as evidenced by observations of large long period amplification and liquefaction. The significance of vertical ground motion intensity is also examined.

Research papers, University of Canterbury Library

A major hazard accompanying earthquake shaking in areas of steep topography is the detachment of rocks from bedrock outcrops that subsequently slide, roll, or bounce downslope (i.e. rockfalls). The 2010-2011 Canterbury earthquake sequence caused recurrent and severe rockfall in parts of southern Christchurch. Coseismic rockfall caused five fatalities and significant infrastructural damage during the 2011 Mw 6.2 Christchurch earthquake. Here we examine a rockfall site in southern Christchurch in detail using geomorphic mapping, lidar analysis, geochronology (cosmogenic 3He dating, radiocarbon dating, optically stimulated luminescence (OSL) from quartz, infrared stimulated luminescence from K-feldspar), numerical modeling of rockfall boulder trajectories, and ground motion prediction equations (GMPEs). Rocks fell from the source cliff only in earthquakes with interpolated peak ground velocities exceeding ~10 cm/s; hundreds of smaller earthquakes did not produce rockfall. On the basis of empirical observations, GMPEs and age chronologies we attribute paleo-rockfalls to strong shaking in prehistoric earthquakes. We conclude that earthquake shaking of comparable intensity to the strongest contemporary earthquakes in Christchurch last occurred at this site approximately 5000 to 7000 years ago, and that in some settings, rockfall deposits provide useful proxies for past strong ground motions.

Research papers, University of Canterbury Library

This dissertation addresses several fundamental and applied aspects of ground motion selection for seismic response analyses. In particular, the following topics are addressed: the theory and application of ground motion selection for scenario earthquake ruptures; the consideration of causal parameter bounds in ground motion selection; ground motion selection in the near-fault region where directivity effect is significant; and methodologies for epistemic uncertainty consideration and propagation in the context of ground motion selection and seismic performance assessment. The paragraphs below outline each contribution in more detail. A scenario-based ground motion selection method is presented which considers the joint distribution of multiple intensity measure (IM) types based on the generalised conditional intensity measure (GCIM) methodology (Bradley, 2010b, 2012c). The ground motion selection algorithm is based on generating realisations of the considered IM distributions for a specific rupture scenario and then finding the prospective ground motions which best fit the realisations using an optimal amplitude scaling factor. In addition, using different rupture scenarios and site conditions, two important aspects of the GCIM methodology are scrutinised: (i) different weight vectors for the various IMs considered; and (ii) quantifying the importance of replicate selections for ensembles with different numbers of desired ground motions. As an application of the developed scenario-based ground motion selection method, ground motion ensembles are selected to represent several major earthquake scenarios in New Zealand that pose a significant seismic hazard, namely, Alpine, Hope and Porters Pass ruptures for Christchurch city; and Wellington, Ohariu, and Wairarapa ruptures for Wellington city. A rigorous basis is developed, and sensitivity analyses performed, for the consideration of bounds on causal parameters (e.g., magnitude, source-to-site distance, and site condition) for ground motion selection. The effect of causal parameter bound selection on both the number of available prospective ground motions from an initial empirical as-recorded database, and the statistical properties of IMs of selected ground motions are examined. It is also demonstrated that using causal parameter bounds is not a reliable approach to implicitly account for ground motion duration and cumulative effects when selection is based on only spectral acceleration (SA) ordinates. Specific causal parameter bounding criteria are recommended for general use as a ‘default’ bounding criterion with possible adjustments from the analyst based on problem-specific preferences. An approach is presented to consider the forward directivity effects in seismic hazard analysis, which does not separate the hazard calculations for pulse-like and non-pulse-like ground motions. Also, the ability of ground motion selection methods to appropriately select records containing forward directivity pulse motions in the near-fault region is examined. Particular attention is given to ground motion selection which is explicitly based on ground motion IMs, including SA, duration, and cumulative measures; rather than a focus on implicit parameters (i.e., distance, and pulse or non-pulse classifications) that are conventionally used to heuristically distinguish between the near-fault and far-field records. No ad hoc criteria, in terms of the number of directivity ground motions and their pulse periods, are enforced for selecting pulse-like records. Example applications are presented with different rupture characteristics, source-to-site geometry, and site conditions. It is advocated that the selection of ground motions in the near-fault region based on IM properties alone is preferred to that in which the proportion of pulse-like motions and their pulse periods are specified a priori as strict criteria for ground motion selection. Three methods are presented to propagate the effect of seismic hazard and ground motion selection epistemic uncertainties to seismic performance metrics. These methods differ in their level of rigor considered to propagate the epistemic uncertainty in the conditional distribution of IMs utilised in ground motion selection, selected ground motion ensembles, and the number of nonlinear response history analyses performed to obtain the distribution of engineering demand parameters. These methods are compared for an example site where it is observed that, for seismic demand levels below the collapse limit, epistemic uncertainty in ground motion selection is a smaller uncertainty contributor relative to the uncertainty in the seismic hazard itself. In contrast, uncertainty in ground motion selection process increases the uncertainty in the seismic demand hazard for near-collapse demand levels.

Research papers, Lincoln University

This thesis is a theoretical exploration of ‘remembrance’ and its production in the interactions between people/s and the landscape. This exploration takes place in the broad context of post earthquake Christchurch with a focus on public spaces along the Ōtākaro – Avon river corridor. Memory is universal to human beings, yet memories are subjective and culturally organized and produced - the relationship between memory and place therefore operates at individual and collective levels. Design responses that facilitate opportunities to create new memories, and also acknowledge the remembered past of human – landscape relationships are critical for social cohesion and wellbeing. I draw on insights from a range of theoretical sources, including critical interpretive methodologies, to validate subjective individual and group responses to memory and place. Such approaches also allowed me, as the researcher, considerable freedom to apply memory theory through film to illustrate ways we can re-member ourselves to our landscapes. The Ōtākaro-Avon river provided the site through and in which film strategies for remembrance are explored. Foregrounding differences in Māori and settler cultural orientations to memory and landscape, has highlighted the need for landscape design to consider remembrance - those cognitive and unseen dimensions that intertwine people and place. I argue it is our task to make space for such diverse relationships, and to ensure these stories and memories, embodied in landscape can be read through generations. I do not prescribe methods or strategies; rather I have sought to encourage thinking and debate and to suggest approaches through which the possibilities for remembrance may be enhanced.

Research papers, University of Canterbury Library

Geologic phenomena produced by earthquake shaking, including rockfalls and liquefaction features, provide important information on the intensity and spatiotemporal distribution of earthquake ground motions. The study of rockfall and liquefaction features produced in contemporary well- instrumented earthquakes increases our knowledge of how natural and anthropogenic environments respond to earthquakes and improves our ability to deduce seismologic information from analogous pre-contemporary (paleo-) geologic features. The study of contemporary and paleo- rockfall and liquefaction features enables improved forecasting of environmental responses to future earthquakes. In this thesis I utilize a combination of field and imagery-based mapping, trenching, stratigraphy, and numerical dating techniques to understand the nature and timing of rockfalls (and hillslope sedimentation) and liquefaction in the eastern South Island of New Zealand, and to examine the influence that anthropogenic activity has had on the geologic expressions of earthquake phenomena. At Rapaki (Banks Peninsula, NZ), field and imagery-based mapping, statistical analysis and numerical modeling was conducted on rockfall boulders triggered by the fatal 2011 Christchurch earthquakes (n=285) and compared with newly identified prehistoric (Holocene and Pleistocene) boulders (n=1049) deposited on the same hillslope. A significant population of modern boulders (n=26) travelled farther downslope (>150 m) than their most-travelled prehistoric counterparts, causing extensive damage to residential dwellings at the foot of the hillslope. Replication of prehistoric boulder distributions using 3-dimensional rigid body numerical models requires the application of a drag-coefficient, attributed to moderate to dense slope vegetation, to account for their spatial distribution. Radiocarbon dating provides evidence for 17th to early 20th century deforestation at the study site during Polynesian and European colonization and after emplacement of prehistoric rockfalls. Anthropocene deforestation enabled modern rockfalls to exceed the limits of their prehistoric predecessors, highlighting a shift in the geologic expression of rockfalls due to anthropogenic activity. Optical and radiocarbon dating of loessic hillslope sediments in New Zealand’s South Island is used to constrain the timing of prehistoric rockfalls and associated seismic events, and quantify spatial and temporal patterns of hillslope sedimentation including responses to seismic and anthropogenic forcing. Luminescence ages from loessic sediments constrain timing of boulder emplacement to between ~3.0 and ~12.5 ka, well before the arrival of Polynesians (ca AD 1280) and Europeans (ca AD 1800) in New Zealand, and suggest loess accumulation was continuing at the study site until 12-13 ka. Large (>5 m3) prehistoric rockfall boulders preserve an important record of Holocene hillslope sedimentation by creating local traps for sediment aggradation and upbuilding soil formation. Sediment accumulation rates increased considerably (>~10 factor increase) following human arrival and associated anthropogenic burning of hillslope vegetation. New numerical ages are presented to place the evolution of loess-mantled hillslopes in New Zealand’s South Island into a longer temporal framework and highlight the roles of earthquakes and humans on hillslope surface process. Extensive field mapping and characterization for 1733 individual prehistoric rockfall boulders was conducted at Rapaki and another Banks Peninsula site, Purau, to understand their origin, frequency, and spatial and volumetric distributions. Boulder characteristics and distributions were compared to 421 boulders deposited at the same sites during the 2010-2011 Canterbury earthquake sequence. Prehistoric boulders at Rapaki and Purau are comprised of two dominant lithofacies types: volcanic breccia and massive (coherent) lava basalt. Volcanic breccia boulders are found in greatest abundance (64-73% of total mapped rockfall) and volume (~90-96% of total rockfall) at both locations and exclusively comprise the largest boulders with the longest runout distances that pose the greatest hazard to life and property. This study highlights the primary influence that volcanic lithofacies architecture has on rockfall hazard. The influence of anthropogenic modifications on the surface and subsurface geologic expression of contemporary liquefaction created during the 2010-2011 Canterbury earthquake sequence (CES) in eastern Christchurch is examined. Trench observations indicate that anthropogenic fill layer boundaries and the composition/texture of discretely placed fill layers play an important role in absorbing fluidized sand/silt and controlling the subsurface architecture of preserved liquefaction features. Surface liquefaction morphologies (i.e. sand blows and linear sand blow arrays) display alignment with existing utility lines and utility excavations (and perforated pipes) provided conduits for liquefaction ejecta during the CES. No evidence of pre-CES liquefaction was identified within the anthropogenic fill layers or underlying native sediment. Radiocarbon dating of charcoal within the youngest native sediment suggests liquefaction has not occurred at the study site for at least the past 750-800 years. The importance of systematically examining the impact of buried infrastructure on channelizing and influencing surface and subsurface liquefaction morphologies is demonstrated. This thesis highlights the importance of using a multi-technique approach for understanding prehistoric and contemporary earthquake phenomena and emphasizes the critical role that humans play in shaping the geologic record and Earth’s surface processes.

Research papers, University of Canterbury Library

Over 900 buildings in the Christchurch central business district and 10,000 residential homes were demolished following the 22nd of February 2011 Canterbury earthquake, significantly disrupting the rebuild progress. This study looks to quantify the time required for demolitions during this event which will be useful for future earthquake recovery planning. This was done using the Canterbury Earthquake Recovery Authority (CERA) demolition database, which allowed an in-depth look into the duration of each phase of the demolition process. The effect of building location, building height, and the stakeholder which initiated the demolition process (i.e. building owner or CERA) was investigated. The demolition process comprises of five phases; (i) decision making, (ii) procurement and planning, (iii) demolition, (iv) site clean-up, and (v) completion certification. It was found that the time required to decide to demolish the building made up majority of the total demolition duration. Demolition projects initiated by CERA had longer procurement and planning durations, but was quicker in other phases. Demolished buildings in the suburbs had a longer decision making duration, but had little effect on other phases of the demolition process. The decision making and procurement and planning phases of the demolition process were shorter for taller buildings, though the other phases took longer. Fragility functions for the duration of each phase in the demolition process are provided for the various categories of buildings for use in future studies.

Research papers, University of Canterbury Library

The purpose of this thesis is to conduct a detailed examination of the forward-directivity characteristics of near-fault ground motions produced in the 2010-11 Canterbury earthquakes, including evaluating the efficacy of several existing empirical models which form the basis of frameworks for considering directivity in seismic hazard assessment. A wavelet-based pulse classification algorithm developed by Baker (2007) is firstly used to identify and characterise ground motions which demonstrate evidence of forward-directivity effects from significant events in the Canterbury earthquake sequence. The algorithm fails to classify a large number of ground motions which clearly exhibit an early-arriving directivity pulse due to: (i) incorrect pulse extraction resulting from the presence of pulse-like features caused by other physical phenomena; and (ii) inadequacy of the pulse indicator score used to carry out binary pulse-like/non-pulse-like classification. An alternative ‘manual’ approach is proposed to ensure 'correct' pulse extraction and the classification process is also guided by examination of the horizontal velocity trajectory plots and source-to-site geometry. Based on the above analysis, 59 pulse-like ground motions are identified from the Canterbury earthquakes , which in the author's opinion, are caused by forward-directivity effects. The pulses are also characterised in terms of their period and amplitude. A revised version of the B07 algorithm developed by Shahi (2013) is also subsequently utilised but without observing any notable improvement in the pulse classification results. A series of three chapters are dedicated to assess the predictive capabilities of empirical models to predict the: (i) probability of pulse occurrence; (ii) response spectrum amplification caused by the directivity pulse; (iii) period and amplitude (peak ground velocity, PGV) of the directivity pulse using observations from four significant events in the Canterbury earthquakes. Based on the results of logistic regression analysis, it is found that the pulse probability model of Shahi (2013) provides the most improved predictions in comparison to its predecessors. Pulse probability contour maps are developed to scrutinise observations of pulses/non-pulses with predicted probabilities. A direct comparison of the observed and predicted directivity amplification of acceleration response spectra reveals the inadequacy of broadband directivity models, which form the basis of the near-fault factor in the New Zealand loadings standard, NZS1170.5:2004. In contrast, a recently developed narrowband model by Shahi & Baker (2011) provides significantly improved predictions by amplifying the response spectra within a small range of periods. The significant positive bias demonstrated by the residuals associated with all models at longer vibration periods (in the Mw7.1 Darfield and Mw6.2 Christchurch earthquakes) is likely due to the influence of basin-induced surface waves and non-linear soil response. Empirical models for the pulse period notably under-predict observations from the Darfield and Christchurch earthquakes, inferred as being a result of both the effect of nonlinear site response and influence of the Canterbury basin. In contrast, observed pulse periods from the smaller magnitude June (Mw6.0) and December (Mw5.9) 2011 earthquakes are in good agreement with predictions. Models for the pulse amplitude generally provide accurate estimates of the observations at source-to-site distances between 1 km and 10 km. At longer distances, observed PGVs are significantly under-predicted due to their slower apparent attenuation. Mixed-effects regression is employed to develop revised models for both parameters using the latest NGA-West2 pulse-like ground motion database. A pulse period relationship which accounts for the effect of faulting mechanism using rake angle as a continuous predictor variable is developed. The use of a larger database in model development, however does not result in improved predictions of pulse period for the Darfield and Christchurch earthquakes. In contrast, the revised model for PGV provides a more appropriate attenuation of the pulse amplitude with distance, and does not exhibit the bias associated with previous models. Finally, the effects of near-fault directivity are explicitly included in NZ-specific probabilistic seismic hazard analysis (PSHA) using the narrowband directivity model of Shahi & Baker (2011). Seismic hazard analyses are conducted with and without considering directivity for typical sites in Christchurch and Otira. The inadequacy of the near-fault factor in the NZS1170.5: 2004 is apparent based on a comparison with the directivity amplification obtained from PSHA.

Research papers, University of Canterbury Library

This paper provides a summary of the ground motions observed in the recent Canterbury, New Zealand earthquake sequence. The sequence occurred in a region of relatively moderate seismicity, 130km to the east of the Alpine Fault, the major plate-boundary in the region. From an engineering perspective, the sequence has been primarily comprised of the initial 04/09/2010 Darfield earthquake (Mw7.1) followed by the 22/02/2011 Christchurch earthquake (Mw6.3), and two aftershocks on 13/06/ 2011 (Mw5.3 and 6.0, respectively). The dense spacing of strong motions in the region, and their close proximity to the respective causative faults, has resulted in strong ground motions far exceeding the previous catalogue of strong motion observed in New Zealand. The observed ground motions have exhibited clear evidence of: (i) near-source directivity; (ii) sedimentary basin focusing, amplification and basin effect refraction; (iii) non-linear site response; (iv) cyclic mobility postliquefaction; and (v) extreme vertical ground motions exceeding 2g, among others.

Research papers, University of Canterbury Library

Generalized conditional intensity measure (GCIM) method is extended to ground motion selection for scenario ruptures. Using different rupture scenarios and site conditions, various aspects of the GCIM methodology are scrutinized, including: (i) implementation of different weight vectors and the composition of the IM vector; (ii) quantifying the importance of replicate selections for different number of desired ground motions; and (iii) the effect of considering bounds on the implicit causal parameters of the prospective ground motions. Using the extended methodology, representative ground motion ensembles for several major earthquake scenarios in New Zealand are developed. Cases considered include representative ground motions for the occurrence of Alpine, Hope, and Porters Pass earthquakes in Christchurch city, and the occurrence of Wellington, Wairarapa, and Ohariu fault ruptures in Wellington city. Challenges in the development of ground motion ensembles for subduction zone earthquakes are also highlighted. The selected scenario-based ground motion sets can be used to complement ground motions which are often selected in conjunction with probabilistic seismic hazard analysis, in order to understand the performance of structures for the question “what if this fault ruptures?”

Research papers, University of Canterbury Library

Generalized conditional intensity measure (GCIM) method is extended to ground motion selection for scenario ruptures. Using different rupture scenarios and site conditions, various aspects of the GCIM methodology are scrutinized, including: (i) implementation of different weight vectors and the composition of the IM vector; (ii) quantifying the importance of replicate selections for different number of desired ground motions; and (iii) the effect of considering bounds on the implicit causal parameters of the prospective ground motions. Using the extended methodology, representative ground motion ensembles for several major earthquake scenarios in New Zealand are developed. Cases considered include representative ground motions for the occurrence of Alpine, Hope, and Porters Pass earthquakes in Christchurch city, and the occurrence of Wellington, Wairarapa, and Ohariu fault ruptures in Wellington city. Challenges in the development of ground motion ensembles for subduction zone earthquakes are also highlighted. The selected scenario-based ground motion sets can be used to complement ground motions which are often selected in conjunction with probabilistic seismic hazard analysis, in order to understand the performance of structures for the question “what if this fault ruptures?”

Research papers, Lincoln University

Global biodiversity is threatened by human actions, including in urban areas. Urbanisation has removed and fragmented indigenous habitats. As one of the 25 biodiversity ’hot spots’, New Zealand is facing the problems of habitat loss and indigenous species extinction. In New Zealand cities, as a result of the land clearance and imported urban planning precepts, many urban areas have little or no original native forest remaining. Urbanisation has also been associated with the introduction of multitudes of species from around the world. Two large earthquakes shook Christchurch in 2010 and 2011 and caused a lot of damage. Parts of the city suffered from soil liquefaction after the earthquakes. In the most damaged parts of Christchurch, particularly in the east, whole neighbourhoods were abandoned and later demolished except for larger trees. Christchurch offers an excellent opportunity to study the biodiversity responses to an urban area with less intensive management, and to learn more about the conditions in urban environments that are most conducive to indigenous plant biodiversity. This study focuses on natural woody plant regeneration of forested sites in Christchurch city, many of which were also surveyed prior to the earthquakes. By repeating the pre-earthquake surveys, I am able to describe the natural regeneration occurring in Christchurch forested areas. By combining this with the regeneration that has occurred in the Residential Red Zone, successional trajectories can be described under a range of management scenarios. Using a comprehensive tree map of the Residential Red Zone, I was also able to document minimum dispersal distances of a range of indigenous trees in Christchurch. This is important for planning reserve connectivity. Moreover, I expand and improve on a previous analysis of the habitat connectivity of Christchurch (made before the earthquakes) to incorporate the Residential Red Zone, to assess the importance for habitat connectivity of restoring the indigenous forest in this area. In combination, these data sets are used to provide patch scenarios and some management options for biodiversity restoration in the Ōtākaro-Avon Red Zone post-earthquake.

Research papers, University of Canterbury Library

Recent surface-rupturing earthquakes in New Zealand have highlighted significant exposure and vulnerability of the road network to fault displacement. Understanding fault displacement hazard and its impact on roads is crucial for mitigating risks and enhancing resilience. There is a need for regional-scale assessments of fault displacement to identify vulnerable areas within the road network for the purposes of planning and prioritising site-specific investigations. This thesis employs updated analysis of data from three historical surface-rupturing earthquakes (Edgecumbe 1987, Darfield 2010, and Kaikoūra 2016) to develop an empirical model that addresses the gap in regional fault displacement hazard analysis. The findings contribute to understanding of • How to use seismic hazard model inputs for regional fault displacement hazard analysis • How faulting type and sediment cover affects the magnitude and spatial distribution of fault displacement • How the distribution of displacement and regional fault displacement hazard is impacted by secondary faulting • The inherent uncertainties and limitations associated with employing an empirical approach at a regional scale • Which sections of New Zealand’s roading network are most susceptible to fault displacement hazard and warrant site-specific investigations • Which regions should prioritise updating emergency management plans to account for post-event disruptions to roading. I used displacement data from the aforementioned historical ruptures to generate displacement versus distance-to-fault curves for various displacement components, fault types, and geological characteristics. Using those relationships and established relationships for along-strike displacement, displacement contours were generated surrounding active faults within the NZ Community Fault Model. Next, I calculated a new measure of 1D strain along roads as well as relative hazard, which integrated 1D strain and normalised slip rate data. Summing these values at the regional level identified areas of heightened relative hazard across New Zealand, and permits an assessment of the susceptibility of road networks using geomorphon land classes as proxies for vulnerability. The results reveal that fault-parallel displacements tend to localise near the fault plane, while vertical and fault-perpendicular displacements sustain over extended distances. Notably, no significant disparities were observed in off-fault displacement between the hanging wall and footwall sides of the fault, or among different surface geology types, potentially attributed to dataset biases. The presence of secondary faulting in the dataset contributes to increased levels of tectonic displacement farther from the fault, highlighting its significance in hazard assessments. Furthermore, fault displacement contours delineate broader zones around dip-slip faults compared to strike-slip faults, with correlations identified between fault length and displacement width. Road ‘strain’ values are higher around dip-slip faults, with notable examples observed in the Westland and Buller Districts. As expected, relative hazard analysis revealed elevated values along faults with high slip rates, notably along the Alpine Fault. A regional-scale analysis of hazard and exposure reveals heightened relative hazard in specific regions, including Wellington, Southern Hawke’s Bay, Central Bay of Plenty, Central West Coast, inland Canterbury, and the Wairau Valley of Marlborough. Notably, the Central West Coast exhibits the highest summed relative hazard value, attributed to the fast-slipping Alpine Fault. The South Island generally experiences greater relative hazard due to larger and faster-slipping faults compared to the North Island, despite having fewer roads. Central regions of New Zealand face heightened risk compared to Southern or Northern regions. Critical road links intersecting high-slipping faults, such as State Highways 6, 73, 1, and 2, necessitate prioritisation for site-specific assessments, emergency management planning and targeted mitigation strategies. Roads intersecting with the Alpine Fault are prone to large parallel displacements, requiring post-quake repair efforts. Mitigation strategies include future road avoidance of nearby faults, modification of road fill and surface material, and acknowledgement of inherent risk, leading to prioritised repair efforts of critical roads post-quake. Implementing these strategies enhances emergency response efforts by improving accessibility to isolated regions following a major surface-rupturing event, facilitating faster supply delivery and evacuation assistance. This thesis contributes to the advancement of understanding fault displacement hazard by introducing a novel regional, empirical approach. The methods and findings highlight the importance of further developing such analyses and extending them to other critical infrastructure types exposed to fault displacement hazard in New Zealand. Enhancing our comprehension of the risks associated with fault displacement hazard offers valuable insights into various mitigation strategies for roading infrastructure and informs emergency response planning, thereby enhancing both national and global infrastructure resilience against geological hazards.

Research papers, Lincoln University

Mitigating the cascade of environmental damage caused by the movement of excess reactive nitrogen (N) from land to sea is currently limited by difficulties in precisely and accurately measuring N fluxes due to variable rates of attenuation (denitrification) during transport. This thesis develops the use of the natural abundance isotopic composition of nitrate (δ15N and δ18O of NO₃-) to integrate the spatialtemporal variability inherent to denitrification, creating an empirical framework for evaluating attenuation during land to water NO₃- transfers. This technique is based on the knowledge that denitrifiers kinetically discriminate against 'heavy' forms of both N and oxygen (O), creating a parallel enrichment in isotopes of both species as the reaction progresses. This discrimination can be quantitatively related to NO₃- attenuation by isotopic enrichment factors (εdenit). However, while these principles are understood, use of NO₃- isotopes to quantify denitrification fluxes in non-marine environments has been limited by, 1) poor understanding of εdenit variability, and, 2) difficulty in distinguishing the extent of mixing of isotopically distinct sources from the imprint of denitrification. Through a combination of critical literature analysis, mathematical modelling, mesocosm to field scale experiments, and empirical studies on two river systems over distance and time, these short comings are parametrised and a template for future NO₃- isotope based attenuation measurements outlined. Published εdenit values (n = 169) are collated in the literature analysis presented in Chapter 2. By evaluating these values in the context of known controllers on the denitrification process, it is found that the magnitude of εdenit, for both δ15N and δ18O, is controlled by, 1) biology, 2) mode of transport through the denitrifying zone (diffusion v. advection), and, 3) nitrification (spatial-temporal distance between nitrification and denitrification). Based on the outcomes of this synthesis, the impact of the three factors identified as controlling εdenit are quantified in the context of freshwater systems by combining simple mathematical modelling and lab incubation studies (comparison of natural variation in biological versus physical expression). Biologically-defined εdenit, measured in sediments collected from four sites along a temperate stream and from three tropical submerged paddy fields, varied from -3‰ to -28‰ depending on the site’s antecedent carbon content. Following diffusive transport to aerobic surface water, εdenit was found to become more homogeneous, but also lower, with the strength of the effect controlled primarily by diffusive distance and the rate of denitrification in the sediments. I conclude that, given the variability in fractionation dynamics at all levels, applying a range of εdenit from -2‰ to -10‰ provides more accurate measurements of attenuation than attempting to establish a site-specific value. Applying this understanding of denitrification's fractionation dynamics, four field studies were conducted to measure denitrification/ NO₃- attenuation across diverse terrestrial → freshwater systems. The development of NO₃- isotopic signatures (i.e., the impact of nitrification, biological N fixation, and ammonia volatilisation on the isotopic 'imprint' of denitrification) were evaluated within two key agricultural regions: New Zealand grazed pastures (Chapter 4) and Philippine lowland submerged rice production (Chapter 5). By measuring the isotopic composition of soil ammonium, NO₃- and volatilised ammonia following the bovine urine deposition, it was determined that the isotopic composition of NO₃ - leached from grazed pastures is defined by the balance between nitrification and denitrification, not ammonia volatilisation. Consequently, NO₃- created within pasture systems was predicted to range from +10‰ (δ15N)and -0.9‰ (δ18O) for non-fertilised fields (N limited) to -3‰ (δ15N) and +2‰ (δ18O) for grazed fertilised fields (N saturated). Denitrification was also the dominant determinant of NO₃- signatures in the Philippine rice paddy. Using a site-specific εdenit for the paddy, N inputs versus attenuation were able to be calculated, revealing that >50% of available N in the top 10 cm of soil was denitrified during land preparation, and >80% of available N by two weeks post-transplanting. Intriguingly, this denitrification was driven by rapid NO₃- production via nitrification of newly mineralised N during land preparation activities. Building on the relevant range of εdenit established in Chapters 2 and 3, as well as the soil-zone confirmation that denitrification was the primary determinant of NO₃- isotopic composition, two long-term longitudinal river studies were conducted to assess attenuation during transport. In Chapter 6, impact and recovery dynamics in an urban stream were assessed over six months along a longitudinal impact gradient using measurements of NO₃- dual isotopes, biological populations, and stream chemistry. Within 10 days of the catastrophic Christchurch earthquake, dissolved oxygen in the lowest reaches was <1 mg l⁻¹, in-stream denitrification accelerated (attenuating 40-80% of sewage N), microbial biofilm communities changed, and several benthic invertebrate taxa disappeared. To test the strength of this method for tackling the diffuse, chronic N loading of streams in agricultural regions, two years of longitudinal measurements of NO₃- isotopes were collected. Attenuation was negatively correlated with NO₃- concentration, and was highly dependent on rainfall: 93% of calculated attenuation (20 kg NO₃--N ha⁻¹ y⁻¹) occurred within 48 h of rainfall. The results of these studies demonstrate the power of intense measurements of NO₃- stable isotope for distinguishing temporal and spatial trends in NO₃ - loss pathways, and potentially allow for improved catchment-scale management of agricultural intensification. Overall this work now provides a more cohesive understanding for expanding the use of NO₃- isotopes measurements to generate accurate understandings of the controls on N losses. This information is becoming increasingly important to predict ecosystem response to future changes, such the increasing agricultural intensity needed to meet global food demand, which is occurring synergistically with unpredictable global climate change.

Research papers, University of Canterbury Library

Study region: Christchurch, New Zealand. Study focus: Low-lying coastal cities worldwide are vulnerable to shallow groundwater salinization caused by saltwater intrusion and anthropogenic activities. Shallow groundwater salinization can have cascading negative impacts on municipal assets, but this is rarely considered compared to impacts of salinization on water supply. Here, shallow groundwater salinity was sampled at high spatial resolution (1.3 piezometer/km2 ), then mapped and spatially interpolated. This was possible due to a uniquely extensive set of shallow piezometers installed in response to the 2010–11 Canterbury Earthquake Sequence to assess liquefaction risk. The municipal assets located within the brackish groundwater areas were highlighted. New hydrological insights for the region: Brackish groundwater areas were centred on a spit of coastal sand dunes and inside the meander of a tidal river with poorly drained soils. The municipal assets located within these areas include: (i) wastewater and stormwater pipes constructed from steel-reinforced concrete, which, if damaged, are vulnerable to premature failure when exposed to chloride underwater, and (ii) 41 parks and reserves totalling 236 ha, within which salt-intolerant groundwater-dependent species are at risk. This research highlights the importance of determining areas of saline shallow groundwater in low-lying coastal urban settings and the co-located municipal assets to allow the prioritisation of sites for future monitoring and management.

Research papers, University of Canterbury Library

This thesis examines the closing of Aranui High School in 2016, a low socio-economic secondary school in eastern Christchurch, New Zealand, and reflects on its history through the major themes of innovation and the impact of central government intervention. The history is explored through the leadership of the school principals, and the necessity for constant adaptation by staff to new ways of teaching and learning, driven by the need to accommodate a more varied student population – academically, behaviourally and culturally – than most other schools in wider Christchurch. Several extreme changes, following a neoliberal approach to education policies at a national government level, impacted severely on the school’s ability to thrive and even survive over the 57 years of its existence, with the final impact of the 2010 and 2011 Canterbury earthquakes leading indirectly to Aranui High’s closure. The earthquakes provided the National government with the impetus to advocate for change to education in Christchurch; changes which impacted negatively on many schools in Christchurch, including Aranui High School. The announcement of the closure of Aranui High shocked many staff and students, who were devastated that the school would no longer exist. Aranui High School, Aranui Primary School, Wainoni Primary School and Avondale Primary School were all closed to make way for Haeata Community Campus, a year 1 to 13 school, which was built on the Aranui High site. Aranui High School served the communities of eastern Christchurch for 57 years from 1960 and deserves acknowledgment and remembrance, and my hope is that this thesis will provide a fair representation of the school’s story, including its successes and challenges, while also explaining the reasons behind the eventual closure. This thesis contributes to New Zealand public history and uses mixed research methods to examine Aranui High School’s role as a secondary school in eastern Christchurch. I argue that the closure of Aranui High School in 2016 was an unjustified act by the Ministry of Education.

Research papers, Lincoln University

Today there is interest in building resilient communities. Identifying and managing the risks of natural hazards with communities who face compounding hazards is challenging. Alpine ski areas provide a unique context to study this challenging and complex process. The traditional approach taken to manage natural hazards is discipline-centric and focuses on common (e.g. high probability low consequence) natural hazards such as avalanches. While this thesis acknowledges that the common approach is rational, it argues that we can extend our communities of practice to include rare (e.g. low probability / high consequence) natural hazards such as earthquakes. The dynamically complex nature of these ‘rare’ hazards limits our understanding about them, but by seeking and using the lived experiences of people in mountain communities some knowledge can be gained to help improve our understanding of how to adapt. This study focuses on such an approach in the context of alpine ski areas prone to earthquakes as a first step toward identifying key policy opportunities for hazard mitigation in general. The contributions can be broken down into methodological, contextual, and theoretical pursuits, as well as opportunities for improving future research. A development mixed method triangulated approach was justified because the research problem (i.e. earthquakes in ski areas) has had little consideration. The context provided the opportunity to test the integration of methods while dealing with the challenges of research in a novel context. Advancement to fuzzy cognitive mapping was achieved through the use of unsupervised neural networks (Self-organizing Maps or Kohonen Maps). The framework applied in the multi-site case study required a synthesis of current approaches, advances to methods and a functional use of cultural theory. Different approaches to participatory policy development were reviewed to develop a research protocol that was accessible. Cultural theory was selected as a foundation for the thesis because of its’ preference for plural rationalities from five ways of organizing. Moreover, the study undertook a shift away from the dichotomy of ‘methodological individualism’ and ‘methodological collectivism’ and instead chose the dividual (i.e. social solidarities that consist of culural biases, behavioral strategies and social relations) as a consistent unit of analysis despite three different methodologies including: field studies, qualitative interviews, and fuzzy cognitive maps. In this sense, the thesis sought to move away from ‘elegant solutions’ from singular solidarities or methods toward a research philosophy that sustains requisite variety and clumsy solutions. Overall the approach was a trandisciplinary framework that is a step toward sustainable hazards mitigation. The results indicate that the selections of risks and adaptation strategies associated with the in-situ hazards are driven by roles that managers, workers, and riders play in the context. Additionally, fuzzy cognitive maps were used as an extension of qualitative interviews and demonstrated the potential for power struggles that may arise between participant groups when considering strategies for preparation, response and recovery. Moreover, the results stress that prolonged engagement with stakeholders is necessary to improve the policy development process. Some comments are made on the compatibility condition of congruence between cultural biases, behavioural strategies, and social relations. As well, inclusion of the hermit/autonomous solidarities is stressed as a necessary component of future applications of cultural theory. The transdisciplinary mixed-method framework is an approach that can be transferred to many other vital areas of research where integration is desirable.

Research papers, University of Canterbury Library

Current research in geotechnical engineering at the University of Canterbury includes a number of laboratory testing programmes focussed on understanding the behaviour of natural soil deposits in Christchurch during the 2010-2011 Canterbury Earthquake Sequence. Many soils found in Christchurch are sands or silty sands with little to no plasticity, making them very difficult to sample using established methods. The gel-push sampling methodology, developed by Kiso-Jiban Consultants in Japan, was developed to address some of the deficiencies of existing sampling techniques and has been deployed on two projects in Christchurch. Gel push sampling is carried out with a range of samplers which are modified versions of existing technology, and the University of Canterbury has acquired three versions of the tools (GP-S, GP-Tr, GP-D). Soil samples are extracted from the bottom of a freshly drilled borehole and are captured within a liner barrel, close to 1m in length. A lubricating polymer gel coats the outside of the soil sample as it enters the liner barrel. The frictional rubbing which normally occurs on the sides of the soil samples using existing techniques is eliminated by the presence of the polymer gel. The operation of the gel-push samplers is significantly more complicated than conventional push-tube samplers, and in the initial trials a number of operational difficulties were encountered, requiring changes to the sampling procedures. Despite these issues, a number of high quality soil samples were obtained on both projects using the GP-S sampler to capture silty soil. Attempts were made to obtain clean sands using a different gel-push sampler (GP-TR) in the Red Zone. The laboratory testing of these sands indicated that they were being significantly disturbed during the sampling and/or transportation procedures. While it remains too early to draw definitive conclusions regarding the performance of the gel-push samplers, the methodology has provided some promising results. Further trialling of the tools are required to refine operating procedures understand the full range of soil conditions which can be successfully sampled using the tools. In parallel with the gel-push trials, a Dames and Moore fixed-piston sampler has been used by our research partners from Berkeley to obtain soil samples at a number of sites within Christchurch. This sampler features relatively short (50cm), thin-walled liner barrels which is advanced into the ground under the action of hydraulic pressure. By reducing the overall length of the soil being captured, the disturbance to the soil as it enters the liner barrel is significantly reduced. The Dames and Moore sampler is significantly easier to operate than the gel-push sampler, and past experience has shown it to be successful in soft, plastic materials (i.e. clays and silty clays). The cyclic resistance of one silty clay obtained using both the gel-push and Dames & Moore samplers has been found to be very similar, and ongoing research aims to establish whether similar results are obtained for different soil types, including silty materials and clean sands.

Research papers, University of Canterbury Library

In the last century, seismic design has undergone significant advancements. Starting from the initial concept of designing structures to perform elastically during an earthquake, the modern seismic design philosophy allows structures to respond to ground excitations in an inelastic manner, thereby allowing damage in earthquakes that are significantly less intense than the largest possible ground motion at the site of the structure. Current performance-based multi-objective seismic design methods aim to ensure life-safety in large and rare earthquakes, and to limit structural damage in frequent and moderate earthquakes. As a result, not many recently built buildings have collapsed and very few people have been killed in 21st century buildings even in large earthquakes. Nevertheless, the financial losses to the community arising from damage and downtime in these earthquakes have been unacceptably high (for example; reported to be in excess of 40 billion dollars in the recent Canterbury earthquakes). In the aftermath of the huge financial losses incurred in recent earthquakes, public has unabashedly shown their dissatisfaction over the seismic performance of the built infrastructure. As the current capacity design based seismic design approach relies on inelastic response (i.e. ductility) in pre-identified plastic hinges, it encourages structures to damage (and inadvertently to incur loss in the form of repair and downtime). It has now been widely accepted that while designing ductile structural systems according to the modern seismic design concept can largely ensure life-safety during earthquakes, this also causes buildings to undergo substantial damage (and significant financial loss) in moderate earthquakes. In a quest to match the seismic design objectives with public expectations, researchers are exploring how financial loss can be brought into the decision making process of seismic design. This has facilitated conceptual development of loss optimisation seismic design (LOSD), which involves estimating likely financial losses in design level earthquakes and comparing against acceptable levels of loss to make design decisions (Dhakal 2010a). Adoption of loss based approach in seismic design standards will be a big paradigm shift in earthquake engineering, but it is still a long term dream as the quantification of the interrelationships between earthquake intensity, engineering demand parameters, damage measures, and different forms of losses for different types of buildings (and more importantly the simplification of the interrelationship into design friendly forms) will require a long time. Dissecting the cost of modern buildings suggests that the structural components constitute only a minor portion of the total building cost (Taghavi and Miranda 2003). Moreover, recent research on seismic loss assessment has shown that the damage to non-structural elements and building contents contribute dominantly to the total building loss (Bradley et. al. 2009). In an earthquake, buildings can incur losses of three different forms (damage, downtime, and death/injury commonly referred as 3Ds); but all three forms of seismic loss can be expressed in terms of dollars. It is also obvious that the latter two loss forms (i.e. downtime and death/injury) are related to the extent of damage; which, in a building, will not just be constrained to the load bearing (i.e. structural) elements. As observed in recent earthquakes, even the secondary building components (such as ceilings, partitions, facades, windows parapets, chimneys, canopies) and contents can undergo substantial damage, which can lead to all three forms of loss (Dhakal 2010b). Hence, if financial losses are to be minimised during earthquakes, not only the structural systems, but also the non-structural elements (such as partitions, ceilings, glazing, windows etc.) should be designed for earthquake resistance, and valuable contents should be protected against damage during earthquakes. Several innovative building technologies have been (and are being) developed to reduce building damage during earthquakes (Buchanan et. al. 2011). Most of these developments are aimed at reducing damage to the buildings’ structural systems without due attention to their effects on non-structural systems and building contents. For example, the PRESSS system or Damage Avoidance Design concept aims to enable a building’s structural system to meet the required displacement demand by rocking without the structural elements having to deform inelastically; thereby avoiding damage to these elements. However, as this concept does not necessarily reduce the interstory drift or floor acceleration demands, the damage to non-structural elements and contents can still be high. Similarly, the concept of externally bracing/damping building frames reduces the drift demand (and consequently reduces the structural damage and drift sensitive non-structural damage). Nevertheless, the acceleration sensitive non-structural elements and contents will still be very vulnerable to damage as the floor accelerations are not reduced (arguably increased). Therefore, these concepts may not be able to substantially reduce the total financial losses in all types of buildings. Among the emerging building technologies, base isolation looks very promising as it seems to reduce both inter-storey drifts and floor accelerations, thereby reducing the damage to the structural/non-structural components of a building and its contents. Undoubtedly, a base isolated building will incur substantially reduced loss of all three forms (dollars, downtime, death/injury), even during severe earthquakes. However, base isolating a building or applying any other beneficial technology may incur additional initial costs. In order to provide incentives for builders/owners to adopt these loss-minimising technologies, real-estate and insurance industries will have to acknowledge the reduced risk posed by (and enhanced resilience of) such buildings in setting their rental/sale prices and insurance premiums.