123 resultados para Quantities

em Queensland University of Technology - ePrints Archive


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Automated Estimator and LCADesign are two early examples of nD modelling software which both rely on the extraction of quantities from CAD models to support their further processing. The issues of building information modelling (BIM), quantity takeoff for different purposes and automating quantity takeoff are discussed by comparing the aims and use of the two programs. The technical features of the two programs are also described. The technical issues around the use of 3D models is described together with implementation issues and comments about the implementation of the IFC specifications. Some user issues that emerged through the development process are described, with a summary of the generic research tasks which are necessary to fully support the use of BIM and nD modelling.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Over the past few years, the Midwest ISO has experienced a surge in requests to interconnect large amounts of wind generation, driven largely by a favorable political environment and an abundant wind resource in the Midwestern US. This tremendous influx of proposed generators along with a highly constrained transmission system adversely impacted interconnection queue processing, resulting in an unmanageable backlog. Under these circumstances, Midwest ISO successfully reformed the interconnection tariff to improve cycle times and provide increased certainty to interconnection customers. One of the key features of the reformed queue process is the System Planning and Analysis (SPA) phase which allows integration of the interconnection studies with regional transmission planning. This paper presents a brief background of the queue reform effort and then delves deeply in to the work performed at the Midwest ISO during the first SPA cycle - the study approach, the challenges faced in having to study over 50,000 MWs of wind generation and the effective solutions designed to complete these studies within tariff timelines.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

net sustainability. At best they reduce relative resource consumption. They still consume vast quantities of materials, energy, water and ecosystems during construction. Moreover, green buildings replace land and ecosystems with structures that, at the very best, only 'mimic' ecosystems<'). Mimicking nature is little compensation when we have lost a third of species that are integral parts of our life support system. Already, development has exceeded the Earth's ecological carrying capacity, so even 'restorative' design is not enough. Urban areas must be retrofitted to increase net bioregional carrying capacity - just to support existing or reduced population levels in cities. The eco-retrofitting of our built environment is therefore an essential precondition of achieving a sustainable society. But we need to eco-retrofit cities in ways that increase net sustainability, not just relative efficiency.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Genuine sustainability would require that urban development provide net positive social and ecological gains to compensate for previous lost natural capital and carrying capacity. Thus far, green buildings do not contribute to net sustainability. While they reduce relative resource consumption, they consume vast quantities of materials, energy and water.i Moreover, they replace land and ecosystems with structures that, at best, ‘mimic’ ecosystems. Elsewhere, the author has proposed a‘sustainability standard’, where development would leave the ecology, as well as society, better off after construction than before.ii To meet this standard, a development would need to add natural and social capital beyond what existed prior to development. Positive DesignTM or Positive DevelopmentTM is that which expands both the ecological base (life support system) and the public estate (equitable access to means of survival). How to achieve this is discussed in Positive Development (Birkeland 2008). This paper examines how net positive gains can be achieved in a ubtropical as well as temperate environment.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Bone graft is generally considered fundamental in achieving solid fusion in scoliosis correction and pseudarthrosis following instrumentation may predispose to implant failure. In endoscopic anterior-instrumented scoliosis surgery, autologous rib or iliac crest graft has been utilised traditionally but both techniques increase operative duration and cause donor site morbidity. Allograft bone and bone- morphogenetic-protein alternatives may improve fusion rates but this remains controversial. This study's objective was to compare two-year postoperative fusion rates in a series of patients who underwent endoscopic anterior instrumentation for thoracic scoliosis utilising various bone graft types. Significantly better rates of fusion occurred in endoscopic anterior instrumented scoliosis correction using femoral allograft compared to autologous rib-heads and iliac crest graft. This may be partly explained by the difficulty obtaining sufficient quantities of autologous graft. Lower fusion rates in the autologous graft group appeared to predispose to rod fracture although the clinical consequence of implant failure is uncertain.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Vibration based damage identification methods examine the changes in primary modal parameters or quantities derived from modal parameters. As one method may have advantages over the other under some circumstances, a multi-criteria approach is proposed. Case studies are conducted separately on beam, plate and plate-on-beam structures. Using the numerically simulated modal data obtained through finite element analysis software, algorithms based on flexibility and strain energy changes before and after damage are obtained and used as the indices for the assessment of the state of structural health. Results show that the proposed multi-criteria method is effective in damage identification in these structures.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The paper presents an interim summary of research and case studies being undertaken in the Sydney Opera House FM Exemplar Project covering procurement, benchmarking and building information models. The final outcomes of the FM Exemplar Project will be presented through various forums open to all FM practitioners and published in Australia and elsewhere through relevant journals. Sydney Opera House is an Australian icon, attracting some 4.5 million visitors per year who admire its built form and enjoy an evening of theatre. The building is the attraction, part of the experience. Therefore, facilities management is critical to the success of the Sydney Opera House enterprise and an ideal subject for the study of facilities management. Significantly the three research themes are heavily intertwined – effective risk sharing in procurement requires historic information and benchmarks for future performance, benchmarking gathers vast quantities of data that can only be exploited if properly related to one another and a building information model provides the means to manage such data. The case studies are emerging as real-life examples of how one organisation is addressing FM issues common to many, and will provide useful lessons for practitioners pursing similar strategies in their own organisations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The ability to assess a commercial building for its impact on the environment at the earliest stage of design is a goal which is achievable by integrating several approaches into a single procedure directly from the 3D CAD representation. Such an approach enables building design professionals to make informed decisions on the environmental impact of building and its alternatives during the design development stage instead of at the post-design stage where options become limited. The indicators of interest are those which relate to consumption of resources and energy, contributions to pollution of air, water and soil, and impacts on the health and wellbeing of people in the built environment as a result of constructing and operating buildings. 3D object-oriented CAD files contain a wealth of building information which can be interrogated for details required for analysis of the performance of a design. The quantities of all components in the building can be automatically obtained from the 3D CAD objects and their constituent materials identified to calculate a complete list of the amounts of all building products such as concrete, steel, timber, plastic etc. When this information is combined with a life cycle inventory database, key internationally recognised environmental indicators can be estimated. Such a fully integrated tool known as LCADesign has been created for automated ecoefficiency assessment of commercial buildings direct from 3D CAD. This paper outlines the key features of LCADesign and its application to environmental assessment of commercial buildings.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Buildings consume resources and energy, contribute to pollution of our air, water and soil, impact the health and well-being of populations and constitute an important part of the built environment in which we live. The ability to assess their design with a view to reducing that impact automatically from their 3D CAD representations enables building design professionals to make informed decisions on the environmental impact of building structures. Contemporary 3D object-oriented CAD files contain a wealth of building information. LCADesign has been designed as a fully integrated approach for automated eco-efficiency assessment of commercial buildings direct from 3D CAD. LCADesign accesses the 3D CAD detail through Industry Foundation Classes (IFCs) - the international standard file format for defining architectural and constructional CAD graphic data as 3D real-world objects - to permit construction professionals to interrogate these intelligent drawing objects for analysis of the performance of a design. The automated take-off provides quantities of all building components whose specific production processes, logistics and raw material inputs, where necessary, are identified to calculate a complete list of quantities for all products such as concrete, steel, timber, plastic etc and combines this information with the life cycle inventory database, to estimate key internationally recognised environmental indicators such as CML, EPS and Eco-indicator 99. This paper outlines the key modules of LCADesign and their role in delivering an automated eco-efficiency assessment for commercial buildings.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper presents the design of self-tuning controllers for a two terminal HVDC link. The controllers are designed utilizing a novel discrete-time converter model based on multirate sampling. The nature of converter firing system necessitates the development of a two-step ahead self-tuning control strategy. A two terminal HVDC system study has been carried out to show the effectiveness of the control strategies proposed which include the design of minimum variance controller, pole assigned controller and PLQG controller. The coordinated control of a two terminal HVDC system has been established deriving the signal from inverter end current and voltage which has been estimated based on the measurements of rectifier end quantities only realized through the robust reduced order observer. A well known scaled down sample system data has been selected for studies and the controllers designed have been tested for worst conditions. The performance of self-tuning controllers has been evaluated through digital simulation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

1. Species' distribution modelling relies on adequate data sets to build reliable statistical models with high predictive ability. However, the money spent collecting empirical data might be better spent on management. A less expensive source of species' distribution information is expert opinion. This study evaluates expert knowledge and its source. In particular, we determine whether models built on expert knowledge apply over multiple regions or only within the region where the knowledge was derived. 2. The case study focuses on the distribution of the brush-tailed rock-wallaby Petrogale penicillata in eastern Australia. We brought together from two biogeographically different regions substantial and well-designed field data and knowledge from nine experts. We used a novel elicitation tool within a geographical information system to systematically collect expert opinions. The tool utilized an indirect approach to elicitation, asking experts simpler questions about observable rather than abstract quantities, with measures in place to identify uncertainty and offer feedback. Bayesian analysis was used to combine field data and expert knowledge in each region to determine: (i) how expert opinion affected models based on field data and (ii) how similar expert-informed models were within regions and across regions. 3. The elicitation tool effectively captured the experts' opinions and their uncertainties. Experts were comfortable with the map-based elicitation approach used, especially with graphical feedback. Experts tended to predict lower values of species occurrence compared with field data. 4. Across experts, consensus on effect sizes occurred for several habitat variables. Expert opinion generally influenced predictions from field data. However, south-east Queensland and north-east New South Wales experts had different opinions on the influence of elevation and geology, with these differences attributable to geological differences between these regions. 5. Synthesis and applications. When formulated as priors in Bayesian analysis, expert opinion is useful for modifying or strengthening patterns exhibited by empirical data sets that are limited in size or scope. Nevertheless, the ability of an expert to extrapolate beyond their region of knowledge may be poor. Hence there is significant merit in obtaining information from local experts when compiling species' distribution models across several regions.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Insulin-like growth factor binding proteins (IGFBPs) are prime regulators of IGF-action in numerous cell types including the retinal pigment epithelium (RPE). The RPE performs several functions essential for vision, including growth factor secretion and waste removal via a phagocytic process mediated in part by vitronectin (Vn). In the course of studying the effects of IGFBPs on IGF-mediated VEGF secretion and Vn-mediated phagocytosis in the RPE cell line ARPE-19, we have discovered that these cells avidly ingest synthetic microspheres (2.0 μm diameter) coated with IGFBPs. Given the novelty of this finding and the established role for endocytosis in mediating IGFBP actions in other cell types, we have explored the potential role of candidate cell surface receptors. Moreover, we have examined the role of key IGFBP structural motifs, by comparing responses to three members of the IGFBP family (IGFBP-3, IGFBP-4 and IGFBP-5) which display overlapping variations in primary structure and glycosylation status. Coating of microspheres (FluoSpheres®, sulfate modified polystyrene filled with a fluorophore) was conducted at 37 °C for 1 h using 20 μg/mL of test protein, followed by extensive washing. Binding of proteins was confirmed using a microBCA assay. The negative control consisted of microspheres treated with 0.1% bovine serum albumin (BSA), and all test samples were post-treated with BSA in an effort to coat any remaining free protein binding sites, which might otherwise encourage non-specific interactions with the cell surface. Serum-starved cultures of ARPE-19 cells were incubated with microspheres for 24 h, using a ratio of approximately 100 microspheres per cell. Uptake of microspheres was quantified using a fluorometer and was confirmed visually by confocal fluorescence microscopy. The ARPE-19 cells displayed little affinity for BSA-treated microspheres, but avidly ingested large quantities of those pre-treated with Vn (ANOVA; p < 0.001). Strong responses were also observed towards recombinant formulations of non-glycosylated IGFBP-3, glycosylated IGFBP-3 and glycosylated IGFBP-5 (all p < 0.001), while glycosylated IGFBP-4 induced a relatively minor response (p < 0.05). The response to IGFBP-3 was unaffected in the presence of excess soluble IGFBP-3, IGF-I or Vn. Likewise, soluble IGFBP-3 did not induce uptake of BSA-treated microspheres. Antibodies to either the transferrin receptor or type 1 IGF-receptor displayed slight inhibitory effects on responses to IGFBPs and Vn. Heparin abolished responses to Vn, IGFBP-5 and non-glycosylated IGFBP-3, but only partially inhibited the response to glycosylated IGFBP-3. Our results demonstrate for the first time IGFBP-mediated endocytosis in ARPE-19 cells and suggest roles for the IGFBP-heparin-binding domain and glycosylation status. These findings have important implications for understanding the mechanisms of IGFBP actions on the RPE, and in particular suggest a role for IGFBP-endocytosis.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Computer forensics is the process of gathering and analysing evidence from computer systems to aid in the investigation of a crime. Typically, such investigations are undertaken by human forensic examiners using purpose-built software to discover evidence from a computer disk. This process is a manual one, and the time it takes for a forensic examiner to conduct such an investigation is proportional to the storage capacity of the computer's disk drives. The heterogeneity and complexity of various data formats stored on modern computer systems compounds the problems posed by the sheer volume of data. The decision to undertake a computer forensic examination of a computer system is a decision to commit significant quantities of a human examiner's time. Where there is no prior knowledge of the information contained on a computer system, this commitment of time and energy occurs with little idea of the potential benefit to the investigation. The key contribution of this research is the design and development of an automated process to describe a computer system and its activity for the purposes of a computer forensic investigation. The term proposed for this process is computer profiling. A model of a computer system and its activity has been developed over the course of this research. Using this model a computer system, which is the subj ect of investigation, can be automatically described in terms useful to a forensic investigator. The computer profiling process IS resilient to attempts to disguise malicious computer activity. This resilience is achieved by detecting inconsistencies in the information used to infer the apparent activity of the computer. The practicality of the computer profiling process has been demonstrated by a proof-of concept software implementation. The model and the prototype implementation utilising the model were tested with data from real computer systems. The resilience of the process to attempts to disguise malicious activity has also been demonstrated with practical experiments conducted with the same prototype software implementation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

An experimental investigation has been made of a round, non-buoyant plume of nitric oxide, NO, in a turbulent grid flow of ozone, 03, using the Turbulent Smog Chamber at the University of Sydney. The measurements have been made at a resolution not previously reported in the literature. The reaction is conducted at non-equilibrium so there is significant interaction between turbulent mixing and chemical reaction. The plume has been characterized by a set of constant initial reactant concentration measurements consisting of radial profiles at various axial locations. Whole plume behaviour can thus be characterized and parameters are selected for a second set of fixed physical location measurements where the effects of varying the initial reactant concentrations are investigated. Careful experiment design and specially developed chemilurninescent analysers, which measure fluctuating concentrations of reactive scalars, ensure that spatial and temporal resolutions are adequate to measure the quantities of interest. Conserved scalar theory is used to define a conserved scalar from the measured reactive scalars and to define frozen, equilibrium and reaction dominated cases for the reactive scalars. Reactive scalar means and the mean reaction rate are bounded by frozen and equilibrium limits but this is not always the case for the reactant variances and covariances. The plume reactant statistics are closer to the equilibrium limit than those for the ambient reactant. The covariance term in the mean reaction rate is found to be negative and significant for all measurements made. The Toor closure was found to overestimate the mean reaction rate by 15 to 65%. Gradient model turbulent diffusivities had significant scatter and were not observed to be affected by reaction. The ratio of turbulent diffusivities for the conserved scalar mean and that for the r.m.s. was found to be approximately 1. Estimates of the ratio of the dissipation timescales of around 2 were found downstream. Estimates of the correlation coefficient between the conserved scalar and its dissipation (parallel to the mean flow) were found to be between 0.25 and the significant value of 0.5. Scalar dissipations for non-reactive and reactive scalars were found to be significantly different. Conditional statistics are found to be a useful way of investigating the reactive behaviour of the plume, effectively decoupling the interaction of chemical reaction and turbulent mixing. It is found that conditional reactive scalar means lack significant transverse dependence as has previously been found theoretically by Klimenko (1995). It is also found that conditional variance around the conditional reactive scalar means is relatively small, simplifying the closure for the conditional reaction rate. These properties are important for the Conditional Moment Closure (CMC) model for turbulent reacting flows recently proposed by Klimenko (1990) and Bilger (1993). Preliminary CMC model calculations are carried out for this flow using a simple model for the conditional scalar dissipation. Model predictions and measured conditional reactive scalar means compare favorably. The reaction dominated limit is found to indicate the maximum reactedness of a reactive scalar and is a limiting case of the CMC model. Conventional (unconditional) reactive scalar means obtained from the preliminary CMC predictions using the conserved scalar p.d.f. compare favorably with those found from experiment except where measuring position is relatively far upstream of the stoichiometric distance. Recommendations include applying a full CMC model to the flow and investigations both of the less significant terms in the conditional mean species equation and the small variation of the conditional mean with radius. Forms for the p.d.f.s, in addition to those found from experiments, could be useful for extending the CMC model to reactive flows in the atmosphere.