909 resultados para Computational complexity.
Resumo:
The Guardian reportage of the United Kingdom Member of Parliament (MP) expenses scandal of 2009 used crowdsourcing and computational journalism techniques. Computational journalism can be broadly defined as the application of computer science techniques to the activities of journalism. Its foundation lies in computer assisted reporting techniques and its importance is increasing due to the: (a) increasing availability of large scale government datasets for scrutiny; (b) declining cost, increasing power and ease of use of data mining and filtering software; and Web 2.0; and (c) explosion of online public engagement and opinion.. This paper provides a case study of the Guardian MP expenses scandal reportage and reveals some key challenges and opportunities for digital journalism. It finds journalists may increasingly take an active role in understanding, interpreting, verifying and reporting clues or conclusions that arise from the interrogations of datasets (computational journalism). Secondly a distinction should be made between information reportage and computational journalism in the digital realm, just as a distinction might be made between citizen reporting and citizen journalism. Thirdly, an opportunity exists for online news providers to take a ‘curatorial’ role, selecting and making easily available the best data sources for readers to use (information reportage). These activities have always been fundamental to journalism, however the way in which they are undertaken may change. Findings from this paper may suggest opportunities and challenges for the implementation of computational journalism techniques in practice by digital Australian media providers, and further areas of research.
Resumo:
After a brief personal orientation, this presentation offers an opening section on „clash, cluster, complexity, cities‟ – making the case that innovation (both creative and economic) proceeds not only from incremental improvements within an expert-pipeline process, but also from the clash of different systems, generations, and cultures. The argument is that cultural complexity arises from such clashes, and that clustering is the solution to problems of complexity. The classic, 10,000-year-old, institutional form taken by such clusters is … cities. Hence, a creative city is one where clashing and competitive complexity is clustered… and, latterly, networked.
Resumo:
There are at least four key challenges in the online news environment that computational journalism may address. Firstly, news providers operate in a rapidly evolving environment and larger businesses are typically slower to adapt to market innovations. News consumption patterns have changed and news providers need to find new ways to capture and retain digital users. Meanwhile, declining financial performance has led to cost cuts in mass market newspapers. Finally investigative reporting is typically slow, high cost and may be tedious, and yet is valuable to the reputation of a news provider. Computational journalism involves the application of software and technologies to the activities of journalism, and it draws from the fields of computer science, social science and communications. New technologies may enhance the traditional aims of journalism, or may require “a new breed of people who are midway between technologists and journalists” (Irfan Essa in Mecklin 2009: 3). Historically referred to as ‘computer assisted reporting’, the use of software in online reportage is increasingly valuable due to three factors: larger datasets are becoming publicly available; software is becoming sophisticated and ubiquitous; and the developing Australian digital economy. This paper introduces key elements of computational journalism – it describes why it is needed; what it involves; benefits and challenges; and provides a case study and examples. Computational techniques can quickly provide a solid factual basis for original investigative journalism and may increase interaction with readers, when correctly used. It is a major opportunity to enhance the delivery of original investigative journalism, which ultimately may attract and retain readers online.
Resumo:
The health of tollbooth workers is seriously threatened by long-term exposure to polluted air from vehicle exhausts. Using traffic data collected at a toll plaza, vehicle movements were simulated by a system dynamics model with different traffic volumes and toll collection procedures. This allowed the average travel time of vehicles to be calculated. A three-dimension Computational Fluid Dynamics (CFD) model was used with a k–ε turbulence model to simulate pollutant dispersion at the toll plaza for different traffic volumes and toll collection procedures. It was shown that pollutant concentration around tollbooths increases as traffic volume increases. Whether traffic volume is low or high (1500 vehicles/h or 2500 vehicles/h), pollutant concentration decreases if electronic toll collection (ETC) is adopted. In addition, pollutant concentration around tollbooths decreases as the proportion of ETC-equipped vehicles increases. However, if the proportion of ETC-equipped vehicles is very low and the traffic volume is not heavy, then pollutant concentration increases as the number of ETC lanes increases.
Resumo:
We have recently demonstrated the geographic isolation of rice tungro bacilliform virus (RTBV) populations in the tungro-endemic provinces of Isabela and North Cotabato, Philippines. In this study, we examined the genetic structure of the virus populations at the tungro-outbreak sites of Lanao del Norte, a province adjacent to North Cotabato. We also analyzed the virus populations at the tungro-endemic sites of Subang, Indonesia, and Dien Khanh, Vietnam. Total DNA extracts from 274 isolates were digested with EcoRV restriction enzyme and hybridized with a full-length probe of RTBV. In the total population, 22 EcoRV-restricted genome profiles (genotypes) were identified. Although overlapping genotypes could be observed, the outbreak sites of Lanao del Norte had a genotype combination distinct from that of Subang or Dien Khanh but a genotype combination similar to that identified earlier from North Cotabato, the adjacent endemic province. Sequence analysis of the intergenic region and part of the ORF1 RTBV genome from randomly selected genotypes confirms the geographic clustering of RTBV genotypes and, combined with restriction analysis, the results suggest a fragmented spatial distribution of RTBV local populations in the three countries. Because RTBV depends on rice tungro spherical virus (RTSV) for transmission, the population dynamics of both tungro viruses were then examined at the endemic and outbreak sites within the Philippines. The RTBV genotypes and the coat protein RTSV genotypes were used as indicators for virus diversity. A shift in population structure of both viruses was observed at the outbreak sites with a reduced RTBV but increased RTSV gene diversity
Resumo:
Recently, the numerical modelling and simulation for fractional partial differential equations (FPDE), which have been found with widely applications in modern engineering and sciences, are attracting increased attentions. The current dominant numerical method for modelling of FPDE is the explicit Finite Difference Method (FDM), which is based on a pre-defined grid leading to inherited issues or shortcomings. This paper aims to develop an implicit meshless approach based on the radial basis functions (RBF) for numerical simulation of time fractional diffusion equations. The discrete system of equations is obtained by using the RBF meshless shape functions and the strong-forms. The stability and convergence of this meshless approach are then discussed and theoretically proven. Several numerical examples with different problem domains are used to validate and investigate accuracy and efficiency of the newly developed meshless formulation. The results obtained by the meshless formations are also compared with those obtained by FDM in terms of their accuracy and efficiency. It is concluded that the present meshless formulation is very effective for the modelling and simulation for FPDE.
Resumo:
The success rate of carrier phase ambiguity resolution (AR) is the probability that the ambiguities are successfully fixed to their correct integer values. In existing works, an exact success rate formula for integer bootstrapping estimator has been used as a sharp lower bound for the integer least squares (ILS) success rate. Rigorous computation of success rate for the more general ILS solutions has been considered difficult, because of complexity of the ILS ambiguity pull-in region and computational load of the integration of the multivariate probability density function. Contributions of this work are twofold. First, the pull-in region mathematically expressed as the vertices of a polyhedron is represented by a multi-dimensional grid, at which the cumulative probability can be integrated with the multivariate normal cumulative density function (mvncdf) available in Matlab. The bivariate case is studied where the pull-region is usually defined as a hexagon and the probability is easily obtained using mvncdf at all the grid points within the convex polygon. Second, the paper compares the computed integer rounding and integer bootstrapping success rates, lower and upper bounds of the ILS success rates to the actual ILS AR success rates obtained from a 24 h GPS data set for a 21 km baseline. The results demonstrate that the upper bound probability of the ILS AR probability given in the existing literatures agrees with the actual ILS success rate well, although the success rate computed with integer bootstrapping method is a quite sharp approximation to the actual ILS success rate. The results also show that variations or uncertainty of the unit–weight variance estimates from epoch to epoch will affect the computed success rates from different methods significantly, thus deserving more attentions in order to obtain useful success probability predictions.
Resumo:
While Business Process Management (BPM) is an established discipline, the increased adoption of BPM technology in recent years has introduced new challenges. One challenge concerns dealing with the ever-growing complexity of business process models. Mechanisms for dealing with this complexity can be classified into two categories: 1) those that are solely concerned with the visual representation of the model and 2) those that change its inner structure. While significant attention is paid to the latter category in the BPM literature, this paper focuses on the former category. It presents a collection of patterns that generalize and conceptualize various existing mechanisms to change the visual representation of a process model. Next, it provides a detailed analysis of the degree of support for these patterns in a number of state-of-the-art languages and tools. This paper concludes with the results of a usability evaluation of the patterns conducted with BPM practitioners.
Resumo:
Scoliosis is a spinal deformity that requires surgical correction in progressive cases. In order to optimize surgical outcomes, patient-specific finite element models are being developed by our group. In this paper, a single rod anterior correction procedure is simulated for a group of six scoliosis patients. For each patient, personalised model geometry was derived from low-dose CT scans, and clinically measured intra-operative corrective forces were applied. However, tissue material properties were not patient-specific, being derived from existing literature. Clinically, the patient group had a mean initial Cobb angle of 47.3 degrees, which was corrected to 17.5 degrees after surgery. The mean simulated post-operative Cobb angle for the group was 18.1 degrees. Although this represents good agreement between clinical and simulated corrections, the discrepancy between clinical and simulated Cobb angle for individual patients varied between -10.3 and +8.6 degrees, with only three of the six patients matching the clinical result to within accepted Cobb measurement error of +-5 degrees. The results of this study suggest that spinal tissue material properties play an important role in governing the correction obtained during surgery, and that patient-specific modelling approaches must address the question of how to prescribe patient-specific soft tissue properties for spine surgery simulation.
Resumo:
A computational fluid dynamics (CFD) analysis has been performed for a flat plate photocatalytic reactor using CFD code FLUENT. Under the simulated conditions (Reynolds number, Re around 2650), a detailed time accurate computation shows the different stages of flow evolution and the effects of finite length of the reactor in creating flow instability, which is important to improve the performance of the reactor for storm and wastewater reuse. The efficiency of a photocatalytic reactor for pollutant decontamination depends on reactor hydrodynamics and configurations. This study aims to investigate the role of different parameters on the optimization of the reactor design for its improved performance. In this regard, more modelling and experimental efforts are ongoing to better understand the interplay of the parameters that influence the performance of the flat plate photocatalytic reactor.
Resumo:
The hydrodynamic behaviour of a novel flat plate photocatalytic reactor for water treatment is investigated using CFD code FLUENT. The reactor consists of a reactive section that features negligible pressure drop and uniform illumination of the photocatalyst to ensure enhanced photocatalytic efficiency. The numerical simulations allowed the identification of several design issues in the original reactor, which include extensive boundary layer separation near the photocatalyst support and regions of flow recirculation that render a significant portion of the reactive area. The simulations reveal that this issue could be addressed by selecting the appropriate inlet positions and configurations. This modification can cause minimal pressure drop across the reactive zone and achieves significant uniformization of the tested pollutant on the photocatalyst surface. The influence of roughness elements type has also been studied with a view to identify their role on the distribution of pollutant concentration on the photocatalyst surface. The results presented here indicate that the flow and pollutant concentration field strongly depend on the geometric parameters and flow conditions.
Resumo:
The literature abounds with descriptions of failures in high-profile projects and a range of initiatives has been generated to enhance project management practice (e.g., Morris, 2006). Estimating from our own research, there are scores of other project failures that are unrecorded. Many of these failures can be explained using existing project management theory; poor risk management, inaccurate estimating, cultures of optimism dominating decision making, stakeholder mismanagement, inadequate timeframes, and so on. Nevertheless, in spite of extensive discussion and analysis of failures and attention to the presumed causes of failure, projects continue to fail in unexpected ways. In the 1990s, three U.S. state departments of motor vehicles (DMV) cancelled major projects due to time and cost overruns and inability to meet project goals (IT-Cortex, 2010). The California DMV failed to revitalize their drivers’ license and registration application process after spending $45 million. The Oregon DMV cancelled their five year, $50 million project to automate their manual, paper-based operation after three years when the estimates grew to $123 million; its duration stretched to eight years or more and the prototype was a complete failure. In 1997, the Washington state DMV cancelled their license application mitigation project because it would have been too big and obsolete by the time it was estimated to be finished. There are countless similar examples of projects that have been abandoned or that have not delivered the requirements.