825 resultados para network traffic analysis
Resumo:
This research identifies roadway, traffic, and environmental factors that influence the injury severity of road traffic crashes in Dhaka. Dhaka provides a rather unusual driving risk environment to study, since virtually anyone can obtain a drivers’ license and very little traffic enforcement and fines are given when drivers violate traffic rules. To examine this city with presumed heightened crash severity risk, police reported crash data from 2007 to 2011 containing about 2714 road traffic crashes were collected. The injury severity of traffic crashes—recorded as either fatal, serious injury, or property damage only—were modeled using an ordered Probit model. Significant factors increasing the probability of fatal injuries include crashes along highways (65%), absence of a road divider (80%), crashes during night time (54%), and vehicle-pedestrian collisions (367%); whereas two-way traffic configuration (21%), and traffic police controlled schemes (41%) decrease the probability of fatalities. Both similarities and differences of the findings between crash risk in Dhaka and developed countries are discussed in policy relevant terms.
Resumo:
Temporary Traffic Control Plans (TCP’s), which provide construction phasing to maintain traffic during construction operations, are integral component of highway construction project design. Using the initial design, designers develop estimated quantities for the required TCP devices that become the basis for bids submitted by highway contractors. However, actual as-built quantities are often significantly different from the engineer’s original estimate. The total cost of TCP phasing on highway construction projects amounts to 6–10% of the total construction cost. Variations between engineer estimated quantities and final quantities contribute to reduced cost control, increased chances of cost related litigations, and bid rankings and selection. Statistical analyses of over 2000 highway construction projects were performed to determine the sources of variation, which later were used as the basis of development for an automated-hybrid prediction model that uses multiple regressions and heuristic rules to provide accurate TCP quantities and costs. The predictive accuracy of the model developed was demonstrated through several case studies.
Resumo:
The use of Wireless Sensor Networks (WSNs) for vibration-based Structural Health Monitoring (SHM) has become a promising approach due to many advantages such as low cost, fast and flexible deployment. However, inherent technical issues such as data asynchronicity and data loss have prevented these distinct systems from being extensively used. Recently, several SHM-oriented WSNs have been proposed and believed to be able to overcome a large number of technical uncertainties. Nevertheless, there is limited research verifying the applicability of those WSNs with respect to demanding SHM applications like modal analysis and damage identification. Based on a brief review, this paper first reveals that Data Synchronization Error (DSE) is the most inherent factor amongst uncertainties of SHM-oriented WSNs. Effects of this factor are then investigated on outcomes and performance of the most robust Output-only Modal Analysis (OMA) techniques when merging data from multiple sensor setups. The two OMA families selected for this investigation are Frequency Domain Decomposition (FDD) and data-driven Stochastic Subspace Identification (SSI-data) due to the fact that they both have been widely applied in the past decade. Accelerations collected by a wired sensory system on a large-scale laboratory bridge model are initially used as benchmark data after being added with a certain level of noise to account for the higher presence of this factor in SHM-oriented WSNs. From this source, a large number of simulations have been made to generate multiple DSE-corrupted datasets to facilitate statistical analyses. The results of this study show the robustness of FDD and the precautions needed for SSI-data family when dealing with DSE at a relaxed level. Finally, the combination of preferred OMA techniques and the use of the channel projection for the time-domain OMA technique to cope with DSE are recommended.
Resumo:
The practice of medicine has always aimed at individualized treatment of disease. The relationship between patient and physician has always been a personal one, and the physician's choice of treatment has been intended to be the best fit for the patient's needs. The necessary pooling/grouping of disease families and their assignment to a number of drugs or treatment methods has, consequently, led to an increase in the number of effective therapies. However, given the heterogeneity of most human diseases, and cancer specifically, it is currently impossible for the treating clinician to effectively predict a patient's response and outcome based on current technologies, much less the idiosyncratic resistances and adverse effects associated with the limited therapeutic options.
Resumo:
Cancer can be defined as a deregulation or hyperactivity in the ongoing network of intracellular and extracellular signaling events. Reverse phase protein microarray technology may offer a new opportunity to measure and profile these signaling pathways, providing data on post-translational phosphorylation events not obtainable by gene microarray analysis. Treatment of ovarian epithelial carcinoma almost always takes place in a metastatic setting since unfortunately the disease is often not detected until later stages. Thus, in addition to elucidation of the molecular network within a tumor specimen, critical questions are to what extent do signaling changes occur upon metastasis and are there common pathway elements that arise in the metastatic microenvironment. For individualized combinatorial therapy, ideal therapeutic selection based on proteomic mapping of phosphorylation end points may require evaluation of the patient's metastatic tissue. Extending these findings to the bedside will require the development of optimized protocols and reference standards. We have developed a reference standard based on a mixture of phosphorylated peptides to begin to address this challenge.
Resumo:
Traffic incidents are key contributors to non-recurrent congestion, potentially generating significant delay. Factors that influence the duration of incidents are important to understand so that effective mitigation strategies can be implemented. To identify and quantify the effects of influential factors, a methodology for studying total incident duration based on historical data from an ‘integrated database’ is proposed. Incident duration models are developed using a selected freeway segment in the Southeast Queensland, Australia network. The models include incident detection and recovery time as components of incident duration. A hazard-based duration modelling approach is applied to model incident duration as a function of a variety of factors that influence traffic incident duration. Parametric accelerated failure time survival models are developed to capture heterogeneity as a function of explanatory variables, with both fixed and random parameters specifications. The analysis reveals that factors affecting incident duration include incident characteristics (severity, type, injury, medical requirements, etc.), infrastructure characteristics (roadway shoulder availability), time of day, and traffic characteristics. The results indicate that event type durations are uniquely different, thus requiring different responses to effectively clear them. Furthermore, the results highlight the presence of unobserved incident duration heterogeneity as captured by the random parameter models, suggesting that additional factors need to be considered in future modelling efforts.
Resumo:
This paper introduces a new method to automate the detection of marine species in aerial imagery using a Machine Learning approach. Our proposed system has at its core, a convolutional neural network. We compare this trainable classifier to a handcrafted classifier based on color features, entropy and shape analysis. Experiments demonstrate that the convolutional neural network outperforms the handcrafted solution. We also introduce a negative training example-selection method for situations where the original training set consists of a collection of labeled images in which the objects of interest (positive examples) have been marked by a bounding box. We show that picking random rectangles from the background is not necessarily the best way to generate useful negative examples with respect to learning.
Resumo:
This article analyses co-movements in a wide group of commodity prices during the time period 1992–2010. Our methodological approach is based on the correlation matrix and the networks inside. Through this approach we are able to summarize global interaction and interdependence, capturing the existing heterogeneity in the degrees of synchronization between commodity prices. Our results produce two main findings: (a) we do not observe a persistent increase in the degree of co-movement of the commodity prices in our time sample, however from mid-2008 to the end of 2009 co-movements almost doubled when compared with the average correlation; (b) we observe three groups of commodities which have exhibited similar price dynamics (metals, oil and grains, and oilseeds) and which have increased their degree of co-movement during the sampled period.
Resumo:
A systematic literature review and a comprehensive meta-analysis that combines the findings from existing studies, was conducted in this thesis to analyse the impact of traffic characteristics on crash occurrence. Sensitivity analyses were conducted to investigate the quality, publication bias and outlier bias of the various studies, and the time intervals used to measure traffic characteristics were considered. Based on this comprehensive and systematic review, and the results of the subsequent meta-analysis, major issues in study design, traffic and crash data, and model development and evaluation are discussed.
Resumo:
The Hong Kong construction industry is currently facing ageing problem and labour shortage. There are opportunities for employing ethnic minority construction workers to join this hazardous industry. These ethnic minority workers are prone to accidents due to communication barriers. Safety communication is playing an important role for avoiding the accidents on construction sites. However, the ethnic minority workers are not very fluent in the local language and facing safety communication problems while working with local workers. Social network analysis (SNA), being an effective tool to identify the safety communication flow on the construction site, is used to attain the measures of safety communication like centrality, density and betweenness within the ethnic minorities and local workers, and to generate sociograms that visually represent communication pattern within the effective and ineffective safety networks. The aim of this paper is to present the application of SNA for improving the safety communication of ethnic minorities in the construction industry of Hong Kong. The paper provides the theoretical background of SNA approaches for the data collection and analysis using the software UCINET and NetDraw, to determine the predominant safety communication network structure and pattern of ethnic minorities on site.
Resumo:
Railways are an important mode of transportation. They are however large and complex and their construction, management and operation is time consuming and costly. Evidently planning the current and future activities is vital. Part of that planning process is an analysis of capacity. To determine what volume of traffic can be achieved over time, a variety of railway capacity analysis techniques have been created. A generic analytical approach that incorporates more complex train paths however has yet to be provided. This article provides such an approach. This article extends a mathematical model for determining the theoretical capacity of a railway network. The main contribution of this paper is the modelling of more complex train paths whereby each section can be visited many times in the course of a train’s journey. Three variant models are formulated and then demonstrated in a case study. This article’s numerical investigations have successively shown the applicability of the proposed models and how they may be used to gain insights into system performance.
Resumo:
The development of methods for real-time crash prediction as a function of current or recent traffic and roadway conditions is gaining increasing attention in the literature. Numerous studies have modeled the relationships between traffic characteristics and crash occurrence, and significant progress has been made. Given the accumulated evidence on this topic and the lack of an articulate summary of research status, challenges, and opportunities, there is an urgent need to scientifically review these studies and to synthesize the existing state-of-the-art knowledge. This paper addresses this need by undertaking a systematic literature review to identify current knowledge, challenges, and opportunities, and then conducts a meta-analysis of existing studies to provide a summary impact of traffic characteristics on crash occurrence. Sensitivity analyses were conducted to assess quality, publication bias, and outlier bias of the various studies; and the time intervals used to measure traffic characteristics were also considered. As a result of this comprehensive and systematic review, issues in study designs, traffic and crash data, and model development and validation are discussed. Outcomes of this study are intended to provide researchers focused on real-time crash prediction with greater insight into the modeling of this important but extremely challenging safety issue.