862 resultados para Computer simulation software
Resumo:
This thesis offers a practical and theoretical evaluations about gossip-epidemic algorithms, comparing those most common in the literature with new proposed algorithms and analyzing their behavior. Tests have been executed using one hundred graphs that has been randomly generated by Large Unstructured NEtwork Simulator (LUNES), a simulation software provided by Parallel and Distributed Simulation Research Group (PADS), of the Department of Computer Science, Università di Bologna and simulated using Advanced RTI System (ARTÌS), based on the High Level Architecture standard. Literatures algorithms have been analyzed and taken as base for new algorithms.
Resumo:
PURPOSE: The advent of imaging software programs has proved to be useful for diagnosis, treatment planning, and outcome measurement, but precision of 3-dimensional (3D) surgical simulation still needs to be tested. This study was conducted to determine whether the virtual surgery performed on 3D models constructed from cone-beam computed tomography (CBCT) can correctly simulate the actual surgical outcome and to validate the ability of this emerging technology to recreate the orthognathic surgery hard tissue movements in 3 translational and 3 rotational planes of space. MATERIALS AND METHODS: Construction of pre- and postsurgery 3D models from CBCTs of 14 patients who had combined maxillary advancement and mandibular setback surgery and 6 patients who had 1-piece maxillary advancement surgery was performed. The postsurgery and virtually simulated surgery 3D models were registered at the cranial base to quantify differences between simulated and actual surgery models. Hotelling t tests were used to assess the differences between simulated and actual surgical outcomes. RESULTS: For all anatomic regions of interest, there was no statistically significant difference between the simulated and the actual surgical models. The right lateral ramus was the only region that showed a statistically significant, but small difference when comparing 2- and 1-jaw surgeries. CONCLUSIONS: Virtual surgical methods were reliably reproduced. Oral surgery residents could benefit from virtual surgical training. Computer simulation has the potential to increase predictability in the operating room.
Resumo:
Simulation is an important resource for researchers in diverse fields. However, many researchers have found flaws in the methodology of published simulation studies and have described the state of the simulation community as being in a crisis of credibility. This work describes the project of the Simulation Automation Framework for Experiments (SAFE), which addresses the issues that undermine credibility by automating the workflow in the execution of simulation studies. Automation reduces the number of opportunities for users to introduce error in the scientific process thereby improvingthe credibility of the final results. Automation also eases the job of simulation users and allows them to focus on the design of models and the analysis of results rather than on the complexities of the workflow.
Resumo:
To what extent is “software engineering” really “engineering” as this term is commonly understood? A hallmark of the products of the traditional engineering disciplines is trustworthiness based on dependability. But in his keynote presentation at ICSE 2006 Barry Boehm pointed out that individuals’, systems’, and peoples’ dependency on software is becoming increasingly critical, yet that dependability is generally not the top priority for software intensive system producers. Continuing in an uncharacteristic pessimistic vein, Professor Boehm said that this situation will likely continue until a major software-induced system catastrophe similar in impact to the 9/11 World Trade Center catastrophe stimulates action toward establishing accountability for software dependability. He predicts that it is highly likely that such a software-induced catastrophe will occur between now and 2025. It is widely understood that software, i.e., computer programs, are intrinsically different from traditionally engineered products, but in one aspect they are identical: the extent to which the well-being of individuals, organizations, and society in general increasingly depend on software. As wardens of the future through our mentoring of the next generation of software developers, we believe that it is our responsibility to at least address Professor Boehm’s predicted catastrophe. Traditional engineering has, and continually addresses its social responsibility through the evolution of the education, practice, and professional certification/licensing of professional engineers. To be included in the fraternity of professional engineers, software engineering must do the same. To get a rough idea of where software engineering currently stands on some of these issues we conducted two surveys. Our main survey was sent to software engineering academics in the U.S., Canada, and Australia. Among other items it sought detail information on their software engineering programs. Our auxiliary survey was sent to U.S. engineering institutions to get some idea about how software engineering programs compared with those in established engineering disciplines of Civil, Electrical, and Mechanical Engineering. Summaries of our findings can be found in the last two sections of our paper.
Resumo:
Many schools do not begin to introduce college students to software engineering until they have had at least one semester of programming. Since software engineering is a large, complex, and abstract subject it is difficult to construct active learning exercises that build on the students’ elementary knowledge of programming and still teach basic software engineering principles. It is also the case that beginning students typically know how to construct small programs, but they have little experience with the techniques necessary to produce reliable and long-term maintainable modules. I have addressed these two concerns by defining a local standard (Montana Tech Method (MTM) Software Development Standard for Small Modules Template) that step-by-step directs students toward the construction of highly reliable small modules using well known, best-practices software engineering techniques. “Small module” is here defined as a coherent development task that can be unit tested, and can be car ried out by a single (or a pair of) software engineer(s) in at most a few weeks. The standard describes the process to be used and also provides a template for the top-level documentation. The instructional module’s sequence of mini-lectures and exercises associated with the use of this (and other) local standards are used throughout the course, which perforce covers more abstract software engineering material using traditional reading and writing assignments. The sequence of mini-lectures and hands-on assignments (many of which are done in small groups) constitutes an instructional module that can be used in any similar software engineering course.
Resumo:
We describe four recent additions to NEURON's suite of graphical tools that make it easier for users to create and manage models: an enhancement to the Channel Builder that facilitates the specification and efficient simulation of stochastic channel models
Resumo:
This tutorial gives a step by step explanation of how one uses experimental data to construct a biologically realistic multicompartmental model. Special emphasis is given on the many ways that this process can be imprecise. The tutorial is intended for both experimentalists who want to get into computer modeling and for computer scientists who use abstract neural network models but are curious about biological realistic modeling. The tutorial is not dependent on the use of a specific simulation engine, but rather covers the kind of data needed for constructing a model, how they are used, and potential pitfalls in the process.
Resumo:
It is often claimed that scientists can obtain new knowledge about nature by running computer simulations. How is this possible? I answer this question by arguing that computer simulations are arguments. This view parallels Norton’s argument view about thought experiments. I show that computer simulations can be reconstructed as arguments that fully capture the epistemic power of the simulations. Assuming the extended mind hypothesis, I furthermore argue that running the computer simulation is to execute the reconstructing argument. I discuss some objections and reject the view that computer simulations produce knowledge because they are experiments. I conclude by comparing thought experiments and computer simulations, assuming that both are arguments.
Resumo:
The use of smaller surgical incisions has become popularized for total hip arthroplasty (THR) because of the potential benefits of shorter recovery and improved cosmetic appearance. However, an increased incidence of serious complications has been reported. To minimize the risks of minimally invasive approaches to THR, we have developed an experimental approach which enables us to evaluate risk factors in these procedures through cadaveric simulations performed within the laboratory. During cadaveric hip replacement procedures performed via posterior and antero-lateral mini-incisions, pressures developed between the wound edges and the retractors were approximately double those recorded during conventional hip replacement using Charnley retractors (p < 0.01). In MIS procedures performed via the dual-incision approach, lack of direct visualisation of the proximal femur led to misalignment of broaches and implants with increased risk of cortical fracture during canal preparation and implant insertion. Cadaveric simulation of surgical procedures allows surgeons to measure variables affecting the technical success of surgery and to master new procedures without placing patients at risk.
Resumo:
Withdrawal reflexes of the mollusk Aplysia exhibit sensitization, a simple form of long-term memory (LTM). Sensitization is due, in part, to long-term facilitation (LTF) of sensorimotor neuron synapses. LTF is induced by the modulatory actions of serotonin (5-HT). Pettigrew et al. developed a computational model of the nonlinear intracellular signaling and gene network that underlies the induction of 5-HT-induced LTF. The model simulated empirical observations that repeated applications of 5-HT induce persistent activation of protein kinase A (PKA) and that this persistent activation requires a suprathreshold exposure of 5-HT. This study extends the analysis of the Pettigrew model by applying bifurcation analysis, singularity theory, and numerical simulation. Using singularity theory, classification diagrams of parameter space were constructed, identifying regions with qualitatively different steady-state behaviors. The graphical representation of these regions illustrates the robustness of these regions to changes in model parameters. Because persistent protein kinase A (PKA) activity correlates with Aplysia LTM, the analysis focuses on a positive feedback loop in the model that tends to maintain PKA activity. In this loop, PKA phosphorylates a transcription factor (TF-1), thereby increasing the expression of an ubiquitin hydrolase (Ap-Uch). Ap-Uch then acts to increase PKA activity, closing the loop. This positive feedback loop manifests multiple, coexisting steady states, or multiplicity, which provides a mechanism for a bistable switch in PKA activity. After the removal of 5-HT, the PKA activity either returns to its basal level (reversible switch) or remains at a high level (irreversible switch). Such an irreversible switch might be a mechanism that contributes to the persistence of LTM. The classification diagrams also identify parameters and processes that might be manipulated, perhaps pharmacologically, to enhance the induction of memory. Rational drug design, to affect complex processes such as memory formation, can benefit from this type of analysis.
Resumo:
A model of Drosophila circadian rhythm generation was developed to represent feedback loops based on transcriptional regulation of per, Clk (dclock), Pdp-1, and vri (vrille). The model postulates that histone acetylation kinetics make transcriptional activation a nonlinear function of [CLK]. Such a nonlinearity is essential to simulate robust circadian oscillations of transcription in our model and in previous models. Simulations suggest that two positive feedback loops involving Clk are not essential for oscillations, because oscillations of [PER] were preserved when Clk, vri, or Pdp-1 expression was fixed. However, eliminating positive feedback by fixing vri expression altered the oscillation period. Eliminating the negative feedback loop in which PER represses per expression abolished oscillations. Simulations of per or Clk null mutations, of per overexpression, and of vri, Clk, or Pdp-1 heterozygous null mutations altered model behavior in ways similar to experimental data. The model simulated a photic phase-response curve resembling experimental curves, and oscillations entrained to simulated light-dark cycles. Temperature compensation of oscillation period could be simulated if temperature elevation slowed PER nuclear entry or PER phosphorylation. The model makes experimental predictions, some of which could be tested in transgenic Drosophila.
Resumo:
It is system dynamics that determines the function of cells, tissues and organisms. To develop mathematical models and estimate their parameters are an essential issue for studying dynamic behaviors of biological systems which include metabolic networks, genetic regulatory networks and signal transduction pathways, under perturbation of external stimuli. In general, biological dynamic systems are partially observed. Therefore, a natural way to model dynamic biological systems is to employ nonlinear state-space equations. Although statistical methods for parameter estimation of linear models in biological dynamic systems have been developed intensively in the recent years, the estimation of both states and parameters of nonlinear dynamic systems remains a challenging task. In this report, we apply extended Kalman Filter (EKF) to the estimation of both states and parameters of nonlinear state-space models. To evaluate the performance of the EKF for parameter estimation, we apply the EKF to a simulation dataset and two real datasets: JAK-STAT signal transduction pathway and Ras/Raf/MEK/ERK signaling transduction pathways datasets. The preliminary results show that EKF can accurately estimate the parameters and predict states in nonlinear state-space equations for modeling dynamic biochemical networks.
Resumo:
This paper presents the development history and specification of a 3D game engine titled "Spark Engine". The term "engine" is used to describe a complex graphics software suite that streamlines application development and provides efficient rendering functionality. A game engine specifically provides tools to simplify game development. Spark Engine is fully shader driven and is built on top of Microsoft's XNA Framework. It is a reusable and flexible platform that can be used to build any type of graphics application ranging from gaming to simulation. The engine was released as open source software under the New BSD License with an interest in furthering its development.
Resumo:
The Geographical Simulation Model developed by IDE-JETRO (IDE-GSM) is a computer simulation model based on spatial economics. IDE-GSM enables us to predict the economic impacts of various trade and transport facilitation measures. Here, we mainly compare the prioritized projects of the Master Plan on ASEAN Connectivity (MPAC) and the Comprehensive Asia Development Plan (CADP). MPAC focus on specific hard or soft infrastructure projects that connect one ASEAN member state to another while the CADP emphasizes the importance of economic corridors or linkages between a large cluster and another cluster. As compared with MPAC projects, the simulation analysis shows that CADP projects have much larger positive impacts on ASEAN countries.
Resumo:
Recently, we have presented some studies concerning the analysis, design and optimization of one experimental device developed in the UK - GPTAD - which has been designed to remove blood clots without the need to make contact with the clot itself, thereby potentially reducing the risk of problems such as downstream embolisation. Based on the idea of a modification of the previous device, in this work, we present a model based in the use of stents like the SolitaireTM FR, which is in contact with the clot itself. In the case of such devices, the stent is self-expandable and the extraction of the blood clot is faciliatated by the stent, which must be inside the clot. Such stents are generally inserted in position by using the guidewire inserted into the catheter. This type of modeling could potentially be useful in showing how the blood clot is moved by the various different forces involved. The modelling has been undertaken by analyzing the resistances, compliances and inertances effects. We model an artery and blood clot for range of forces for the guidewire. In each case we determine the interaction between blood clot, stent and artery.