912 resultados para Development Projects
Resumo:
Currently, individuals including designers, contractors, and owners learn about the project requirements by studying a combination of paper and electronic copies of the construction documents including the drawings, specifications (standard and supplemental), road and bridge standard drawings, design criteria, contracts, addenda, and change orders. This can be a tedious process since one needs to go back and forth between the various documents (paper or electronic) to obtain information about the entire project. Object-oriented computer-aided design (OO-CAD) is an innovative technology that can bring a change to this process by graphical portrayal of information. OO-CAD allows users to point and click on portions of an object-oriented drawing that are then linked to relevant databases of information (e.g., specifications, procurement status, and shop drawings). The vision of this study is to turn paper-based design standards and construction specifications into an object-oriented design and specification (OODAS) system or a visual electronic reference library (ERL). Individuals can use the system through a handheld wireless book-size laptop that includes all of the necessary software for operating in a 3D environment. All parties involved in transportation projects can access all of the standards and requirements simultaneously using a 3D graphical interface. By using this system, users will have all of the design elements and all of the specifications readily available without concerns of omissions. A prototype object-oriented model was created and demonstrated to potential users representing counties, cities, and the state. Findings suggest that a system like this could improve productivity to find information by as much as 75% and provide a greater sense of confidence that all relevant information had been identified. It was also apparent that this system would be used by more people in construction than in design. There was also concern related to the cost to develop and maintain the complete system. The future direction should focus on a project-based system that can help the contractors and DOT inspectors find information (e.g., road standards, specifications, instructional memorandums) more rapidly as it pertains to a specific project.
Resumo:
For well over 100 years, the Working Stress Design (WSD) approach has been the traditional basis for geotechnical design with regard to settlements or failure conditions. However, considerable effort has been put forth over the past couple of decades in relation to the adoption of the Load and Resistance Factor Design (LRFD) approach into geotechnical design. With the goal of producing engineered designs with consistent levels of reliability, the Federal Highway Administration (FHWA) issued a policy memorandum on June 28, 2000, requiring all new bridges initiated after October 1, 2007, to be designed according to the LRFD approach. Likewise, regionally calibrated LRFD resistance factors were permitted by the American Association of State Highway and Transportation Officials (AASHTO) to improve the economy of bridge foundation elements. Thus, projects TR-573, TR-583 and TR-584 were undertaken by a research team at Iowa State University’s Bridge Engineering Center with the goal of developing resistance factors for pile design using available pile static load test data. To accomplish this goal, the available data were first analyzed for reliability and then placed in a newly designed relational database management system termed PIle LOad Tests (PILOT), to which this first volume of the final report for project TR-573 is dedicated. PILOT is an amalgamated, electronic source of information consisting of both static and dynamic data for pile load tests conducted in the State of Iowa. The database, which includes historical data on pile load tests dating back to 1966, is intended for use in the establishment of LRFD resistance factors for design and construction control of driven pile foundations in Iowa. Although a considerable amount of geotechnical and pile load test data is available in literature as well as in various State Department of Transportation files, PILOT is one of the first regional databases to be exclusively used in the development of LRFD resistance factors for the design and construction control of driven pile foundations. Currently providing an electronically organized assimilation of geotechnical and pile load test data for 274 piles of various types (e.g., steel H-shaped, timber, pipe, Monotube, and concrete), PILOT (http://srg.cce.iastate.edu/lrfd/) is on par with such familiar national databases used in the calibration of LRFD resistance factors for pile foundations as the FHWA’s Deep Foundation Load Test Database. By narrowing geographical boundaries while maintaining a high number of pile load tests, PILOT exemplifies a model for effective regional LRFD calibration procedures.
Resumo:
In response to the mandate on Load and Resistance Factor Design (LRFD) implementations by the Federal Highway Administration (FHWA) on all new bridge projects initiated after October 1, 2007, the Iowa Highway Research Board (IHRB) sponsored these research projects to develop regional LRFD recommendations. The LRFD development was performed using the Iowa Department of Transportation (DOT) Pile Load Test database (PILOT). To increase the data points for LRFD development, develop LRFD recommendations for dynamic methods, and validate the results of LRFD calibration, 10 full-scale field tests on the most commonly used steel H-piles (e.g., HP 10 x 42) were conducted throughout Iowa. Detailed in situ soil investigations were carried out, push-in pressure cells were installed, and laboratory soil tests were performed. Pile responses during driving, at the end of driving (EOD), and at re-strikes were monitored using the Pile Driving Analyzer (PDA), following with the CAse Pile Wave Analysis Program (CAPWAP) analysis. The hammer blow counts were recorded for Wave Equation Analysis Program (WEAP) and dynamic formulas. Static load tests (SLTs) were performed and the pile capacities were determined based on the Davisson’s criteria. The extensive experimental research studies generated important data for analytical and computational investigations. The SLT measured load-displacements were compared with the simulated results obtained using a model of the TZPILE program and using the modified borehole shear test method. Two analytical pile setup quantification methods, in terms of soil properties, were developed and validated. A new calibration procedure was developed to incorporate pile setup into LRFD.
Resumo:
The Federal Highway Administration (FHWA) mandated utilizing the Load and Resistance Factor Design (LRFD) approach for all new bridges initiated in the United States after October 1, 2007. As a result, there has been a progressive move among state Departments of Transportation (DOTs) toward an increased use of the LRFD in geotechnical design practices. For the above reasons, the Iowa Highway Research Board (IHRB) sponsored three research projects: TR-573, TR-583 and TR-584. The research information is summarized in the project web site (http://srg.cce.iastate.edu/lrfd/). Two reports of total four volumes have been published. Report volume I by Roling et al. (2010) described the development of a user-friendly and electronic database (PILOT). Report volume II by Ng et al. (2011) summarized the 10 full-scale field tests conducted throughout Iowa and data analyses. This report presents the development of regionally calibrated LRFD resistance factors for bridge pile foundations in Iowa based on reliability theory, focusing on the strength limit states and incorporating the construction control aspects and soil setup into the design process. The calibration framework was selected to follow the guidelines provided by the American Association of State Highway and Transportation Officials (AASHTO), taking into consideration the current local practices. The resistance factors were developed for general and in-house static analysis methods used for the design of pile foundations as well as for dynamic analysis methods and dynamic formulas used for construction control. The following notable benefits to the bridge foundation design were attained in this project: 1) comprehensive design tables and charts were developed to facilitate the implementation of the LRFD approach, ensuring uniform reliability and consistency in the design and construction processes of bridge pile foundations; 2) the results showed a substantial gain in the factored capacity compared to the 2008 AASHTO-LRFD recommendations; and 3) contribution to the existing knowledge, thereby advancing the foundation design and construction practices in Iowa and the nation.
Resumo:
Bio-binders can be utilized as asphalt modifiers, extenders, and replacements for conventional asphalt in bituminous binders. From the rheology results of Phase I of this project, it was found that the bio-binders tested had good performance, similar to conventional asphalt, except at low temperatures. Phase II of this project addresses this shortcoming and evaluates the Superpave performance of laboratory mixes produced with the enhanced bio-binders. The main objective of this research was to develop a bio-binder capable of replacing conventional asphalt in flexible pavements by incorporating ground tire rubber (GTR) into bio-oil derived from fast pyrolysis of agriculture and forestry residues. The chemical compatibility of the new bio-binder with GTR was assessed, and the low-temperature performance of the bio-binders was enhanced by the use of GTR. The newly developed binder, which consisted of 80 percent conventional binder and 20 percent rubber-modified bio-oil (85 percent bio-oil with 15 percent GTR), was used to produce mixes at two different air void contents, 4 and 7 percent. The laboratory performance test results showed that the performance of the newly developed bio-binder mixes is as good as or better than conventional asphalt mixes for fatigue cracking, rutting resistance, moisture sensitivity, and low-temperature cracking. These results need to be validated in field projects in order to demonstrate adequate performance for this innovative and sustainable technology for flexible pavements.
Resumo:
The Department of Education, Division of Community Colleges, will annually provide the State Board of Education with the Workforce Training and Economic Development Fund Annual Progress Report. Administration and oversight responsibility for the fund was transferred from the Iowa Economic Development Authority to the Iowa Department of Education effective July 1, 2013 (FY 2014). This report is the first annual progress report produced and distributed by the Iowa Department of Education. The full report outlines the programs, projects, and initiatives that the community colleges have implemented during the past fiscal year.
Resumo:
Managing existing and newly constructed highway corridors has recently become a significant concern in many states, including Iowa. As urban land and land on the urban fringe develops, there is pressure to add features such as commercial driveways, at-grade public road intersections, and traffic signals to arterial highway routes that should primarily serve high-speed traffic. This diminishes the speed and traffic capacity of such roadways and can also cause significant safety issues. if mobility and safety are diminished, the value of the highway investment is diminished. Since a major highway corridor improvement may cost tens of millions of dollars or more, corridor management is as critical to preserving that investment as is more "hard side" management practices such as pavement or bridge management. Corridor management is a process that applies access management principles to highway corridors in an attempt to balance the competing needs of traffic service, safety, and support for land development. This project helped to identify routes that should be given high priority for corridor management. The pilot study in the form of two corridor management case studies provides an analytical process that can be replicated along the other Iowa commuting corridors using commonly available transportation and land use data resources. It also offers a general set of guidelines for the Iowa Department of Transportation to use in the development of its own comprehensive corridor management program.
Resumo:
This report proposes, that for certain types of highway construction projects undertaken by the Iowa Department of Transportation, a scheduling technique commonly referred to as linear scheduling may be more effective than the Critical Path Method scheduling technique that is currently being used. The types of projects that appear to be good candidates for the technique are those projects that have a strong linear orientation. Like a bar chart, this technique shows when an activity is scheduled to occur and like a CPM schedule it shows the sequence in which activities are expected to occur. During the 1992 construction season, the authors worked with an inlay project on Interstate 29 to demonstrate the linear scheduling technique to the Construction Office. The as-planned schedule was developed from the CPM schedule that the contractor had developed for the project. Therefore, this schedule represents what a linear representation of a CPM schedule would look like, and not necessarily what a true linear schedule would look like if it had been the only scheduling technique applied to the project. There is a need to expand the current repertoire of scheduling techniques to address those projects for which the bar chart and CPM may not be appropriate either because of the lack of control information or due to overly complex process for the actual project characteristics. The scheduling approaches used today on transportation projects have many shortcomings for properly modeling the real world constraints and conditions which are encountered. Linear project's predilection for activities with variable production rates, a concept very difficult to handle with the CPM, is easily handled and visualized with the linear technique. It is recommended that work proceed with the refinement of the method of linear scheduling described above and the development of a microcomputer based system for use by the Iowa Department of Transportation and contractors for its implementation. The system will be designed to provide the information needed to adjust schedules in a rational understandable method for monitoring progress on the projects and alerting Iowa Department of Transportation personnel when the contractor is deviating from the plan.
Resumo:
BACKGROUND: Several European HIV observational data bases have, over the last decade, accumulated a substantial number of resistance test results and developed large sample repositories, There is a need to link these efforts together, We here describe the development of such a novel tool that allows to bind these data bases together in a distributed fashion for which the control and data remains with the cohorts rather than classic data mergers.METHODS: As proof-of-concept we entered two basic queries into the tool: available resistance tests and available samples. We asked for patients still alive after 1998-01-01, and between 180 and 195 cm of height, and how many samples or resistance tests there would be available for these patients, The queries were uploaded with the tool to a central web server from which each participating cohort downloaded the queries with the tool and ran them against their database, The numbers gathered were then submitted back to the server and we could accumulate the number of available samples and resistance tests.RESULTS: We obtained the following results from the cohorts on available samples/resistance test: EuResist: not availableI11,194; EuroSIDA: 20,71611,992; ICONA: 3,751/500; Rega: 302/302; SHCS: 53,78311,485, In total, 78,552 samples and 15,473 resistance tests were available amongst these five cohorts. Once these data items have been identified, it is trivial to generate lists of relevant samples that would be usefuI for ultra deep sequencing in addition to the already available resistance tests, Saon the tool will include small analysis packages that allow each cohort to pull a report on their cohort profile and also survey emerging resistance trends in their own cohort,CONCLUSIONS: We plan on providing this tool to all cohorts within the Collaborative HIV and Anti-HIV Drug Resistance Network (CHAIN) and will provide the tool free of charge to others for any non-commercial use, The potential of this tool is to ease collaborations, that is, in projects requiring data to speed up identification of novel resistance mutations by increasing the number of observations across multiple cohorts instead of awaiting single cohorts or studies to reach the critical number needed to address such issues.
Resumo:
Tämän hetken trendit kuten globalisoituminen, ympäristömme turbulenttisuus, elintason nousu, turvallisuuden tarpeen kasvu ja teknologian kehitysnopeus korostavatmuutosten ennakoinnin tarpeellisuutta. Pysyäkseen kilpailukykyisenä yritysten tulee kerätä, analysoida ja hyödyntää liiketoimintatietoa, jokatukee niiden toimintaa viranomaisten, kilpailijoiden ja asiakkaiden toimenpiteiden ennakoinnissa. Innovoinnin ja uusien konseptien kehittäminen, kilpailijoiden toiminnan arviointi, asiakkaiden tarpeet muun muassa vaativatennakoivaa arviointia. Heikot signaalit ovat keskeisessä osassa organisaatioiden valmistautumisessa tulevaisuuden tapahtumiin. Opinnäytetyön tarkoitus on luoda ja kehittää heikkojen signaalien ymmärrystä ja hallintaa sekäkehittää konseptuaalinen ja käytännöllinen lähestymistapa ennakoivan toiminnan edistämiselle. Heikkojen signaalien tyyppien luokittelu perustuu ominaisuuksiin ajan, voimakkuuden ja liiketoimintaan integroinnin suhteen. Erityyppiset heikot signaalit piirteineen luovat reunaehdot laatutekijöiden keräämiselle ja siitä edelleen laatujärjestelmän ja matemaattiseen malliin perustuvan työvälineen kehittämiselle. Heikkojen signaalien laatutekijät on kerätty yhteen kaikista heikkojen signaalien konseptin alueista. Analysoidut ja kohdistetut laatumuuttujat antavat mahdollisuuden kehittää esianalyysiä ja ICT - työvälineitä perustuen matemaattisen mallin käyttöön. Opinnäytetyön tavoitteiden saavuttamiseksi tehtiin ensin Business Intelligence -kirjallisuustutkimus. Hiekkojen signaalien prosessi ja systeemi perustuvat koottuun Business Intelligence - systeemiin. Keskeisinä kehitysalueina tarkasteltiin liiketoiminnan integraatiota ja systemaattisen menetelmän kehitysaluetta. Heikkojen signaalien menetelmien ja määritelmien kerääminen sekä integrointi määriteltyyn prosessiin luovat uuden konseptin perustan, johon tyypitys ja laatutekijät kytkeytyvät. Käytännöllisen toiminnan tarkastelun ja käyttöönoton mahdollistamiseksi toteutettiin Business Intelligence markkinatutkimus (n=156) sekä yhteenveto muihin saatavilla oleviin markkinatutkimuksiin. Syvähaastatteluilla (n=21) varmennettiin laadullisen tarkastelun oikeellisuus. Lisäksi analysoitiin neljä käytännön projektia, joiden yhteenvedot kytkettiin uuden konseptin kehittämiseen. Prosessi voidaan jakaa kahteen luokkaan: yritysten markkinasignaalit vuoden ennakoinnilla ja julkisen sektorin verkostoprojektit kehittäen ennakoinnin struktuurin luonnin 7-15 vuoden ennakoivalle toiminnalle. Tutkimus rajattiin koskemaan pääasiassa ulkoisen tiedon aluetta. IT työvälineet ja lopullisen laatusysteemin kehittäminen jätettiin tutkimuksen ulkopuolelle. Opinnäytetyön tavoitteena ollut heikkojen signaalien konseptin kehittäminen toteutti sille asetetut odotusarvot. Heikkojen signaalien systemaattista tarkastelua ja kehittämistyötä on mahdollista edistää Business Intelligence - systematiikan hyödyntämisellä. Business Intelligence - systematiikkaa käytetään isojen yritysten liiketoiminnan suunnittelun tukena.Organisaatioiden toiminnassa ei ole kuitenkaan yleisesti hyödynnetty laadulliseen analyysiin tukeutuvaa ennakoinnin weak signals - toimintaa. Ulkoisenja sisäisen tiedon integroinnin ja systematiikan hyödyt PK -yritysten tukena vaativat merkittävää panostusta julkishallinnon rahoituksen ja kehitystoiminnan tukimuotoina. Ennakointi onkin tuottanut lukuisia julkishallinnon raportteja, mutta ei käytännön toteutuksia. Toisaalta analysoitujen case-tapausten tuloksena voidaan nähdä, ettei organisaatioissa välttämättä tarvita omaa projektipäällikköä liiketoiminnan tuen kehittämiseksi. Business vastuun ottamiseksi ja asiaan sitoutumiseen on kuitenkin löydyttävä oikea henkilö
Resumo:
This paper reviews the literature on managerially actionable new product development success factors and summarises the field in a classic managerial framework. Because of the varying quality, breadth and scope of the field, the review only contains post-1980 studies of tangible product development that are of a rigorous scientific standard. Success is interpreted as a commercial success. The field has gained insight into a broad set of factors that vary in scope, abstraction and context. Main areas that contribute to NPD success are top management support exhibited through resource allocation and communicating the strategic importance of NPD in the organisation. The right projects need to be selected for investment at the beginning of the process and should be aligned to the organisation's internal competencies and the external environment. The NPD process should use cross-functional teams and a competent project champions. Marketing research competency is crucial, as an understanding of the market, customers and competitors is repeatedly highlighted. Product launch competency was also consistently shown to be important. In terms of controlling the NPD process, strict project gates are required to maintain control.
Resumo:
Software engineering is criticized as not being engineering or 'well-developed' science at all. Software engineers seem not to know exactly how long their projects will last, what they will cost, and will the software work properly after release. Measurements have to be taken in software projects to improve this situation. It is of limited use to only collect metrics afterwards. The values of the relevant metrics have to be predicted, too. The predictions (i.e. estimates) form the basis for proper project management. One of the most painful problems in software projects is effort estimation. It has a clear and central effect on other project attributes like cost and schedule, and to product attributes like size and quality. Effort estimation can be used for several purposes. In this thesis only the effort estimation in software projects for project management purposes is discussed. There is a short introduction to the measurement issues, and some metrics relevantin estimation context are presented. Effort estimation methods are covered quite broadly. The main new contribution in this thesis is the new estimation model that has been created. It takes use of the basic concepts of Function Point Analysis, but avoids the problems and pitfalls found in the method. It is relativelyeasy to use and learn. Effort estimation accuracy has significantly improved after taking this model into use. A major innovation related to the new estimationmodel is the identified need for hierarchical software size measurement. The author of this thesis has developed a three level solution for the estimation model. All currently used size metrics are static in nature, but this new proposed metric is dynamic. It takes use of the increased understanding of the nature of the work as specification and design work proceeds. It thus 'grows up' along with software projects. The effort estimation model development is not possible without gathering and analyzing history data. However, there are many problems with data in software engineering. A major roadblock is the amount and quality of data available. This thesis shows some useful techniques that have been successful in gathering and analyzing the data needed. An estimation process is needed to ensure that methods are used in a proper way, estimates are stored, reported and analyzed properly, and they are used for project management activities. A higher mechanism called measurement framework is also introduced shortly. The purpose of the framework is to define and maintain a measurement or estimationprocess. Without a proper framework, the estimation capability of an organization declines. It requires effort even to maintain an achieved level of estimationaccuracy. Estimation results in several successive releases are analyzed. It isclearly seen that the new estimation model works and the estimation improvementactions have been successful. The calibration of the hierarchical model is a critical activity. An example is shown to shed more light on the calibration and the model itself. There are also remarks about the sensitivity of the model. Finally, an example of usage is shown.
Resumo:
La présente thèse s'intitule "Développent et Application des Méthodologies Computationnelles pour la Modélisation Qualitative". Elle comprend tous les différents projets que j'ai entrepris en tant que doctorante. Plutôt qu'une mise en oeuvre systématique d'un cadre défini a priori, cette thèse devrait être considérée comme une exploration des méthodes qui peuvent nous aider à déduire le plan de processus regulatoires et de signalisation. Cette exploration a été mue par des questions biologiques concrètes, plutôt que par des investigations théoriques. Bien que tous les projets aient inclus des systèmes divergents (réseaux régulateurs de gènes du cycle cellulaire, réseaux de signalisation de cellules pulmonaires) ainsi que des organismes (levure à fission, levure bourgeonnante, rat, humain), nos objectifs étaient complémentaires et cohérents. Le projet principal de la thèse est la modélisation du réseau de l'initiation de septation (SIN) du S.pombe. La cytokinèse dans la levure à fission est contrôlée par le SIN, un réseau signalant de protéines kinases qui utilise le corps à pôle-fuseau comme échafaudage. Afin de décrire le comportement qualitatif du système et prédire des comportements mutants inconnus, nous avons décidé d'adopter l'approche de la modélisation booléenne. Dans cette thèse, nous présentons la construction d'un modèle booléen étendu du SIN, comprenant la plupart des composantes et des régulateurs du SIN en tant que noeuds individuels et testable expérimentalement. Ce modèle utilise des niveaux d'activité du CDK comme noeuds de contrôle pour la simulation d'évènements du SIN à différents stades du cycle cellulaire. Ce modèle a été optimisé en utilisant des expériences d'un seul "knock-out" avec des effets phénotypiques connus comme set d'entraînement. Il a permis de prédire correctement un set d'évaluation de "knock-out" doubles. De plus, le modèle a fait des prédictions in silico qui ont été validées in vivo, permettant d'obtenir de nouvelles idées de la régulation et l'organisation hiérarchique du SIN. Un autre projet concernant le cycle cellulaire qui fait partie de cette thèse a été la construction d'un modèle qualitatif et minimal de la réciprocité des cyclines dans la S.cerevisiae. Les protéines Clb dans la levure bourgeonnante présentent une activation et une dégradation caractéristique et séquentielle durant le cycle cellulaire, qu'on appelle communément les vagues des Clbs. Cet évènement est coordonné avec la courbe d'activation inverse du Sic1, qui a un rôle inhibitoire dans le système. Pour l'identification des modèles qualitatifs minimaux qui peuvent expliquer ce phénomène, nous avons sélectionné des expériences bien définies et construit tous les modèles minimaux possibles qui, une fois simulés, reproduisent les résultats attendus. Les modèles ont été filtrés en utilisant des simulations ODE qualitatives et standardisées; seules celles qui reproduisaient le phénotype des vagues ont été gardées. L'ensemble des modèles minimaux peut être utilisé pour suggérer des relations regulatoires entre les molécules participant qui peuvent ensuite être testées expérimentalement. Enfin, durant mon doctorat, j'ai participé au SBV Improver Challenge. Le but était de déduire des réseaux spécifiques à des espèces (humain et rat) en utilisant des données de phosphoprotéines, d'expressions des gènes et des cytokines, ainsi qu'un réseau de référence, qui était mis à disposition comme donnée préalable. Notre solution pour ce concours a pris la troisième place. L'approche utilisée est expliquée en détail dans le dernier chapitre de la thèse. -- The present dissertation is entitled "Development and Application of Computational Methodologies in Qualitative Modeling". It encompasses the diverse projects that were undertaken during my time as a PhD student. Instead of a systematic implementation of a framework defined a priori, this thesis should be considered as an exploration of the methods that can help us infer the blueprint of regulatory and signaling processes. This exploration was driven by concrete biological questions, rather than theoretical investigation. Even though the projects involved divergent systems (gene regulatory networks of cell cycle, signaling networks in lung cells), as well as organisms (fission yeast, budding yeast, rat, human), our goals were complementary and coherent. The main project of the thesis is the modeling of the Septation Initiation Network (SIN) in S.pombe. Cytokinesis in fission yeast is controlled by the SIN, a protein kinase signaling network that uses the spindle pole body as scaffold. In order to describe the qualitative behavior of the system and predict unknown mutant behaviors we decided to adopt a Boolean modeling approach. In this thesis, we report the construction of an extended, Boolean model of the SIN, comprising most SIN components and regulators as individual, experimentally testable nodes. The model uses CDK activity levels as control nodes for the simulation of SIN related events in different stages of the cell cycle. The model was optimized using single knock-out experiments of known phenotypic effect as a training set, and was able to correctly predict a double knock-out test set. Moreover, the model has made in silico predictions that have been validated in vivo, providing new insights into the regulation and hierarchical organization of the SIN. Another cell cycle related project that is part of this thesis was to create a qualitative, minimal model of cyclin interplay in S.cerevisiae. CLB proteins in budding yeast present a characteristic, sequential activation and decay during the cell cycle, commonly referred to as Clb waves. This event is coordinated with the inverse activation curve of Sic1, which has an inhibitory role in the system. To generate minimal qualitative models that can explain this phenomenon, we selected well-defined experiments and constructed all possible minimal models that, when simulated, reproduce the expected results. The models were filtered using standardized qualitative ODE simulations; only the ones reproducing the wave-like phenotype were kept. The set of minimal models can be used to suggest regulatory relations among the participating molecules, which will subsequently be tested experimentally. Finally, during my PhD I participated in the SBV Improver Challenge. The goal was to infer species-specific (human and rat) networks, using phosphoprotein, gene expression and cytokine data and a reference network provided as prior knowledge. Our solution to the challenge was selected as in the final chapter of the thesis.
Resumo:
Agile software development methods are attempting to provide an answer to the software development industry's need of lighter weight, more agile processes that offer the possibility to react to changes during the software development process. The objective of this thesis is to analyze and experiment the possibility of using agile methods or practices also in small software projects, even in projects containing only one developer. In the practical part of the thesis a small software project was executed with some agile methods and practices that in the theoretical part of the thesis were found possible to be applied to the project. In the project a Bluetooth proxy application that is run in the S60 smartphone platform and PC was developed further to contain some new features. As a result it was found that certain agile practices can be useful even in the very small projects. The selection of the suitable practices depends on the project and the size of the project team.
Resumo:
Vaatimusmäärittelyn tavoitteena on luoda halutun järjestelmän kokonaisen, yhtenäisen vaatimusluettelon vaatimusten määrittämiseksi käsitteellisellä tasolla. Liiketoimintaprosessien mallintaminen on varsin hyödyllinen vaatimusmäärittelyn varhaisissa vaiheissa. Tämä työ tutkii liiketoimintaprosessien mallintamista tietojärjestelmien kehittämistä varten. Nykyään on olemassa erilaisia liiketoimintaprosessien mallintamiseen tarkoitettuja tekniikoita. Tämä työ tarkastaa liiketoimintaprosessien mallintamisen periaatteet ja näkökohdat sekä eri mallinnustekniikoita. Uusi menetelmä, joka on suunniteltu erityisesti pienille ja keskisuurille ohjelmistoprojekteille, on kehitetty prosessinäkökohtien ja UML-kaavioiden perusteella.