937 resultados para Distributed model predictive control
Resumo:
The number of electronic devices connected to agricultural machinery is increasing to support new agricultural practices tasks related to the Precision Agriculture such as spatial variability mapping and Variable Rate Technology (VRT). The Distributed Control System (DCS) is a suitable solution for decentralization of the data acquisition system and the Controller Area Network (CAN) is the major trend among the embedded communications protocols for agricultural machinery and vehicles. The application of soil correctives is a typical problem in Brazil. The efficiency of this correction process is highly dependent of the inputs way at soil and the occurrence of errors affects directly the agricultural yield. To handle this problem, this paper presents the development of a CAN-based distributed control system for a VRT system of soil corrective in agricultural machinery. The VRT system is composed by a tractor-implement that applies a desired rate of inputs according to the georeferenced prescription map of the farm field to support PA (Precision Agriculture). The performance evaluation of the CAN-based VRT system was done by experimental tests and analyzing the CAN messages transmitted in the operation of the entire system. The results of the control error according to the necessity of agricultural application allow conclude that the developed VRT system is suitable for the agricultural productions reaching an acceptable response time and application error. The CAN-Based DCS solution applied in the VRT system reduced the complexity of the control system, easing the installation and maintenance. The use of VRT system allowed applying only the required inputs, increasing the efficiency operation and minimizing the environmental impact.
Resumo:
We analyze new results on a magnetically levitated body (a block including a magnet whose bottom pole is set in such a way as to repel the upper pole of a magnetic base) excited by a non-ideal energy source (an unbalanced electric motor of limited power supply). These new results are related to the jump phenomena and increase of power required of such sources near resonance are manifestations of a non-ideal system and they are referred as the Sommerfeld effect, which emulates an energy sink. In this work, we also discuss control strategies to be applied to this system, in resonance conditions, in order to decrease its vibration amplitude and avoiding this apparent energy sink.
Resumo:
Access control is a fundamental concern in any system that manages resources, e.g., operating systems, file systems, databases and communications systems. The problem we address is how to specify, enforce, and implement access control in distributed environments. This problem occurs in many applications such as management of distributed project resources, e-newspaper and payTV subscription services. Starting from an access relation between users and resources, we derive a user hierarchy, a resource hierarchy, and a unified hierarchy. The unified hierarchy is then used to specify the access relation in a way that is compact and that allows efficient queries. It is also used in cryptographic schemes that enforce the access relation. We introduce three specific cryptography based hierarchical schemes, which can effectively enforce and implement access control and are designed for distributed environments because they do not need the presence of a central authority (except perhaps for set- UP).
Resumo:
A data set of a commercial Nellore beef cattle selection program was used to compare breeding models that assumed or not markers effects to estimate the breeding values, when a reduced number of animals have phenotypic, genotypic and pedigree information available. This herd complete data set was composed of 83,404 animals measured for weaning weight (WW), post-weaning gain (PWG), scrotal circumference (SC) and muscle score (MS), corresponding to 116,652 animals in the relationship matrix. Single trait analyses were performed by MTDFREML software to estimate fixed and random effects solutions using this complete data. The additive effects estimated were assumed as the reference breeding values for those animals. The individual observed phenotype of each trait was adjusted for fixed and random effects solutions, except for direct additive effects. The adjusted phenotype composed of the additive and residual parts of observed phenotype was used as dependent variable for models' comparison. Among all measured animals of this herd, only 3160 animals were genotyped for 106 SNP markers. Three models were compared in terms of changes on animals' rank, global fit and predictive ability. Model 1 included only polygenic effects, model 2 included only markers effects and model 3 included both polygenic and markers effects. Bayesian inference via Markov chain Monte Carlo methods performed by TM software was used to analyze the data for model comparison. Two different priors were adopted for markers effects in models 2 and 3, the first prior assumed was a uniform distribution (U) and, as a second prior, was assumed that markers effects were distributed as normal (N). Higher rank correlation coefficients were observed for models 3_U and 3_N, indicating a greater similarity of these models animals' rank and the rank based on the reference breeding values. Model 3_N presented a better global fit, as demonstrated by its low DIC. The best models in terms of predictive ability were models 1 and 3_N. Differences due prior assumed to markers effects in models 2 and 3 could be attributed to the better ability of normal prior in handle with collinear effects. The models 2_U and 2_N presented the worst performance, indicating that this small set of markers should not be used to genetically evaluate animals with no data, since its predictive ability is restricted. In conclusion, model 3_N presented a slight superiority when a reduce number of animals have phenotypic, genotypic and pedigree information. It could be attributed to the variation retained by markers and polygenic effects assumed together and the normal prior assumed to markers effects, that deals better with the collinearity between markers. (C) 2012 Elsevier B.V. All rights reserved.
Resumo:
Adult stem cells are distributed through the whole organism, and present a great potential for the therapy of different types of disease. For the design of efficient therapeutic strategies, it is important to have a more detailed understanding of their basic biological characteristics, as well as of the signals produced by damaged tissues and to which they respond. Myocardial infarction (MI), a disease caused by a lack of blood flow supply in the heart, represents the most common cause of morbidity and mortality in the Western world. Stem cell therapy arises as a promising alternative to conventional treatments, which are often ineffective in preventing loss of cardiomyocytes and fibrosis. Cell therapy protocols must take into account the molecular events that occur in the regenerative niche of MI. In the present study, we investigated the expression profile of ten genes coding for chemokines or cytokines in a murine model of MI, aiming at the characterization of the regenerative niche. MI was induced in adult C57BL/6 mice and heart samples were collected after 24 h and 30 days, as well as from control animals, for quantitative RT-PCR. Expression of the chemokine genes CCL2, CCL3, CCL4, CCL7, CXCL2 and CXCL10 was significantly increased 24 h after infarction, returning to baseline levels on day 30. Expression of the CCL8 gene significantly increased only on day 30, whereas gene expression of CXCL12 and CX3CL1 were not significantly increased in either ischemic period. Finally, expression of the IL-6 gene increased 24 h after infarction and was maintained at a significantly higher level than control samples 30 days later. These results contribute to the better knowledge of the regenerative niche in MI, allowing a more efficient selection or genetic manipulation of cells in therapeutic protocols.
Resumo:
Abstract Background Smear negative pulmonary tuberculosis (SNPT) accounts for 30% of pulmonary tuberculosis cases reported yearly in Brazil. This study aimed to develop a prediction model for SNPT for outpatients in areas with scarce resources. Methods The study enrolled 551 patients with clinical-radiological suspicion of SNPT, in Rio de Janeiro, Brazil. The original data was divided into two equivalent samples for generation and validation of the prediction models. Symptoms, physical signs and chest X-rays were used for constructing logistic regression and classification and regression tree models. From the logistic regression, we generated a clinical and radiological prediction score. The area under the receiver operator characteristic curve, sensitivity, and specificity were used to evaluate the model's performance in both generation and validation samples. Results It was possible to generate predictive models for SNPT with sensitivity ranging from 64% to 71% and specificity ranging from 58% to 76%. Conclusion The results suggest that those models might be useful as screening tools for estimating the risk of SNPT, optimizing the utilization of more expensive tests, and avoiding costs of unnecessary anti-tuberculosis treatment. Those models might be cost-effective tools in a health care network with hierarchical distribution of scarce resources.
Resumo:
Abstract Background To understand the molecular mechanisms underlying important biological processes, a detailed description of the gene products networks involved is required. In order to define and understand such molecular networks, some statistical methods are proposed in the literature to estimate gene regulatory networks from time-series microarray data. However, several problems still need to be overcome. Firstly, information flow need to be inferred, in addition to the correlation between genes. Secondly, we usually try to identify large networks from a large number of genes (parameters) originating from a smaller number of microarray experiments (samples). Due to this situation, which is rather frequent in Bioinformatics, it is difficult to perform statistical tests using methods that model large gene-gene networks. In addition, most of the models are based on dimension reduction using clustering techniques, therefore, the resulting network is not a gene-gene network but a module-module network. Here, we present the Sparse Vector Autoregressive model as a solution to these problems. Results We have applied the Sparse Vector Autoregressive model to estimate gene regulatory networks based on gene expression profiles obtained from time-series microarray experiments. Through extensive simulations, by applying the SVAR method to artificial regulatory networks, we show that SVAR can infer true positive edges even under conditions in which the number of samples is smaller than the number of genes. Moreover, it is possible to control for false positives, a significant advantage when compared to other methods described in the literature, which are based on ranks or score functions. By applying SVAR to actual HeLa cell cycle gene expression data, we were able to identify well known transcription factor targets. Conclusion The proposed SVAR method is able to model gene regulatory networks in frequent situations in which the number of samples is lower than the number of genes, making it possible to naturally infer partial Granger causalities without any a priori information. In addition, we present a statistical test to control the false discovery rate, which was not previously possible using other gene regulatory network models.
Resumo:
Background The increase in fructose consumption is paralleled by a higher incidence of metabolic syndrome, and consequently, cardiovascular disease mortality. We examined the effects of 8 weeks of low intensity exercise training (LET) on metabolic, hemodynamic, ventricular and vascular morphological changes induced by fructose drinking in male rats. Methods Male Wistar rats were divided into (n = 8 each) control (C), sedentary fructose (F) and ET fructose (FT) groups. Fructose-drinking rats received D-fructose (100 g/l). FT rats were assigned to a treadmill training protocol at low intensity (30% of maximal running speed) during 1 h/day, 5 days/week for 8 weeks. Measurements of triglyceride concentrations, white adipose tissue (WAT) and glycemia were carried out together with insulin tolerance test to evaluate metabolic profile. Arterial pressure (AP) signals were directly recorded. Baroreflex sensitivity (BS) was evaluated by the tachycardic and bradycardic responses. Right atria, left ventricle (LV) and ascending aorta were prepared to morphoquantitative analysis. Results LET reduced WAT (−37.7%), triglyceride levels (−33%), systolic AP (−6%), heart weight/body weight (−20.5%), LV (−36%) and aortic (−76%) collagen fibers, aortic intima-media thickness and circumferential wall tension in FT when compared to F rats. Additionally, FT group presented improve of BS, numerical density of atrial natriuretic peptide granules (+42%) and LV capillaries (+25%), as well as the number of elastic lamellae in aorta compared with F group. Conclusions Our data suggest that LET, a widely recommended practice, seems to be particularly effective for preventing metabolic, hemodynamic and morphological disorders triggered by MS.
Resumo:
Abstract Introduction Sclerostin levels have been reported to be low in ankylosing spondylitis (AS), but there is no data regarding the possible role of this Wnt inhibitor during anti-tumor necrosis factor (TNF) therapy. The present study longitudinally evaluated sclerostin levels, inflammatory markers and bone mineral density (BMD) in AS patients under anti-TNF therapy. Methods Thirty active AS patients were assessed at baseline, 6 and 12 months after anti-TNF therapy regarding clinical parameters, inflammatory markers, BMD and baseline radiographic damage (mSASSS). Thirty age- and sex-matched healthy individuals comprised the control group. Patients' sclerostin levels, sclerostin binding low-density lipoprotein receptor-related protein 6 (LRP6) and BMD were evaluated at the same time points and compared to controls. Results At baseline, AS patients had lower sclerostin levels (60.5 ± 32.7 vs. 96.7 ± 52.9 pmol/L, P = 0.002) and comparable sclerostin binding to LRP6 (P = 0.387) than controls. Improvement of Bath Ankylosing Spondylitis Disease Activity Index (BASDAI), Bath Ankylosing Spondylitis Functional Index (BASFI), Bath Ankylosing Spondylitis Metrology Index (BASMI), Ankylosing Spondylitis quality of life (ASQoL) was observed at baseline vs. 6 vs. 12 months (P < 0.01). Concomitantly, a gradual increase in spine BMD (P < 0.001) and a positive correlation between baseline mSASSS and spine BMD was found (r = 0.468, P < 0.01). Inflammatory parameters reduction was observed comparing baseline vs. 6 vs. 12 months (P <0.01). Sclerostin levels progressively increased [baseline (60.5 ± 32.7) vs. 6 months (67.1 ± 31.9) vs. 12 months (72.7 ± 32.3) pmol/L, P <0.001]. At 12 months, the sclerostin levels remained significantly lower in patients compared to controls (72.7 ± 32.3 vs. 96.70 ± 52.85 pmol/L, P = 0.038). Moreover, sclerostin serum levels at 12 months were lower in the 10 patients with high C reactive protein (CRP) (≥ 5 mg/l) compared to the other 20 patients with normal CRP (P = 0.004). Of note, these 10 patients with persistent inflammation also had lower sclerostin serum levels at baseline compared to the other patients (P = 0.023). Univariate logistic regression analysis demonstrated that AS patients with lower sclerostin serum levels had an increased risk to have high CRP at 12 months (odds ratio = 7.43, 95% CI 1.23 to 45.01, P = 0.020) than those with higher sclerostin values. Conclusions Persistent low sclerostin levels may underlie continuous inflammation in AS patients under anti-TNF therapy.
Resumo:
Abstract Background Cell adhesion molecules (CAMs) are essential for maintaining tissue integrity by regulating intercellular and cell to extracellular matrix interactions. Cadherins and catenins are CAMs that are located on the cell membrane and are important for adherens junction (AJ) function. This study aims to verify if hypercholesterolemic diet (HCD) or bladder outlet obstruction (BOO) promotes structural bladder wall modifications specific to alterations in the expression of cadherins and catenins in detrusor muscle cells. Methods Forty-five 4-week-old female Wistar rats were divided into the following three groups: group 1 was a control group that was fed a normal diet (ND); group 2 was the BOO model and was fed a ND; and group 3 was a control group that was fed a HCD (1.25% cholesterol). Initially, serum cholesterol, LDL cholesterol and body weight were determined. Four weeks later, groups 1 and 3 underwent a sham operation; whereas group 2 underwent a partial BOO procedure that included a suture tied around the urethra. Six weeks later, all rats had their bladders removed, and previous exams were repeated. The expression levels of N-, P-, and E-cadherin, cadherin-11 and alpha-, beta- and gamma-catenins were evaluated by immunohistochemistry with a semiquantitative analysis. Results Wistar rats fed a HCD (group 3) exhibited a significant increase in LDL cholesterol levels (p=0.041) and body weight (p=0.017) when compared to both groups that were fed a normal diet in a ten-week period. We found higher β- and γ-catenin expression in groups 2 and 3 when compared to group 1 (p = 0.042 and p = 0.044, respectively). We also observed Cadherin-11 overexpression in group 3 when compared to groups 1 and 2 (p = 0.002). Conclusions A HCD in Wistar rats promoted, in addition to higher body weight gain and increased serum LDL cholesterol levels, overexpression of β- and γ-catenin in the detrusor muscle cells. Similar finding was observed in the BOO group. Higher Cadherin-11 expression was observed only in the HCD-treated rats. These findings may be associated with bladder dysfunctions that occur under such situations.
Resumo:
[EN]A predictive solar radiation numerical model is presented. Starting from the works of, a solar radiation numerical model is developed considering the terrain surface through 2-D adaptive meshes of triangles which are constructed using a refinement/derefinement procedure in accordance with the variations of terrain surface and albedo. The effect of shadows is considered in each time step. Solar radiation is first computed for clear-sky (CS) conditions and then, real-sky values are computed daily in terms of the CS index computed using all the observational data which are available for each day at several points of the studied zone…
Resumo:
Recently in most of the industrial automation process an ever increasing degree of automation has been observed. This increasing is motivated by the higher requirement of systems with great performance in terms of quality of products/services generated, productivity, efficiency and low costs in the design, realization and maintenance. This trend in the growth of complex automation systems is rapidly spreading over automated manufacturing systems (AMS), where the integration of the mechanical and electronic technology, typical of the Mechatronics, is merging with other technologies such as Informatics and the communication networks. An AMS is a very complex system that can be thought constituted by a set of flexible working stations, one or more transportation systems. To understand how this machine are important in our society let considerate that every day most of us use bottles of water or soda, buy product in box like food or cigarets and so on. Another important consideration from its complexity derive from the fact that the the consortium of machine producers has estimated around 350 types of manufacturing machine. A large number of manufacturing machine industry are presented in Italy and notably packaging machine industry,in particular a great concentration of this kind of industry is located in Bologna area; for this reason the Bologna area is called “packaging valley”. Usually, the various parts of the AMS interact among them in a concurrent and asynchronous way, and coordinate the parts of the machine to obtain a desiderated overall behaviour is an hard task. Often, this is the case in large scale systems, organized in a modular and distributed manner. Even if the success of a modern AMS from a functional and behavioural point of view is still to attribute to the design choices operated in the definition of the mechanical structure and electrical electronic architecture, the system that governs the control of the plant is becoming crucial, because of the large number of duties associated to it. Apart from the activity inherent to the automation of themachine cycles, the supervisory system is called to perform other main functions such as: emulating the behaviour of traditional mechanical members thus allowing a drastic constructive simplification of the machine and a crucial functional flexibility; dynamically adapting the control strategies according to the different productive needs and to the different operational scenarios; obtaining a high quality of the final product through the verification of the correctness of the processing; addressing the operator devoted to themachine to promptly and carefully take the actions devoted to establish or restore the optimal operating conditions; managing in real time information on diagnostics, as a support of the maintenance operations of the machine. The kind of facilities that designers can directly find on themarket, in terms of software component libraries provides in fact an adequate support as regard the implementation of either top-level or bottom-level functionalities, typically pertaining to the domains of user-friendly HMIs, closed-loop regulation and motion control, fieldbus-based interconnection of remote smart devices. What is still lacking is a reference framework comprising a comprehensive set of highly reusable logic control components that, focussing on the cross-cutting functionalities characterizing the automation domain, may help the designers in the process of modelling and structuring their applications according to the specific needs. Historically, the design and verification process for complex automated industrial systems is performed in empirical way, without a clear distinction between functional and technological-implementation concepts and without a systematic method to organically deal with the complete system. Traditionally, in the field of analog and digital control design and verification through formal and simulation tools have been adopted since a long time ago, at least for multivariable and/or nonlinear controllers for complex time-driven dynamics as in the fields of vehicles, aircrafts, robots, electric drives and complex power electronics equipments. Moving to the field of logic control, typical for industrial manufacturing automation, the design and verification process is approached in a completely different way, usually very “unstructured”. No clear distinction between functions and implementations, between functional architectures and technological architectures and platforms is considered. Probably this difference is due to the different “dynamical framework”of logic control with respect to analog/digital control. As a matter of facts, in logic control discrete-events dynamics replace time-driven dynamics; hence most of the formal and mathematical tools of analog/digital control cannot be directly migrated to logic control to enlighten the distinction between functions and implementations. In addition, in the common view of application technicians, logic control design is strictly connected to the adopted implementation technology (relays in the past, software nowadays), leading again to a deep confusion among functional view and technological view. In Industrial automation software engineering, concepts as modularity, encapsulation, composability and reusability are strongly emphasized and profitably realized in the so-calledobject-oriented methodologies. Industrial automation is receiving lately this approach, as testified by some IEC standards IEC 611313, IEC 61499 which have been considered in commercial products only recently. On the other hand, in the scientific and technical literature many contributions have been already proposed to establish a suitable modelling framework for industrial automation. During last years it was possible to note a considerable growth in the exploitation of innovative concepts and technologies from ICT world in industrial automation systems. For what concerns the logic control design, Model Based Design (MBD) is being imported in industrial automation from software engineering field. Another key-point in industrial automated systems is the growth of requirements in terms of availability, reliability and safety for technological systems. In other words, the control system should not only deal with the nominal behaviour, but should also deal with other important duties, such as diagnosis and faults isolations, recovery and safety management. Indeed, together with high performance, in complex systems fault occurrences increase. This is a consequence of the fact that, as it typically occurs in reliable mechatronic systems, in complex systems such as AMS, together with reliable mechanical elements, an increasing number of electronic devices are also present, that are more vulnerable by their own nature. The diagnosis problem and the faults isolation in a generic dynamical system consists in the design of an elaboration unit that, appropriately processing the inputs and outputs of the dynamical system, is also capable of detecting incipient faults on the plant devices, reconfiguring the control system so as to guarantee satisfactory performance. The designer should be able to formally verify the product, certifying that, in its final implementation, it will perform itsrequired function guarantying the desired level of reliability and safety; the next step is that of preventing faults and eventually reconfiguring the control system so that faults are tolerated. On this topic an important improvement to formal verification of logic control, fault diagnosis and fault tolerant control results derive from Discrete Event Systems theory. The aimof this work is to define a design pattern and a control architecture to help the designer of control logic in industrial automated systems. The work starts with a brief discussion on main characteristics and description of industrial automated systems on Chapter 1. In Chapter 2 a survey on the state of the software engineering paradigm applied to industrial automation is discussed. Chapter 3 presentes a architecture for industrial automated systems based on the new concept of Generalized Actuator showing its benefits, while in Chapter 4 this architecture is refined using a novel entity, the Generalized Device in order to have a better reusability and modularity of the control logic. In Chapter 5 a new approach will be present based on Discrete Event Systems for the problemof software formal verification and an active fault tolerant control architecture using online diagnostic. Finally conclusive remarks and some ideas on new directions to explore are given. In Appendix A are briefly reported some concepts and results about Discrete Event Systems which should help the reader in understanding some crucial points in chapter 5; while in Appendix B an overview on the experimental testbed of the Laboratory of Automation of University of Bologna, is reported to validated the approach presented in chapter 3, chapter 4 and chapter 5. In Appendix C some components model used in chapter 5 for formal verification are reported.
Resumo:
Spatial prediction of hourly rainfall via radar calibration is addressed. The change of support problem (COSP), arising when the spatial supports of different data sources do not coincide, is faced in a non-Gaussian setting; in fact, hourly rainfall in Emilia-Romagna region, in Italy, is characterized by abundance of zero values and right-skeweness of the distribution of positive amounts. Rain gauge direct measurements on sparsely distributed locations and hourly cumulated radar grids are provided by the ARPA-SIMC Emilia-Romagna. We propose a three-stage Bayesian hierarchical model for radar calibration, exploiting rain gauges as reference measure. Rain probability and amounts are modeled via linear relationships with radar in the log scale; spatial correlated Gaussian effects capture the residual information. We employ a probit link for rainfall probability and Gamma distribution for rainfall positive amounts; the two steps are joined via a two-part semicontinuous model. Three model specifications differently addressing COSP are presented; in particular, a stochastic weighting of all radar pixels, driven by a latent Gaussian process defined on the grid, is employed. Estimation is performed via MCMC procedures implemented in C, linked to R software. Communication and evaluation of probabilistic, point and interval predictions is investigated. A non-randomized PIT histogram is proposed for correctly assessing calibration and coverage of two-part semicontinuous models. Predictions obtained with the different model specifications are evaluated via graphical tools (Reliability Plot, Sharpness Histogram, PIT Histogram, Brier Score Plot and Quantile Decomposition Plot), proper scoring rules (Brier Score, Continuous Rank Probability Score) and consistent scoring functions (Root Mean Square Error and Mean Absolute Error addressing the predictive mean and median, respectively). Calibration is reached and the inclusion of neighbouring information slightly improves predictions. All specifications outperform a benchmark model with incorrelated effects, confirming the relevance of spatial correlation for modeling rainfall probability and accumulation.
Resumo:
This thesis is focused on Smart Grid applications in medium voltage distribution networks. For the development of new applications it appears useful the availability of simulation tools able to model dynamic behavior of both the power system and the communication network. Such a co-simulation environment would allow the assessment of the feasibility of using a given network technology to support communication-based Smart Grid control schemes on an existing segment of the electrical grid and to determine the range of control schemes that different communications technologies can support. For this reason, is presented a co-simulation platform that has been built by linking the Electromagnetic Transients Program Simulator (EMTP v3.0) with a Telecommunication Network Simulator (OPNET-Riverbed v18.0). The simulator is used to design and analyze a coordinate use of Distributed Energy Resources (DERs) for the voltage/var control (VVC) in distribution network. This thesis is focused control structure based on the use of phase measurement units (PMUs). In order to limit the required reinforcements of the communication infrastructures currently adopted by Distribution Network Operators (DNOs), the study is focused on leader-less MAS schemes that do not assign special coordinating rules to specific agents. Leader-less MAS are expected to produce more uniform communication traffic than centralized approaches that include a moderator agent. Moreover, leader-less MAS are expected to be less affected by limitations and constraint of some communication links. The developed co-simulator has allowed the definition of specific countermeasures against the limitations of the communication network, with particular reference to the latency and loss and information, for both the case of wired and wireless communication networks. Moreover, the co-simulation platform has bee also coupled with a mobility simulator in order to study specific countermeasures against the negative effects on the medium voltage/current distribution network caused by the concurrent connection of electric vehicles.
Resumo:
This is the first part of a study investigating a model-based transient calibration process for diesel engines. The motivation is to populate hundreds of parameters (which can be calibrated) in a methodical and optimum manner by using model-based optimization in conjunction with the manual process so that, relative to the manual process used by itself, a significant improvement in transient emissions and fuel consumption and a sizable reduction in calibration time and test cell requirements is achieved. Empirical transient modelling and optimization has been addressed in the second part of this work, while the required data for model training and generalization are the focus of the current work. Transient and steady-state data from a turbocharged multicylinder diesel engine have been examined from a model training perspective. A single-cylinder engine with external air-handling has been used to expand the steady-state data to encompass transient parameter space. Based on comparative model performance and differences in the non-parametric space, primarily driven by a high engine difference between exhaust and intake manifold pressures (ΔP) during transients, it has been recommended that transient emission models should be trained with transient training data. It has been shown that electronic control module (ECM) estimates of transient charge flow and the exhaust gas recirculation (EGR) fraction cannot be accurate at the high engine ΔP frequently encountered during transient operation, and that such estimates do not account for cylinder-to-cylinder variation. The effects of high engine ΔP must therefore be incorporated empirically by using transient data generated from a spectrum of transient calibrations. Specific recommendations on how to choose such calibrations, how many data to acquire, and how to specify transient segments for data acquisition have been made. Methods to process transient data to account for transport delays and sensor lags have been developed. The processed data have then been visualized using statistical means to understand transient emission formation. Two modes of transient opacity formation have been observed and described. The first mode is driven by high engine ΔP and low fresh air flowrates, while the second mode is driven by high engine ΔP and high EGR flowrates. The EGR fraction is inaccurately estimated at both modes, while EGR distribution has been shown to be present but unaccounted for by the ECM. The two modes and associated phenomena are essential to understanding why transient emission models are calibration dependent and furthermore how to choose training data that will result in good model generalization.