987 resultados para process interface
Resumo:
The motivation for this research initiated from the abrupt rise and fall of minicomputers which were initially used both for industrial automation and business applications due to their significantly lower cost than their predecessors, the mainframes. Later industrial automation developed its own vertically integrated hardware and software to address the application needs of uninterrupted operations, real-time control and resilience to harsh environmental conditions. This has led to the creation of an independent industry, namely industrial automation used in PLC, DCS, SCADA and robot control systems. This industry employs today over 200'000 people in a profitable slow clockspeed context in contrast to the two mainstream computing industries of information technology (IT) focused on business applications and telecommunications focused on communications networks and hand-held devices. Already in 1990s it was foreseen that IT and communication would merge into one Information and communication industry (ICT). The fundamental question of the thesis is: Could industrial automation leverage a common technology platform with the newly formed ICT industry? Computer systems dominated by complex instruction set computers (CISC) were challenged during 1990s with higher performance reduced instruction set computers (RISC). RISC started to evolve parallel to the constant advancement of Moore's law. These developments created the high performance and low energy consumption System-on-Chip architecture (SoC). Unlike to the CISC processors RISC processor architecture is a separate industry from the RISC chip manufacturing industry. It also has several hardware independent software platforms consisting of integrated operating system, development environment, user interface and application market which enables customers to have more choices due to hardware independent real time capable software applications. An architecture disruption merged and the smartphone and tablet market were formed with new rules and new key players in the ICT industry. Today there are more RISC computer systems running Linux (or other Unix variants) than any other computer system. The astonishing rise of SoC based technologies and related software platforms in smartphones created in unit terms the largest installed base ever seen in the history of computers and is now being further extended by tablets. An underlying additional element of this transition is the increasing role of open source technologies both in software and hardware. This has driven the microprocessor based personal computer industry with few dominating closed operating system platforms into a steep decline. A significant factor in this process has been the separation of processor architecture and processor chip production and operating systems and application development platforms merger into integrated software platforms with proprietary application markets. Furthermore the pay-by-click marketing has changed the way applications development is compensated: Three essays on major trends in a slow clockspeed industry: The case of industrial automation 2014 freeware, ad based or licensed - all at a lower price and used by a wider customer base than ever before. Moreover, the concept of software maintenance contract is very remote in the app world. However, as a slow clockspeed industry, industrial automation has remained intact during the disruptions based on SoC and related software platforms in the ICT industries. Industrial automation incumbents continue to supply systems based on vertically integrated systems consisting of proprietary software and proprietary mainly microprocessor based hardware. They enjoy admirable profitability levels on a very narrow customer base due to strong technology-enabled customer lock-in and customers' high risk leverage as their production is dependent on fault-free operation of the industrial automation systems. When will this balance of power be disrupted? The thesis suggests how industrial automation could join the mainstream ICT industry and create an information, communication and automation (ICAT) industry. Lately the Internet of Things (loT) and weightless networks, a new standard leveraging frequency channels earlier occupied by TV broadcasting, have gradually started to change the rigid world of Machine to Machine (M2M) interaction. It is foreseeable that enough momentum will be created that the industrial automation market will in due course face an architecture disruption empowered by these new trends. This thesis examines the current state of industrial automation subject to the competition between the incumbents firstly through a research on cost competitiveness efforts in captive outsourcing of engineering, research and development and secondly researching process re- engineering in the case of complex system global software support. Thirdly we investigate the industry actors', namely customers, incumbents and newcomers, views on the future direction of industrial automation and conclude with our assessments of the possible routes industrial automation could advance taking into account the looming rise of the Internet of Things (loT) and weightless networks. Industrial automation is an industry dominated by a handful of global players each of them focusing on maintaining their own proprietary solutions. The rise of de facto standards like IBM PC, Unix and Linux and SoC leveraged by IBM, Compaq, Dell, HP, ARM, Apple, Google, Samsung and others have created new markets of personal computers, smartphone and tablets and will eventually also impact industrial automation through game changing commoditization and related control point and business model changes. This trend will inevitably continue, but the transition to a commoditized industrial automation will not happen in the near future.
Resumo:
Many assays to evaluate the nature, breadth, and quality of antigen-specific T cell responses are currently applied in human medicine. In most cases, assay-related protocols are developed on an individual laboratory basis, resulting in a large number of different protocols being applied worldwide. Together with the inherent complexity of cellular assays, this leads to unnecessary limitations in the ability to compare results generated across institutions. Over the past few years a number of critical assay parameters have been identified which influence test performance irrespective of protocol, material, and reagents used. Describing these critical factors as an integral part of any published report will both facilitate the comparison of data generated across institutions and lead to improvements in the assays themselves. To this end, the Minimal Information About T Cell Assays (MIATA) project was initiated. The objective of MIATA is to achieve a broad consensus on which T cell assay parameters should be reported in scientific publications and to propose a mechanism for reporting these in a systematic manner. To add maximum value for the scientific community, a step-wise, open, and field-spanning approach has been taken to achieve technical precision, user-friendliness, adequate incorporation of concerns, and high acceptance among peers. Here, we describe the past, present, and future perspectives of the MIATA project. We suggest that the approach taken can be generically applied to projects in which a broad consensus has to be reached among scientists working in fragmented fields, such as immunology. An additional objective of this undertaking is to engage the broader scientific community to comment on MIATA and to become an active participant in the project.
Resumo:
This paper highlights the role of non-functional information when reusing from a component library. We describe a method for selecting appropriate implementations of Ada packages taking non-functional constraints into account; these constraints model the context of reuse. Constraints take the form of queries using an interface description language called NoFun, which is also used to state non-functional information in Ada packages; query results are trees of implementations, following the import relationships between components. We define two different situations when reusing components, depending whether we take the library being searched as closed or extendible. The resulting tree of implementations can be manipulated by the user to solve ambiguities, to state default behaviours, and by the like. As part of the proposal, we face the problem of computing from code the non-functional information that determines the selection process.
Resumo:
In this work we present a simulation of a recognition process with perimeter characterization of a simple plant leaves as a unique discriminating parameter. Data coding allowing for independence of leaves size and orientation may penalize performance recognition for some varieties. Border description sequences are then used, and Principal Component Analysis (PCA) is applied in order to study which is the best number of components for the classification task, implemented by means of a Support Vector Machine (SVM) System. Obtained results are satisfactory, and compared with [4] our system improves the recognition success, diminishing the variance at the same time.
Resumo:
La Universitat de Vic disposa, entre altres equips, d’una cèl·lula flexible de fabricació, del fabricant Festo, que simula un procés de formació de palets amb els productes que es disposen en un magatzem intermedi. Aquesta cèl·lula està composta de quatre estacions de muntatge diferenciades (càrrega de palets, càrrega de plaques, magatzem intermedi i transport). Cada una disposa d'un PLC SIEMENS S7-300 per la seva automatització, i tots aquests es troben interconnectats amb una xarxa industrial Profibus. L'objectiu d'aquest projecte és implantar el sistema SCADA Vijeo Citect pel control i supervisió de l'estació magatzem d'aquesta cèl·lula flexible de fabricació, establint també un intercanvi de dades entre l'SCADA i el Microsoft Access, per poder ser utilitzat per la docència. Aquest projecte s'ha desenvolupat en cinc fases diferents: 1. La primera fase s'ha dedicat a l'automatització pròpiament de l'estació magatzem a partir de l'autòmat programable Siemens S7-300 i complint amb les necessitats plantejades. 2. En la segona fase s'ha programat i establert la comunicació per l'intercanvi de dades (lectura i escriptura) entre el sistema SCADA Vijeo Citect i la base de dades de Microsoft Access. 3. En la tercera fase s'ha elaborat i programat l'entorn gràfic de supervisió i control del procés a partir del sistema SCADA Vijeo Citect. 4. En la quarta fase s'ha instal·lat un OPC Server en el PC i s'ha establert la comunicació entre el PLC i el sistema SCADA. 5. Finalment s'ha anat revisant i depurant les diferents programacions i comunicacions per tal de que el sistema funcioni com a un conjunt.
Resumo:
Agroforestry systems are indicated as an alternative for sugarcane (Saccharum officinarum) cultivation in Piracicaba, SP, Brazil, however there are not many field experiments on plant performance under these conditions in the world. The objective of this work was to assess crop yield and partitioning in a sugarcane-rubber (Hevea brasiliensis) interface in on-farm conditions. The availability of irradiance for the crop along the interface was simulated and its effe ct over sugarcane dry matter production was tested. Crop yield was negatively affected by distance of the trees, but development and sucrose were not affected. Above ground dry matter increased from 16.6 to 51.5 t ha-1 from trees. Partitioning did not have a defined standard, as harvest index increased from 0.85 to 0.93, but specific leaf area was not significant along the transect, ranging from 13.48 to 15.73 m² kg-1. Light is the main factor of competition between the trees and the crop, but the relative importance of below ground interactions increases closer to the trees. Feasibility of the system depends on maturity of the trees and management strategies.
Resumo:
This issue review provides a review of Iowa's expenditure law and recent statutory and session-law changes that have changed the process. This issue review also includes examples from the fiscal year 2013 budget enacted during the 2012 legislative session to help explain how the various aspects of the expenditure limitation law are applied to legislative decision making. At the end of this issue review is a schematic showing the distribution of fiscal year 2012 general fund surplus revenues when the expenditure limitation provisions are applied.
Resumo:
Mouse NK cells express MHC class I-specific inhibitory Ly49 receptors. Since these receptors display distinct ligand specificities and are clonally distributed, their expression generates a diverse NK cell receptor repertoire specific for MHC class I molecules. We have previously found that the Dd (or Dk)-specific Ly49A receptor is usually expressed from a single allele. However, a small fraction of short-term NK cell clones expressed both Ly49A alleles, suggesting that the two Ly49A alleles are independently and randomly expressed. Here we show that the genes for two additional Ly49 receptors (Ly49C and Ly49G2) are also expressed in a (predominantly) mono-allelic fashion. Since single NK cells can co-express multiple Ly49 receptors, we also investigated whether mono-allelic expression from within the tightly linked Ly49 gene cluster is coordinate or independent. Our clonal analysis suggests that the expression of alleles of distinct Ly49 genes is not coordinate. Thus Ly49 alleles are apparently independently and randomly chosen for stable expression, a process that directly restricts the number of Ly49 receptors expressed per single NK cell. We propose that the Ly49 receptor repertoire specific for MHC class I is generated by an allele-specific, stochastic gene expression process that acts on the entire Ly49 gene cluster.
Resumo:
The flow of two immiscible fluids through a porous medium depends on the complex interplay between gravity, capillarity, and viscous forces. The interaction between these forces and the geometry of the medium gives rise to a variety of complex flow regimes that are difficult to describe using continuum models. Although a number of pore-scale models have been employed, a careful investigation of the macroscopic effects of pore-scale processes requires methods based on conservation principles in order to reduce the number of modeling assumptions. In this work we perform direct numerical simulations of drainage by solving Navier-Stokes equations in the pore space and employing the Volume Of Fluid (VOF) method to track the evolution of the fluid-fluid interface. After demonstrating that the method is able to deal with large viscosity contrasts and model the transition from stable flow to viscous fingering, we focus on the macroscopic capillary pressure and we compare different definitions of this quantity under quasi-static and dynamic conditions. We show that the difference between the intrinsic phase-average pressures, which is commonly used as definition of Darcy-scale capillary pressure, is subject to several limitations and it is not accurate in presence of viscous effects or trapping. In contrast, a definition based on the variation of the total surface energy provides an accurate estimate of the macroscopic capillary pressure. This definition, which links the capillary pressure to its physical origin, allows a better separation of viscous effects and does not depend on the presence of trapped fluid clusters.
Resumo:
Recent reports indicate that of the over 25,000 bridges in Iowa, slightly over 7,000 (29%) are either structurally deficient or functionally obsolete. While many of these bridges may be strengthened or rehabilitated, some simply need to be replaced. Before implementing one of these options, one should consider performing a diagnostic load test on the structure to more accurately assess its load carrying capacity. Frequently, diagnostic load tests reveal strength and serviceability characteristics that exceed the predicted codified parameters. Usually, codified parameters are very conservative in predicting lateral load distribution characteristics and the influence of other structural attributes. As a result, the predicted rating factors are typically conservative. In cases where theoretical calculations show a structural deficiency, it may be very beneficial to apply a "tool" that utilizes a more accurate theoretical model which incorporates field-test data. At a minimum, this approach results in more accurate load ratings and many times results in increased rating factors. Bridge Diagnostics, Inc. (BDI) developed hardware and software that are specially designed for performing bridge ratings based on data obtained from physical testing. To evaluate the BDI system, the research team performed diagnostic load tests on seven "typical" bridge structures: three steel-girder bridges with concrete decks, two concrete slab bridges, and two steel-girder bridges with timber decks. In addition, a steel-girder bridge with a concrete deck previously tested and modeled by BDI was investigated for model verification purposes. The tests were performed by attaching strain transducers on the bridges at critical locations to measure strains resulting from truck loading positioned at various locations on the bridge. The field test results were used to develop and validate analytical rating models. Based on the experimental and analytical results, it was determined that bridge tests could be conducted relatively easy, that accurate models could be generated with the BDI software, and that the load ratings, in general, were greater than the ratings, obtained using the codified LFD Method (according to AASHTO Standard Specifications for Highway Bridges).