976 resultados para Constrained network mapping
Resumo:
Value network has been studied greatly in the academic research, but a tool for value network mapping is missing. The objective of this study was to design a tool (process) for value network mapping in cross-sector collaboration. Furthermore, the study addressed a future perspective of collaboration, aiming to map the value network potential. During the study was investigated and pondered how to get the full potential of collaboration, by creating new value in collaboration process. These actions are parts of mapping process proposed in the study. The implementation and testing of the mapping process were realized through a case study of cross-sector collaboration in welfare services for elderly in the Eastern Finland. Key representatives in elderly care from public, private and third sectors were interviewed and a workshop with experts from every sector was also conducted in this regard. The value network mapping process designed in this study consists of specific steps that help managers and experts to understand how to get a complex value network map and how to enhance it. Furthermore, it make easier the understanding of how new value can be created in collaboration process. The map can be used in order to motivate participants to be engaged with responsibility in collaboration and to be fully committed in their interactions. It can be also used as a motivator tool for those organizations that intend to engage in collaboration process. Additionally, value network map is a starting point in many value network analyses. Furthermore, the enhanced value network map can be used as a performance measurement tool in cross-sector collaboration.
Resumo:
A neural network enhanced self-tuning controller is presented, which combines the attributes of neural network mapping with a generalised minimum variance self-tuning control (STC) strategy. In this way the controller can deal with nonlinear plants, which exhibit features such as uncertainties, nonminimum phase behaviour, coupling effects and may have unmodelled dynamics, and whose nonlinearities are assumed to be globally bounded. The unknown nonlinear plants to be controlled are approximated by an equivalent model composed of a simple linear submodel plus a nonlinear submodel. A generalised recursive least squares algorithm is used to identify the linear submodel and a layered neural network is used to detect the unknown nonlinear submodel in which the weights are updated based on the error between the plant output and the output from the linear submodel. The procedure for controller design is based on the equivalent model therefore the nonlinear submodel is naturally accommodated within the control law. Two simulation studies are provided to demonstrate the effectiveness of the control algorithm.
Resumo:
Network virtualization is a promising technique for building the Internet of the future since it enables the low cost introduction of new features into network elements. An open issue in such virtualization is how to effect an efficient mapping of virtual network elements onto those of the existing physical network, also called the substrate network. Mapping is an NP-hard problem and existing solutions ignore various real network characteristics in order to solve the problem in a reasonable time frame. This paper introduces new algorithms to solve this problem based on 0–1 integer linear programming, algorithms based on a whole new set of network parameters not taken into account by previous proposals. Approximative algorithms proposed here allow the mapping of virtual networks on large network substrates. Simulation experiments give evidence of the efficiency of the proposed algorithms.
Resumo:
The main theme of research of this project concerns the study of neutral networks to control uncertain and non-linear control systems. This involves the control of continuous time, discrete time, hybrid and stochastic systems with input, state or output constraints by ensuring good performances. A great part of this project is devoted to the opening of frontiers between several mathematical and engineering approaches in order to tackle complex but very common non-linear control problems. The objectives are: 1. Design and develop procedures for neutral network enhanced self-tuning adaptive non-linear control systems; 2. To design, as a general procedure, neural network generalised minimum variance self-tuning controller for non-linear dynamic plants (Integration of neural network mapping with generalised minimum variance self-tuning controller strategies); 3. To develop a software package to evaluate control system performances using Matlab, Simulink and Neural Network toolbox. An adaptive control algorithm utilising a recurrent network as a model of a partial unknown non-linear plant with unmeasurable state is proposed. Appropriately, it appears that structured recurrent neural networks can provide conveniently parameterised dynamic models for many non-linear systems for use in adaptive control. Properties of static neural networks, which enabled successful design of stable adaptive control in the state feedback case, are also identified. A survey of the existing results is presented which puts them in a systematic framework showing their relation to classical self-tuning adaptive control application of neural control to a SISO/MIMO control. Simulation results demonstrate that the self-tuning design methods may be practically applicable to a reasonably large class of unknown linear and non-linear dynamic control systems.
Resumo:
A tese tem como objectivo principal a criação de um modelo equivalente eléctrico da rede de nervuras de algumas folhas vegetais e analisar o seu comportamento a estímulos eléctricos, analisando-se também a respectiva resposta em frequência. A motivação desta tese passa pela observação dos sistemas existentes na natureza. Neste caso, as folhas vegetais e analisar se são sistemas de ordem fraccionária ou não. Para a sua elaboração, fez-se uma breve abordagem à estrutura das plantas, sob o ponto de vista da botânica e elaborou-se um método de fotografia das amostras, mapeamento da rede de nervuras e medição dos segmentos que compõem essa mesma rede. A tese termina com um capítulo de resultados experimentais.
Resumo:
Tutkimus kartoitetaan tulevaisuuden älykotiliiketoiminnan palvelutuottaja-verkostoa. Lisäksi delphi-prosessia seuraten, pyritään kyselyn kautta vali-doimaan tulevaisuudesta tehty skenaario. Kysely tehtiin 42 yrityksen stra-tegiasta vastaavalle johtajalle. Selvitettäviä asioita olivat skenaarion liike-toiminnallinen kiinnostavuus, toteutumiskelpoisuus ja yrityksen siinä ha-luama rooli. Lisäksi selvitettiin mm. yritysten yhteistyökyvykkyyttä, innova-tiivisuutta, asiakassuuntautuneisuutta ja teknologinen valmiutta. Pääpaino oli selvittää verkoston muodostaja eli johtaja yritys. Kysely validoi kaikkien vastanneiden osalta tulevan skenaarion. Tulosten perusteella pystyttiin erottamaan 12 potentiaalista johtajaa. Nämä erottui-vat kyselyn kaikilla osa-alueilla parempina kuin muut yritykset. Potentiaali-set johtajat näkivät tulevaisuuden optimistisempana kuin muut ja lisäksi ne harjoittavat jo nyt liiketoimintaa, joka on lähellä kuvattua älykotipalveluver-kostoa. Tuloksia voidaan hyödyntää muodostettaessa palveluverkostoa uusille markkinoille. Kuvatun mallin toteutuminen vaatii kuitenkin julkisen sektorin palvelutoiminnan uudistusta, sillä se sisältää useita rinnakkaisia prosesseja julkisen terveydenhuollon kanssa.
Resumo:
Cross-sector collaboration and partnerships have become an emerging and desired strategy in addressing huge social and environmental challenges. Despite its popularity, cross-sector collaboration management has proven to be very challenging. Even though cross-sector collaboration and partnership management have been widely studied and discussed in recent years, their effectiveness as well as their ability to create value with respect to the problems they address has remained very challenging. There is little or no evidence of their ability to create value. Regarding all these challenges, this study aims to explore how to manage cross-sector collaborations and partnerships to be able to improve their effectiveness and to create more value for all partners involved in collaboration as well as for customers. The thesis is divided into two parts. The first part comprises an overview of relevant literature (including strategic management, value networks and value creation theories), followed by presenting the results of the whole thesis and the contribution made by the study. The second part consists of six research publications, including both quantitative and qualitative studies. The chosen research strategy is triangulation, as the study includes four types of triangulation: (1) theoretical triangulation, (2) methodological triangulation, (3) data triangulation and (4) researcher triangulation. Two publications represent conceptual development, which are based on secondary data research. One publication is a quantitative study, carried out through a survey. The other three publications represent qualitative studies, based on case studies, where data was collected through interviews and workshops, with participation of managers from all three sectors: public, private and the third (nonprofit). The study consolidates the field of “strategic management of value networks,” which is proposed to be applied in the context of cross-sector collaboration and partnerships, with the aim of increasing their effectiveness and the process of value creation. Furthermore, the study proposes a first definition for the strategic management of value networks. The study also proposes and develops two strategy tools that are recommended to be used for the strategic management of value networks in cross-sector collaboration and partnerships. Taking a step forward, the study implements the strategy tools in practice, aiming to show and to demonstrate how new value can be created by using the developed strategy tools for the strategic management of value networks. This study makes four main contributions. (1) First, it brings a theoretical contribution by providing new insights and consolidating the field of strategic management of value networks, also proposing a first definition for the strategic management of value networks. (2) Second, the study makes a methodical contribution by proposing and developing two strategy tools for value networks of cross-sector collaboration: (a) value network mapping, a method that allows us to assess the current and the potential value network and (b) the Value Network Scorecard, a method of performance measurement and performance prediction in cross-sector collaboration. (3) Third, the study has managerial implications, offering new solutions and empirical evidence on how to increase the effectiveness of cross-sector collaboration and also allow managers to understand how new value can be created in cross-sector partnerships and how to get the full potential of collaboration. (4) And fourth, the study also has practical implications, allowing managers to understand how to use in practice the strategy tools developed in this study, providing discussions on the limitations regarding the proposed tools as well as general limitations involved in the study.
Resumo:
Organizations have been facing several challenges to survive in a world in constant transformation. In light of that, new management models need to be incorporated to the organizational dynamics in order to achieve competitive advantages. The inter-firm networks establish a particular form means of cooperation as a fundamental element, as there can be ties of diverse nature to justify the relationships. These are interorganizations or inter-firms, formed by people who maintain the relationships to improve the overall performance of the network and of the participants. Thus, the Education Institutions, considered as organizations, which have the objective to disseminate knowledge and form professionals who contribute to the growth of a nation, also make use of this type of arrangement. Considering the raised questions, this article aims to demonstrate the applicability of a business network mapping technique at Higher Education Institutions. It also presents the possibility to perform the comparative analysis of the intensity of the relationship observed in a private and in a public institution, in the State of São Paulo. By applying the comparative analysis, this study also presents the relationship among the people involved at strategic and tactical levels of each institution, taking into account each one s peculiarities.
Resumo:
Chronic stress is associated with hippocampal atrophy and cognitive dysfunction. This study investigates how long-lasting administration of corticosterone as a mimic of experimentally induced stress affects psychometric performance and the expression of the phosphatidylethanolamine binding protein (PEBP1) in the adult hippocampus of one-year-old male rats. Psychometric investigations were conducted in rats before and after corticosterone treatment using a holeboard test system. Rats were randomly attributed to 2 groups (n = 7) for daily subcutaneous injection of either 26.8 mg/kg body weight corticosterone or sesame oil (vehicle control). Treatment was continued for 60 days, followed by cognitive retesting in the holeboard system. For protein analysis, the hippocampal proteome was separated by 2D electrophoresis (2DE) followed by image processing, statistical analysis, protein identification via peptide mass fingerprinting and gel matching and subsequent functional network mapping and molecular pathway analysis. Differential expression of PEBP1 was additionally quantified by Western blot analysis. Results show that chronic corticosterone significantly decreased rat hippocampal PEBP1 expression and induced a working and reference memory dysfunction. From this, we derive the preliminary hypothesis that PEBP1 may be a novel molecular mediator influencing cognitive integrity during chronic corticosterone exposure in rat hippocampus.
Resumo:
El aumento de las capacidades de interconexión de dispositivos de todo tipo está suponiendo una revolución en el campo de la prestación de servicios, tanto en la cantidad como en la variedad. Esta evolución ha puesto de manifiesto la necesidad de abordar un desarrollo tecnológico sin precedentes, donde la previsión de dispositivos interconectados e interoperando entre sí y con las personas alcanza cifras del orden de los millardos. Esta idea de un mundo de cosas interconectadas ha dado lugar a una visión que se ha dado en llamar Internet de las Cosas. Un mundo donde las cosas de cualquier tipo pueden interactuar con otras cosas, incluyendo las que forman parte de redes con recurso limitados. Y esto además conduce a la creación de servicios compuestos que superan a la suma de las partes. Además de la relevancia tecnológica, esta nueva visión enlaza con la de la Ciudad del Futuro. Un concepto que recurre a la convergencia de la energía, el transporte y las tecnologías de la información y las comunicaciones para definir una forma mediante la que lograr el crecimiento sostenible y competitivo, mejorando así la calidad de vida y abriendo el gobierno de las ciudades a la participación ciudadana. En la línea de desarrollo que permite avanzar hacia la consecución de tales objetivos, este Proyecto Fin de Carrera propone una forma de virtualizar los servicios ofrecidos por la diversidad de dispositivos que van adquiriendo la capacidad de interoperar en una red. Para ello se apoya en el uso de una capa de intermediación orientada a servicios, nSOM, desarrollada en la EUITT. Sobre esta arquitectura se proponen como objetivos el diseño y desarrollo de una pasarela de servicios que haga accesibles desde la web los recursos ofrecidos en una red de sensores; el diseño y desarrollo de un registro de dispositivos y servicios en concordancia a la propuesta de arquitectura de referencia para Internet de las Cosas; y el estudio y diseño de un marco para la composición de servicios orquestados en redes de recursos limitados. Para alcanzar estos objetivos primero se abordará un estudio del estado del arte donde se profundizará en el conocimiento de la las tecnologías para la interoperatividad entre cosas, abordando los principios de las redes inalámbricas de sensores y actuadores, las arquitecturas para las comunicaciones Máquina a Máquina e Internet de las Cosas, y la visión de la Web de las Cosas. Seguidamente se tratarán las tecnologías de red y de servicios de interés, para finalizar con un breve repaso a las tecnologías para la composición de servicios. Le seguirá una descripción detallada de la arquitectura nSOM y del diseño propuesto para este proyecto. Finalmente se propondrá un escenario sobre el que se llevarán a cabo diferentes pruebas de validación. ABSTRACT. The increasing of the capabilities of all kind of devices is causing a revolution in the field of the provision of services, both in quantity and in diversity. This situation has highlighted the need to address unprecedented technological development, where the forecast of interconnected and interoperable devices between them and human beings reaches the order of billions. And these numbers go further when the connectivity of constrained networks is taken into account. This idea of an interconnected world of things has led to a vision that has been called "The Internet of Things". It’s a vision of a world where things of any kind can interact with other things, even those in the domain of a constrained network. This also leads to the creation of new composed services that exceed the sum of the parts. Besides the technological interest, this new vision relates with the one from the Smart City. A concept that uses the convergence of the energy, the transport, and the information and communication technologies to define a way to achieve sustainable and competitive growth, improving the quality of life, and opening the governance of the cities to the participation. In the development pathway to reach these goals, this Final Degree Dissertation proposes a way for the virtualization of the services offered by the variety of devices that are reaching the ability to interoperate in a network. For this it is supported by a service oriented middleware called nSOM that has been developed at EUITT. Using this architecture the goals proposed for this project are the design and development of a service gateway that makes available the resources of a sensor network through a web interface; the design and development of a Device & Service Registry according to the reference architecture proposal for the Internet of Things; and the study and design of a composition framework for orchestrated services in constrained networks. To achieve these goals this dissertation begins with a State of the Art study where the background knowledge about the technologies in use for the interoperation of things will be settled. At first it starts talking about Wireless Sensor and Actuator Networks, the architectures for Machine-to-Machine communication and Internet of Things, and also the concepts for the Web of Things vision. Next the related network and services technologies are explored, ending with a brief review of service composition technologies. Then will follow a detailed description of the nSOM architecture, and also of the proposed design for this project. Finally a scenario will be proposed where a series of validation tests will be conducted.
Resumo:
El presente trabajo tiene como objetivo general el análisis de las técnicas de diseño y optimización de redes topográficas, observadas mediante topografía convencional (no satelital) el desarrollo e implementación de un sistema informático capaz de ayudar a la definición de la geometría más fiable y precisa, en función de la orografía del terreno donde se tenga que ubicar. En primer lugar se realizará un estudio de la metodología del ajuste mediante mínimos cuadrados y la propagación de varianzas, para posteriormente analizar su dependencia de la geometría que adopte la red. Será imprescindible determinar la independencia de la matriz de redundancia (R) de las observaciones y su total dependencia de la geometría, así como la influencia de su diagonal principal (rii), números de redundancia, para garantizar la máxima fiabilidad interna de la misma. También se analizará el comportamiento de los números de redundancia (rii) en el diseño de una red topográfica, la variación de dichos valores en función de la geometría, analizando su independencia respecto de las observaciones así como los diferentes niveles de diseño en función de los parámetros y datos conocidos. Ha de señalarse que la optimización de la red, con arreglo a los criterios expuestos, está sujeta a los condicionantes que impone la necesidad de que los vértices sean accesibles, y además sean visibles entre sí, aquellos relacionados por observaciones, situaciones que dependen esencialmente del relieve del terreno y de los obstáculos naturales o artificiales que puedan existir. Esto implica la necesidad de incluir en el análisis y en el diseño, cuando menos de un modelo digital del terreno (MDT), aunque lo más útil sería la inclusión en el estudio del modelo digital de superficie (MDS), pero esta opción no siempre será posible. Aunque el tratamiento del diseño esté basado en un sistema bidimensional se estudiará la posibilidad de incorporar un modelo digital de superficie (MDS); esto permitirá a la hora de diseñar el emplazamiento de los vértices de la red la viabilidad de las observaciones en función de la orografía y los elementos, tanto naturales como artificiales, que sobre ella estén ubicados. Este sistema proporcionaría, en un principio, un diseño óptimo de una red constreñida, atendiendo a la fiabilidad interna y a la precisión final de sus vértices, teniendo en cuenta la orografía, lo que equivaldría a resolver un planteamiento de diseño en dos dimensiones y media1; siempre y cuando se dispusiera de un modelo digital de superficie o del terreno. Dado que la disponibilidad de obtener de manera libre el MDS de las zonas de interés del proyecto, hoy en día es costoso2, se planteará la posibilidad de conjuntar, para el estudio del diseño de la red, de un modelo digital del terreno. Las actividades a desarrollar en el trabajo de esta tesis se describen en esta memoria y se enmarcan dentro de la investigación para la que se plantean los siguientes objetivos globales: 1. Establecer un modelo matemático del proceso de observación de una red topográfica, atendiendo a todos los factores que intervienen en el mismo y a su influencia sobre las estimaciones de las incógnitas que se obtienen como resultado del ajuste de las observaciones. 2. Desarrollar un sistema que permita optimizar una red topográfica en sus resultados, aplicando técnicas de diseño y simulación sobre el modelo anterior. 3. Presentar una formulación explícita y rigurosa de los parámetros que valoran la fiabilidad de una red topográfica y de sus relaciones con el diseño de la misma. El logro de este objetivo se basa, además de en la búsqueda y revisión de las fuentes, en una intensa labor de unificación de notaciones y de construcción de pasos intermedios en los desarrollos matemáticos. 4. Elaborar una visión conjunta de la influencia del diseño de una red, en los seis siguientes factores (precisiones a posteriori, fiabilidad de las observaciones, naturaleza y viabilidad de las mismas, instrumental y metodología de estacionamiento) como criterios de optimización, con la finalidad de enmarcar el tema concreto que aquí se aborda. 5. Elaborar y programar los algoritmos necesarios para poder desarrollar una aplicación que sea capaz de contemplar las variables planteadas en el apartado anterior en el problema del diseño y simulación de redes topográficas, contemplando el modelo digital de superficie. Podrían considerarse como objetivos secundarios, los siguientes apartados: Desarrollar los algoritmos necesarios para interrelacionar el modelo digital del terreno con los propios del diseño. Implementar en la aplicación informática la posibilidad de variación, por parte del usuario, de los criterios de cobertura de los parámetros (distribución normal o t de Student), así como los grados de fiabilidad de los mismos ABSTRACT The overall purpose of this work is the analysis of the techniques of design and optimization for geodetic networks, measured with conventional survey methods (not satellite), the development and implementation of a computational system capable to help on the definition of the most liable and accurate geometry, depending on the land orography where the network has to be located. First of all, a study of the methodology by least squares adjustment and propagation of variances will be held; then, subsequently, analyze its dependency of the geometry that the network will take. It will be essential to determine the independency of redundancy matrix (R) from the observations and its absolute dependency from the network geometry, as well as the influence of the diagonal terms of the R matrix (rii), redundancy numbers, in order to ensure maximum re liability of the network. It will also be analyzed first the behavior of redundancy numbers (rii) in surveying network design, then the variation of these values depending on the geometry with the analysis of its independency from the observations, and finally the different design levels depending on parameters and known data. It should be stated that network optimization, according to exposed criteria, is subject to the accessibility of the network points. In addition, common visibility among network points, which of them are connected with observations, has to be considered. All these situations depends essentially on the terrain relief and the natural or artificial obstacles that should exist. Therefore, it is necessary to include, at least, a digital terrain model (DTM), and better a digital surface model (DSM), not always available. Although design treatment is based on a bidimensional system, the possibility of incorporating a digital surface model (DSM) will be studied; this will allow evaluating the observations feasibility based on the terrain and the elements, both natural and artificial, which are located on it, when selecting network point locations. This system would provide, at first, an optimal design of a constrained network, considering both the internal reliability and the accuracy of its points (including the relief). This approach would amount to solving a “two and a half dimensional”3 design, if a digital surface model is available. As the availability of free DSM4 of the areas of interest of the project today is expensive, the possibility of combining a digital terrain model will arise. The activities to be developed on this PhD thesis are described in this document and are part of the research for which the following overall objectives are posed: 1. To establish a mathematical model for the process of observation of a survey network, considering all the factors involved and its influence on the estimates of the unknowns that are obtained as a result of the observations adjustment. 2. To develop a system to optimize a survey network results, applying design and simulation techniques on the previous model. 3. To present an explicit and rigorous formulation of parameters which assess the reliability of a survey network and its relations with the design. The achievement of this objective is based, besides on the search and review of sources, in an intense work of unification of notation and construction of intermediate steps in the mathematical developments. 4. To develop an overview of the influence on the network design of six major factors (posterior accuracy, observations reliability, viability of observations, instruments and station methodology) as optimization criteria, in order to define the subject approached on this document. 5. To elaborate and program the algorithms needed to develop an application software capable of considering the variables proposed in the previous section, on the problem of design and simulation of surveying networks, considering the digital surface model. It could be considered as secondary objectives, the following paragraphs: To develop the necessary algorithms to interrelate the digital terrain model with the design ones. To implement in the software application the possibility of variation of the coverage criteria parameters (normal distribution or Student t test) and therefore its degree of reliability.
Resumo:
The performance of feed-forward neural networks in real applications can be often be improved significantly if use is made of a-priori information. For interpolation problems this prior knowledge frequently includes smoothness requirements on the network mapping, and can be imposed by the addition to the error function of suitable regularization terms. The new error function, however, now depends on the derivatives of the network mapping, and so the standard back-propagation algorithm cannot be applied. In this paper, we derive a computationally efficient learning algorithm, for a feed-forward network of arbitrary topology, which can be used to minimize the new error function. Networks having a single hidden layer, for which the learning algorithm simplifies, are treated as a special case.
Resumo:
It is well known that the addition of noise to the input data of a neural network during training can, in some circumstances, lead to significant improvements in generalization performance. Previous work has shown that such training with noise is equivalent to a form of regularization in which an extra term is added to the error function. However, the regularization term, which involves second derivatives of the error function, is not bounded below, and so can lead to difficulties if used directly in a learning algorithm based on error minimization. In this paper we show that, for the purposes of network training, the regularization term can be reduced to a positive definite form which involves only first derivatives of the network mapping. For a sum-of-squares error function, the regularization term belongs to the class of generalized Tikhonov regularizers. Direct minimization of the regularized error function provides a practical alternative to training with noise.
Resumo:
Design is being performed on an ever-increasing spectrum of complex practices arising in response to emerging markets and technologies, co-design, digital interaction, service design and cultures of innovation. This emerging notion of design has led to an expansive array of collaborative and facilitation skills to demonstrate and share how such methods can shape innovation. The meaning of these design things in practice can't be taken for granted as matters of fact, which raises a key challenge for design to represent its role through the contradictory nature of matters of concern. This paper explores an innovative, object-oriented approach within the field of design research, visually combining an actor-network theory framework with situational analysis, to report on the role of design for fledgling companies in Scotland, established and funded through the knowledge exchange hub Design in Action (DiA). Key findings and visual maps are presented from reflective discussions with actors from a selection of the businesses within DiA's portfolio. The suggestion is that any notions of strategic value, of engendering meaningful change, of sharing the vision of design, through design things, should be grounded in the reflexive interpretations of matters of concern that emerge.
Resumo:
Defining digital humanities might be an endless debate if we stick to the discussion about the boundaries of this concept as an academic "discipline". In an attempt to concretely identify this field and its actors, this paper shows that it is possible to analyse them through Twitter, a social media widely used by this "community of practice". Based on a network analysis of 2,500 users identified as members of this movement, the visualisation of the "who's following who?" graph allows us to highlight the structure of the network's relationships, and identify users whose position is particular. Specifically, we show that linguistic groups are key factors to explain clustering within a network whose characteristics look similar to a small world.