976 resultados para client-centered design
Resumo:
The article is part of an ongoing theoretical research about the emergence of parametric architecture. It discusses some of the last developments in digital design from the transition of the discourse and operativity by “diagrams” to the theories and processes derived from “parameters”. Centered in Peter Eisenman’s and Patrik Schumacher’s propositions, it is in its horizon to comprehend what relations are established between “diagram” and “parameter” -‐ similarities, complementarities and differences -‐, contributing for the critical contextualization of theoretical movements and design processes in contemporary architecture.
Resumo:
The aim of this Ph.D. project has been the design and characterization of new and more efficient luminescent tools, in particular sensors and labels, for analytical chemistry, medical diagnostics and imaging. Actually both the increasing temporal and spatial resolutions that are demanded by those branches, coupled to a sensitivity that is required to reach the single molecule resolution, can be provided by the wide range of techniques based on luminescence spectroscopy. As far as the development of new chemical sensors is concerned, as chemists we were interested in the preparation of new, efficient, sensing materials. In this context, we kept developing new molecular chemosensors, by exploiting the supramolecular approach, for different classes of analytes. In particular we studied a family of luminescent tetrapodal-hosts based on aminopyridinium units with pyrenyl groups for the detection of anions. These systems exhibited noticeable changes in the photophysical properties, depending on the nature of the anion; in particular, addition of chloride resulted in a conformational change, giving an initial increase in excimeric emission. A good selectivity for dicarboxylic acid was also found. In the search for higher sensitivities, we moved our attention also to systems able to perform amplification effects. In this context we described the metal ion binding properties of three photoactive poly-(arylene ethynylene) co-polymers with different complexing units and we highlighted, for one of them, a ten-fold amplification of the response in case of addition of Zn2+, Cu2+ and Hg2+ ions. In addition, we were able to demonstrate the formation of complexes with Yb3+ an Er3+ and an efficient sensitization of their typical metal centered NIR emission upon excitation of the polymer structure, this feature being of particular interest for their possible applications in optical imaging and in optical amplification for telecommunication purposes. An amplification effect was also observed during this research in silica nanoparticles derivatized with a suitable zinc probe. In this case we were able to prove, for the first time, that nanoparticles can work as “off-on” chemosensors with signal amplification. Fluorescent silica nanoparticles can be thus seen as innovative multicomponent systems in which the organization of photophysically active units gives rise to fruitful collective effects. These precious effects can be exploited for biological imaging, medical diagnostic and therapeutics, as evidenced also by some results reported in this thesis. In particular, the observed amplification effect has been obtained thanks to a suitable organization of molecular probe units onto the surface of the nanoparticles. In the effort of reaching a deeper inside in the mechanisms which lead to the final amplification effects, we also attempted to find a correlation between the synthetic route and the final organization of the active molecules in the silica network, and thus with those mutual interactions between one another which result in the emerging, collective behavior, responsible for the desired signal amplification. In this context, we firstly investigated the process of formation of silica nanoparticles doped with pyrene derivative and we showed that the dyes are not uniformly dispersed inside the silica matrix; thus, core-shell structures can be formed spontaneously in a one step synthesis. Moreover, as far as the design of new labels is concerned, we reported a new synthetic approach to obtain a class of robust, biocompatible silica core-shell nanoparticles able to show a long-term stability. Taking advantage of this new approach we also showed the synthesis and photophysical properties of core-shell NIR absorbing and emitting materials that proved to be very valuable for in-vivo imaging. In general, the dye doped silica nanoparticles prepared in the framework of this project can conjugate unique properties, such as a very high brightness, due to the possibility to include many fluorophores per nanoparticle, high stability, because of the shielding effect of the silica matrix, and, to date, no toxicity, with a simple and low-cost preparation. All these features make these nanostructures suitable to reach the low detection limits that are nowadays required for effective clinical and environmental applications, fulfilling in this way the initial expectations of this research project.
Resumo:
Broad consensus has been reached within the Education and Cognitive Psychology research communities on the need to center the learning process on experimentation and concrete application of knowledge, rather than on a bare transfer of notions. Several advantages arise from this educational approach, ranging from the reinforce of students learning, to the increased opportunity for a student to gain greater insight into the studied topics, up to the possibility for learners to acquire practical skills and long-lasting proficiency. This is especially true in Engineering education, where integrating conceptual knowledge and practical skills assumes a strategic importance. In this scenario, learners are called to play a primary role. They are actively involved in the construction of their own knowledge, instead of passively receiving it. As a result, traditional, teacher-centered learning environments should be replaced by novel learner-centered solutions. Information and Communication Technologies enable the development of innovative solutions that provide suitable answers to the need for the availability of experimentation supports in educational context. Virtual Laboratories, Adaptive Web-Based Educational Systems and Computer-Supported Collaborative Learning environments can significantly foster different learner-centered instructional strategies, offering the opportunity to enhance personalization, individualization and cooperation. More specifically, they allow students to explore different kinds of materials, to access and compare several information sources, to face real or realistic problems and to work on authentic and multi-facet case studies. In addition, they encourage cooperation among peers and provide support through coached and scaffolded activities aimed at fostering reflection and meta-cognitive reasoning. This dissertation will guide readers within this research field, presenting both the theoretical and applicative results of a research aimed at designing an open, flexible, learner-centered virtual lab for supporting students in learning Information Security.
Resumo:
The following research thesis is about a retrofit project made in Denmark, Copenhagen, and carried out on one of the buildings belonging to the Royal Danish Academy. The key assumption and base of the entire research process is that, up to now, the standard procedure in retrofit cases like this provides as comparative method between de facto and design, the use of Energy Simulation software. These programs generally divide the space into different thermal zones, assigning to each of them different levels of employment, activities, set-point temperatures set for cooling and heating analysis and so on, but always providing average and constant values, usually taken in the middle point of the single thermal zone. Therefore, the project and its research path stems from the attempt to investigate the potentialities of this kind of designing for retrofit process, as previously anticipated not antithetical but complementary to that classic energy-based retrofit, thus passing from the building scale, and all its thermal zones, to the users' scale, related to humans and microclimates. The main software used in this process is Autodesk Simulation CFD. The idea behind the project is that in certain situations, for example, it will not be necessary to add throughout insulation layers (previously parameterized and optimized with Design Builder), and that even in Winter conditions, due maybe to the users' activities, the increased level of clothing (clo) and the heat produced by equipments, thermal comfort could be achieved also in areas characterized by considerably lower MRT. After the analysis of the State of Art and its simulations, the project has still been supported by the tool itself, the CFD Software, in an iterative process aimed at achieving visible improvements in terms of MRT, on spaces with different needs and characteristics, both in Winter and Summer regimes.
Resumo:
Intensive Family Preservation Services seek to reflect the values of focusing on client strengths and viewing clients as colleagues. To promote those values, Intensive Family Preservation Programs should include a systematic form of client self monitoring in their packages of outcome measures. This paper presents a model of idiographic self-monitoring used in time series, single system research design developed for Family Partners, a family preservation program of the School for Contemporary Education in Annandale, Virginia. The evaluation model provides a means of empowering client families to utilize their strengths and promote their status as colleague in determining their own goals, participating in the change process, and measuring their own progress.
Resumo:
The renewed interest in Family Centered Practice, prompted by the funding of Family Preservation and Support Programs, has created a need for training practitioners at a number of different levels and for a variety of roles. This paper will describe a training program for Family Centered Practice. Building on an empowerment model, the author presents an approach for working with families and children that views the tragedies of the past as resources, rather than the major cause of present problems. Collaborative Conversations for Change adapts the solution-focused therapy model to nontherapy roles that are required for a program to be family centered. Although these roles are not therapy, they are nevertheless therapeutic and reinforce clients' strengths. These collaborative conversations, however brief they may be, recognize that the client is the expert on his/her pain and struggles and the practitioner is the expert on assisting her/him plan change. Additionally, illustrations from a cross-cultural perspective demonstrate the utility of collaborative conversation in enhancing cultural competence.
Resumo:
Gaining economic benefits from substantially lower labor costs has been reported as a major reason for offshoring labor-intensive information systems services to low-wage countries. However, if wage differences are so high, why is there such a high level of variation in the economic success between offshored IS projects? This study argues that offshore outsourcing involves a number of extra costs for the ^his paper was recommended for acceptance by Associate Guest Editor Erran Carmel. client organization that account for the economic failure of offshore projects. The objective is to disaggregate these extra costs into their constituent parts and to explain why they differ between offshored software projects. The focus is on software development and maintenance projects that are offshored to Indian vendors. A theoretical framework is developed a priori based on transaction cost economics (TCE) and the knowledge-based view of the firm, comple mented by factors that acknowledge the specific offshore context The framework is empirically explored using a multiple case study design including six offshored software projects in a large German financial service institution. The results of our analysis indicate that the client incurs post contractual extra costs for four types of activities: (1) re quirements specification and design, (2) knowledge transfer, (3) control, and (4) coordination. In projects that require a high level of client-specific knowledge about idiosyncratic business processes and software systems, these extra costs were found to be substantially higher than in projects where more general knowledge was needed. Notably, these costs most often arose independently from the threat of oppor tunistic behavior, challenging the predominant TCE logic of market failure. Rather, the client extra costs were parti cularly high in client-specific projects because the effort for managing the consequences of the knowledge asymmetries between client and vendor was particularly high in these projects. Prior experiences of the vendor with related client projects were found to reduce the level of extra costs but could not fully offset the increase in extra costs in highly client-specific projects. Moreover, cultural and geographic distance between client and vendor as well as personnel turnover were found to increase client extra costs. Slight evidence was found, however, that the cost-increasing impact of these factors was also leveraged in projects with a high level of required client-specific knowledge (moderator effect).
Resumo:
BACKGROUND: Reducing the complexity of major depressive disorder by symptom-based subtypes constitutes the basis of more specific treatments. To date, few studies have empirically derived symptom subtypes separated by sex, although the impact of sex has been widely accepted in depression research. METHODS: The community-based sample included 373 males and 443 females from the Zurich Program for Sustainable Development of Mental Health Services (ZInEP) manifesting depressive symptoms in the past 12 months. Latent Class Analysis (LCA) was performed separately by sex to extract sex-related depression subtypes. The subtypes were characterized by psychosocial characteristics. RESULTS: Three similar subtypes were found in both sexes: a severe typical subtype (males: 22.8%; females: 35.7%), a severe atypical subtype (males: 17.4%; females: 22.6%), and a moderate subtype (males: 25.2%; females: 41.8%). In males, two additional subgroups were identified: a severe irritable/angry-rejection sensitive (IARS) subtype (30%) comprising the largest group, and a small psychomotor retarded subtype (4%). Males belonging to the severe typical subtype exhibited the lowest masculine gender role orientation, while females of the typical subtype showed more anxiety disorders. The severe atypical subtype was associated with eating disorders in both sexes and with alcohol/drug abuse/dependence in females. In contrast, alcohol/drug abuse/dependence was associated with the severe IARS subtype in males. LIMITATIONS:The study had a cross-sectional design, allowing for no causal inferences. CONCLUSIONS:This study contributes to a better understanding of sex-related depression subtypes, which can be well distinguished on the basis of symptom profiles. This provides the base for future research investigating the etiopathogenesis and effective treatment of the heterogeneous depression disorder.
Resumo:
OBJECTIVE The aim of this cross-sectional study was to estimate bone loss of implants with platform-switching design and analyze possible risk indicators after 5 years of loading in a multi-centered private practice network. METHOD AND MATERIALS Peri-implant bone loss was measured radiographically as the distance from the implant shoulder to the mesial and distal alveolar crest, respectively. Risk factor analysis for marginal bone loss included type of implant prosthetic treatment concept and dental status of the opposite arch. RESULTS A total of 316 implants in 98 study patients after 5 years of loading were examined. The overall mean value for radiographic bone loss was 1.02 mm (SD ± 1.25 mm, 95% CI 0.90- 1.14). Correlation analyses indicated a strong association of peri-implant bone loss > 2 mm for removable implant-retained prostheses with an odds ratio of 53.8. CONCLUSION The 5-year-results of the study show clinically acceptable values of mean bone loss after 5 years of loading. Implant-supported removable prostheses seem to be a strong co-factor for extensive bone level changes compared to fixed reconstructions. However, these results have to be considered for evaluation of the included special cohort under private dental office conditions.
Resumo:
Background: Hypertension and Diabetes is a public health and economic concern in the United States. The utilization of medical home concepts increases the receipt of preventive services, however, do they also increase adherence to treatments? This study examined the effect of patient-centered medical home technologies such as the electronic health record, clinical support system, and web-based care management in improving health outcomes related to hypertension and diabetes. Methods: A systematic review of the literature used a best evidence synthesis approach to address the general question " Do patient-centered medical home technologies have an effect of diabetes and hypertension treatment?" This was followed by an evaluation of specific examples of the technologies utilized such as computer-assisted recommendations and web-based care management provided by the patient's electronic health record. Ebsco host, Ovid host, and Google Scholar were the databases used to conduct the literature search. Results: The initial search identified over 25 studies based on content and quality that implemented technology interventions to improve communication between provider and patient. After further assessing the articles for risk of bias and study design, 13 randomized controlled studies were chosen. All of the studies chosen were conducted in various primary care settings in both private practices and hospitals between the years 2000 and 2007. The sample sizes of the studies ranged from 42 to 2924 participants. The mean age for all of the studies ranged from 56 to 71 years. The percent women in the studies ranged from one to 78 percent. Over one-third of the studies did not provide the racial composition of the participants. For the seven studies that did provide information about the ethnic composition, 64% of the intervention participants were White. All of the studies utilized some type of web-based or computer-based communication to manage hypertension or diabetes care. Findings on outcomes were mixed, with nine out of 13 studies showing no significant effect on outcomes examined, and four of the studies showing significant and positive impact on health outcomes related to hypertension or diabetes Conclusion: Although the technologies improved patient and provider satisfaction, the outcomes measures such as blood pressure control and glucose control were inconclusive. Further research is needed with diverse ethnic and SES population to investigate the role of patient-centered technologies on hypertension and diabetes control. Also, further research is needed to investigate the effects of innovative medical home technologies that can be used by both patients and providers to increase quality of communication concerning adherence to treatments.^
Resumo:
These three manuscripts are presented as a PhD dissertation for the study of using GeoVis application to evaluate telehealth programs. The primary reason of this research was to understand how the GeoVis applications can be designed and developed using combined approaches of HC approach and cognitive fit theory and in terms utilized to evaluate telehealth program in Brazil. First manuscript The first manuscript in this dissertation presented a background about the use of GeoVisualization to facilitate visual exploration of public health data. The manuscript covered the existing challenges that were associated with an adoption of existing GeoVis applications. The manuscript combines the principles of Human Centered approach and Cognitive Fit Theory and a framework using a combination of these approaches is developed that lays the foundation of this research. The framework is then utilized to propose the design, development and evaluation of “the SanaViz” to evaluate telehealth data in Brazil, as a proof of concept. Second manuscript The second manuscript is a methods paper that describes the approaches that can be employed to design and develop “the SanaViz” based on the proposed framework. By defining the various elements of the HC approach and CFT, a mixed methods approach is utilized for the card sorting and sketching techniques. A representative sample of 20 study participants currently involved in the telehealth program at the NUTES telehealth center at UFPE, Recife, Brazil was enrolled. The findings of this manuscript helped us understand the needs of the diverse group of telehealth users, the tasks that they perform and helped us determine the essential features that might be necessary to be included in the proposed GeoVis application “the SanaViz”. Third manuscript The third manuscript involved mix- methods approach to compare the effectiveness and usefulness of the HC GeoVis application “the SanaViz” against a conventional GeoVis application “Instant Atlas”. The same group of 20 study participants who had earlier participated during Aim 2 was enrolled and a combination of quantitative and qualitative assessments was done. Effectiveness was gauged by the time that the participants took to complete the tasks using both the GeoVis applications, the ease with which they completed the tasks and the number of attempts that were taken to complete each task. Usefulness was assessed by System Usability Scale (SUS), a validated questionnaire tested in prior studies. In-depth interviews were conducted to gather opinions about both the GeoVis applications. This manuscript helped us in the demonstration of the usefulness and effectiveness of HC GeoVis applications to facilitate visual exploration of telehealth data, as a proof of concept. Together, these three manuscripts represent challenges of combining principles of Human Centered approach, Cognitive Fit Theory to design and develop GeoVis applications as a method to evaluate Telehealth data. To our knowledge, this is the first study to explore the usefulness and effectiveness of GeoVis to facilitate visual exploration of telehealth data. The results of the research enabled us to develop a framework for the design and development of GeoVis applications related to the areas of public health and especially telehealth. The results of our study showed that the varied users were involved with the telehealth program and the tasks that they performed. Further it enabled us to identify the components that might be essential to be included in these GeoVis applications. The results of our research answered the following questions; (a) Telehealth users vary in their level of understanding about GeoVis (b) Interaction features such as zooming, sorting, and linking and multiple views and representation features such as bar chart and choropleth maps were considered the most essential features of the GeoVis applications. (c) Comparing and sorting were two important tasks that the telehealth users would perform for exploratory data analysis. (d) A HC GeoVis prototype application is more effective and useful for exploration of telehealth data than a conventional GeoVis application. Future studies should be done to incorporate the proposed HC GeoVis framework to enable comprehensive assessment of the users and the tasks they perform to identify the features that might be necessary to be a part of the GeoVis applications. The results of this study demonstrate a novel approach to comprehensively and systematically enhance the evaluation of telehealth programs using the proposed GeoVis Framework.
Resumo:
La Tesis Doctoral nace con una intensa vocación pedagógica. La hipótesis de trabajo se establece en torno a una cuestión de interés personal, un tema sobre el que se vertebran, desde el comienzo del doctorado, los diferentes cursos y trabajos de investigación: LA CASA DOMÍNGUEZ como paradigma de la dialéctica en la obra de Alejandro de la Sota. La clasificación de la realidad en categorías antagónicas determina un orden conceptual polarizado, una red de filiaciones excluyentes sobre las que Sota construye su personal protocolo operativo: la arquitectura intelectual o popular, experimental o tradicional, universal o local, ligera o pesada, elevada o enterrada, etc. Se propone el abordaje de una cuestión latente en el conjunto de la obra ‘sotiana’, desde la disección y el análisis de una de sus obras más pequeñas: la casa Domínguez. Se trata de una organización sin precedentes, que eleva la estrategia dialéctica al paroxismo: la vivienda se separa en dos estratos independientes, la zona de día, elevada, y la zona de noche, enterrada; cada uno de los estratos establece su propio orden geométrico y constructivo, su propio lenguaje y carácter, su propia identidad e incluso su propio presupuesto. Las relaciones entre interior y exterior se especializan en función de la actividad o el reposo, estableciéndose una compleja red de relaciones, algunas evidentes y otras celosamente veladas, entre los diferentes niveles. La estancia destinada a las tareas activas se proyecta como un objeto de armazón ligero y piel fría; la precisa geometría del cubo delimita la estancia vigilante sobre el paisaje conquistado. La ladera habitada se destina al reposo y se configura como una topografía verde bajo la que se desarrollan los dormitorios en torno a patios, grietas y lucernarios, generando un paisaje propio: la construcción del objeto frente a la construcción del lugar La casa Domínguez constituye uno de los proyectos menos estudiados, y por lo tanto menos celebrados, de la obra de Don Alejandro. Las publicaciones sucesivas reproducen la documentación gráfica junto a la memoria (epopeya) que el propio Sota compone para la publicación del proyecto. Apenas un par de breves textos críticos de Miguel Ángel Baldellou y, recientemente de Moisés Puente, abordan la vivienda como tema monográfico. Sin embargo, la producción de proyecto y obra ocupó a De la Sota un periodo no inferior a diez años, con casi cien planos dibujados para dos versiones de proyecto, la primera de ellas, inédita. El empeño por determinar hasta el último detalle de la ‘pequeña’ obra, conduce a Sota a controlar incluso el mobiliario interior, como hiciera en otras obras ‘importantes’ como el Gobierno Civil de Tarragona, el colegio mayor César Carlos o el edificio de Correos y Telecomunicaciones de León. La complicidad del cliente, mantenida durante casi cuarenta años, habilita el despliegue de una importante colección de recursos y herramientas de proyecto. La elección de la casa Domínguez como tema central de la tesis persigue por lo tanto un triple objetivo: en primer lugar, el abordaje del proyecto como paradigma de la dialéctica ‘sotiana’, analizando la coherencia entre el discurso de carácter heroico y la obra finalmente construida; en segundo lugar, la investigación rigurosa, de corte científico, desde la disección y progresivo desmontaje del objeto arquitectónico; y por último, la reflexión sobre los temas y dispositivos de proyecto que codifican la identificación entre la acción de construir y el hecho de habitar, registrando los aciertos y valorando con actitud crítica aquellos elementos poco coherentes con el orden interno de la propuesta. This doctoral thesis is the fruit of a profound pedagogical vocation. The central hypothesis was inspired by a question of great personal interest, and this interest has, since the very beginning of the doctorate, been the driving force behind all subsequent lines of research and investigation. The “Casa Domínguez” represents a paradigm of the dialectics found in the work of Alejandro de la Sota. The perception of reality as antagonistic categories determines a polarized conceptual order, a network of mutually excluding associations upon which Sota builds his own personal operative protocol: intellectual or popular architecture, experimental or traditional, universal or local, heavy or light, above or below ground, etc. Through the analysis and dissection of the “Casa Domínguez”, one of Sota’s smallest projects, an attempt is made to approach the underlying question posed in “Sotian” work as a whole. This is about organization without precedent, raising the strategic dialectics to levels of paroxysm. The house is divided into two separate levels, the day-time level above ground, and the lower night-time level beneath the surface of the ground. Each level has its own geometrical and stuctural order, its own language and character, its own identity and even has its own construction budget. The interaction between the two areas is centered on the two functions of rest and activity, and this in turn establishes a complex relationship network between both, which is sometimes self-evident, but at other times jealously guarded. The living area designed for daily activity is presented as an object of light structure and delicate skin; the precise geometry of the cube delimiting the ever watchful living area’s domain over the land it has conquered. A green topography is created on the slope below which lies an area adapted for rest and relaxation. Two bedrooms, built around patios, skylights and light crevices, generate an entirely independent environment: the construction of an object as opposed to the creation of a landscape. The “Casa Domínguez” project has been subject to much less scrutiny and examination than Don Alejandro’s other works, and is consequently less well-known. A succession of journals have printed the blueprint document together with a poetic description (epopee), composed by Sota himself, to mark the project’s publication. There has, however, scarcely been more than two brief critical appraisals, those by Miguel Ángel Baldellou and more recently by Moisés Puente, that have regarded the project as a monographic work. The project and works nevertheless occupied no less than ten years of De La Sota’s life, with over a hundred draft drawings for two separate versions of the project, the first of which remains unpublished. The sheer determination to design this “small” work in the most meticulous detail, drove Sota to manage and select its interior furniture, as indeed he had previously done with more “important” works like the Tarragona Civil Government, César Carlos College, or the Post Office telecommunications building in León. Client collaboration, maintained over a period of almost forty years, has facilitated an impressive array of the project’s tools and resources. The choice of “Casa Domínguez” as the central subject matter of this thesis, was made in pursuance of a triple objective: firstly, to approach the project as a paradigm of the “Sotian” dialectic, the analysis of the discourse between the heroic character and the finished building; secondly, a rigorous scientific investigation, and progressive disassembling and dissecting of the architectonic object; and finally, a reflection on aspects of the project and its technology which codify the identification between the action of construction and the reality of living, thus marking its achievements, whilst at the same time subjecting incoherent elements of the proposal’s established order to a critical evaluation.
Resumo:
La temperatura es una preocupación que juega un papel protagonista en el diseño de circuitos integrados modernos. El importante aumento de las densidades de potencia que conllevan las últimas generaciones tecnológicas ha producido la aparición de gradientes térmicos y puntos calientes durante el funcionamiento normal de los chips. La temperatura tiene un impacto negativo en varios parámetros del circuito integrado como el retardo de las puertas, los gastos de disipación de calor, la fiabilidad, el consumo de energía, etc. Con el fin de luchar contra estos efectos nocivos, la técnicas de gestión dinámica de la temperatura (DTM) adaptan el comportamiento del chip en función en la información que proporciona un sistema de monitorización que mide en tiempo de ejecución la información térmica de la superficie del dado. El campo de la monitorización de la temperatura en el chip ha llamado la atención de la comunidad científica en los últimos años y es el objeto de estudio de esta tesis. Esta tesis aborda la temática de control de la temperatura en el chip desde diferentes perspectivas y niveles, ofreciendo soluciones a algunos de los temas más importantes. Los niveles físico y circuital se cubren con el diseño y la caracterización de dos nuevos sensores de temperatura especialmente diseñados para los propósitos de las técnicas DTM. El primer sensor está basado en un mecanismo que obtiene un pulso de anchura variable dependiente de la relación de las corrientes de fuga con la temperatura. De manera resumida, se carga un nodo del circuito y posteriormente se deja flotando de tal manera que se descarga a través de las corrientes de fugas de un transistor; el tiempo de descarga del nodo es la anchura del pulso. Dado que la anchura del pulso muestra una dependencia exponencial con la temperatura, la conversión a una palabra digital se realiza por medio de un contador logarítmico que realiza tanto la conversión tiempo a digital como la linealización de la salida. La estructura resultante de esta combinación de elementos se implementa en una tecnología de 0,35 _m. El sensor ocupa un área muy reducida, 10.250 nm2, y consume muy poca energía, 1.05-65.5nW a 5 muestras/s, estas cifras superaron todos los trabajos previos en el momento en que se publicó por primera vez y en el momento de la publicación de esta tesis, superan a todas las implementaciones anteriores fabricadas en el mismo nodo tecnológico. En cuanto a la precisión, el sensor ofrece una buena linealidad, incluso sin calibrar; se obtiene un error 3_ de 1,97oC, adecuado para tratar con las aplicaciones de DTM. Como se ha explicado, el sensor es completamente compatible con los procesos de fabricación CMOS, este hecho, junto con sus valores reducidos de área y consumo, lo hacen especialmente adecuado para la integración en un sistema de monitorización de DTM con un conjunto de monitores empotrados distribuidos a través del chip. Las crecientes incertidumbres de proceso asociadas a los últimos nodos tecnológicos comprometen las características de linealidad de nuestra primera propuesta de sensor. Con el objetivo de superar estos problemas, proponemos una nueva técnica para obtener la temperatura. La nueva técnica también está basada en las dependencias térmicas de las corrientes de fuga que se utilizan para descargar un nodo flotante. La novedad es que ahora la medida viene dada por el cociente de dos medidas diferentes, en una de las cuales se altera una característica del transistor de descarga |la tensión de puerta. Este cociente resulta ser muy robusto frente a variaciones de proceso y, además, la linealidad obtenida cumple ampliamente los requisitos impuestos por las políticas DTM |error 3_ de 1,17oC considerando variaciones del proceso y calibrando en dos puntos. La implementación de la parte sensora de esta nueva técnica implica varias consideraciones de diseño, tales como la generación de una referencia de tensión independiente de variaciones de proceso, que se analizan en profundidad en la tesis. Para la conversión tiempo-a-digital, se emplea la misma estructura de digitalización que en el primer sensor. Para la implementación física de la parte de digitalización, se ha construido una biblioteca de células estándar completamente nueva orientada a la reducción de área y consumo. El sensor resultante de la unión de todos los bloques se caracteriza por una energía por muestra ultra baja (48-640 pJ) y un área diminuta de 0,0016 mm2, esta cifra mejora todos los trabajos previos. Para probar esta afirmación, se realiza una comparación exhaustiva con más de 40 propuestas de sensores en la literatura científica. Subiendo el nivel de abstracción al sistema, la tercera contribución se centra en el modelado de un sistema de monitorización que consiste de un conjunto de sensores distribuidos por la superficie del chip. Todos los trabajos anteriores de la literatura tienen como objetivo maximizar la precisión del sistema con el mínimo número de monitores. Como novedad, en nuestra propuesta se introducen nuevos parámetros de calidad aparte del número de sensores, también se considera el consumo de energía, la frecuencia de muestreo, los costes de interconexión y la posibilidad de elegir diferentes tipos de monitores. El modelo se introduce en un algoritmo de recocido simulado que recibe la información térmica de un sistema, sus propiedades físicas, limitaciones de área, potencia e interconexión y una colección de tipos de monitor; el algoritmo proporciona el tipo seleccionado de monitor, el número de monitores, su posición y la velocidad de muestreo _optima. Para probar la validez del algoritmo, se presentan varios casos de estudio para el procesador Alpha 21364 considerando distintas restricciones. En comparación con otros trabajos previos en la literatura, el modelo que aquí se presenta es el más completo. Finalmente, la última contribución se dirige al nivel de red, partiendo de un conjunto de monitores de temperatura de posiciones conocidas, nos concentramos en resolver el problema de la conexión de los sensores de una forma eficiente en área y consumo. Nuestra primera propuesta en este campo es la introducción de un nuevo nivel en la jerarquía de interconexión, el nivel de trillado (o threshing en inglés), entre los monitores y los buses tradicionales de periféricos. En este nuevo nivel se aplica selectividad de datos para reducir la cantidad de información que se envía al controlador central. La idea detrás de este nuevo nivel es que en este tipo de redes la mayoría de los datos es inútil, porque desde el punto de vista del controlador sólo una pequeña cantidad de datos |normalmente sólo los valores extremos| es de interés. Para cubrir el nuevo nivel, proponemos una red de monitorización mono-conexión que se basa en un esquema de señalización en el dominio de tiempo. Este esquema reduce significativamente tanto la actividad de conmutación sobre la conexión como el consumo de energía de la red. Otra ventaja de este esquema es que los datos de los monitores llegan directamente ordenados al controlador. Si este tipo de señalización se aplica a sensores que realizan conversión tiempo-a-digital, se puede obtener compartición de recursos de digitalización tanto en tiempo como en espacio, lo que supone un importante ahorro de área y consumo. Finalmente, se presentan dos prototipos de sistemas de monitorización completos que de manera significativa superan la características de trabajos anteriores en términos de área y, especialmente, consumo de energía. Abstract Temperature is a first class design concern in modern integrated circuits. The important increase in power densities associated to recent technology evolutions has lead to the apparition of thermal gradients and hot spots during run time operation. Temperature impacts several circuit parameters such as speed, cooling budgets, reliability, power consumption, etc. In order to fight against these negative effects, dynamic thermal management (DTM) techniques adapt the behavior of the chip relying on the information of a monitoring system that provides run-time thermal information of the die surface. The field of on-chip temperature monitoring has drawn the attention of the scientific community in the recent years and is the object of study of this thesis. This thesis approaches the matter of on-chip temperature monitoring from different perspectives and levels, providing solutions to some of the most important issues. The physical and circuital levels are covered with the design and characterization of two novel temperature sensors specially tailored for DTM purposes. The first sensor is based upon a mechanism that obtains a pulse with a varying width based on the variations of the leakage currents on the temperature. In a nutshell, a circuit node is charged and subsequently left floating so that it discharges away through the subthreshold currents of a transistor; the time the node takes to discharge is the width of the pulse. Since the width of the pulse displays an exponential dependence on the temperature, the conversion into a digital word is realized by means of a logarithmic counter that performs both the timeto- digital conversion and the linearization of the output. The structure resulting from this combination of elements is implemented in a 0.35_m technology and is characterized by very reduced area, 10250 nm2, and power consumption, 1.05-65.5 nW at 5 samples/s, these figures outperformed all previous works by the time it was first published and still, by the time of the publication of this thesis, they outnumber all previous implementations in the same technology node. Concerning the accuracy, the sensor exhibits good linearity, even without calibration it displays a 3_ error of 1.97oC, appropriate to deal with DTM applications. As explained, the sensor is completely compatible with standard CMOS processes, this fact, along with its tiny area and power overhead, makes it specially suitable for the integration in a DTM monitoring system with a collection of on-chip monitors distributed across the chip. The exacerbated process fluctuations carried along with recent technology nodes jeop-ardize the linearity characteristics of the first sensor. In order to overcome these problems, a new temperature inferring technique is proposed. In this case, we also rely on the thermal dependencies of leakage currents that are used to discharge a floating node, but now, the result comes from the ratio of two different measures, in one of which we alter a characteristic of the discharging transistor |the gate voltage. This ratio proves to be very robust against process variations and displays a more than suficient linearity on the temperature |1.17oC 3_ error considering process variations and performing two-point calibration. The implementation of the sensing part based on this new technique implies several issues, such as the generation of process variations independent voltage reference, that are analyzed in depth in the thesis. In order to perform the time-to-digital conversion, we employ the same digitization structure the former sensor used. A completely new standard cell library targeting low area and power overhead is built from scratch to implement the digitization part. Putting all the pieces together, we achieve a complete sensor system that is characterized by ultra low energy per conversion of 48-640pJ and area of 0.0016mm2, this figure outperforms all previous works. To prove this statement, we perform a thorough comparison with over 40 works from the scientific literature. Moving up to the system level, the third contribution is centered on the modeling of a monitoring system consisting of set of thermal sensors distributed across the chip. All previous works from the literature target maximizing the accuracy of the system with the minimum number of monitors. In contrast, we introduce new metrics of quality apart form just the number of sensors; we consider the power consumption, the sampling frequency, the possibility to consider different types of monitors and the interconnection costs. The model is introduced in a simulated annealing algorithm that receives the thermal information of a system, its physical properties, area, power and interconnection constraints and a collection of monitor types; the algorithm yields the selected type of monitor, the number of monitors, their position and the optimum sampling rate. We test the algorithm with the Alpha 21364 processor under several constraint configurations to prove its validity. When compared to other previous works in the literature, the modeling presented here is the most complete. Finally, the last contribution targets the networking level, given an allocated set of temperature monitors, we focused on solving the problem of connecting them in an efficient way from the area and power perspectives. Our first proposal in this area is the introduction of a new interconnection hierarchy level, the threshing level, in between the monitors and the traditional peripheral buses that applies data selectivity to reduce the amount of information that is sent to the central controller. The idea behind this new level is that in this kind of networks most data are useless because from the controller viewpoint just a small amount of data |normally extreme values| is of interest. To cover the new interconnection level, we propose a single-wire monitoring network based on a time-domain signaling scheme that significantly reduces both the switching activity over the wire and the power consumption of the network. This scheme codes the information in the time domain and allows a straightforward obtention of an ordered list of values from the maximum to the minimum. If the scheme is applied to monitors that employ TDC, digitization resource sharing is achieved, producing an important saving in area and power consumption. Two prototypes of complete monitoring systems are presented, they significantly overcome previous works in terms of area and, specially, power consumption.
Resumo:
El concepto de funicularidad se puede extender a estructuras lineales espaciales como, por ejemplo, los puentes arco con tablero curvo. Estas estructuras, especialmente pasarelas peatonales, son consecuencia de la necesidad de encajar trazados exigentes y de dar respuesta a nuevas demandas arquitectónicas. En las estructuras curvas el diseño conceptual juega un papel absolutamente esencial. Siempre ha sido así, pero en el caso presente, cabe resaltar que una errónea elección de la geometría conlleva una serie de problemas que se irán acumulando a lo largo del proceso de proyecto, de la construcción y de la vida de la estructura. En este trabajo se presenta SOFIA (Shaping Optimal Form with an Interactive Approach), una herramienta capaz de, conocida la geometría del tablero, de buscar automáticamente la forma del arco antifunicular correspondiente. El planteamiento seguido es conceptualmente el mismo que el utilizado en la búsqueda de formas óptimas en estructuras en dos dimensiones: el arco antifunicular es el que representa, para unas cargas dadas, el lugar geométrico de los puntos con momento flector nulo. La herramienta ha sido desarrollada en un entorno integrado, interactivo y paramético. Su implementación está ilustrada y unos ejemplos de análisis paramétricos están desarrollados. La posición transversal relativa entre tablero y arco ha sido investigada para obtener la configuración del puente estructuralmente más eficiente. Las pasarelas curvas se han convertido en un problema de ingeniería más común de lo habitual en el contexto de los desarrollos urbanos cuando el cliente está buscando un fuerte componente estético: un diseño conceptual adecuado permite obtener una estructura eficiente y elegante. Spatial arch bridges represent an innovative answer to demands on functionality, structural optimization and aesthetics for curved decks, popular in urban contexts. This thesis presents SOFIA (Shaping Optimal Form with an Interactive Approach), a methodology for conceptual designing of antifunicular spatial arch bridges with curved deck in a parametric, interactive and integrated environment. The approach and its implementation are in-depth described and detailed examples of parametric analyses are illustrated. The optimal deck-arch relative transversal position has been investigated for obtaining the most cost-effective bridge. Curved footbridges have become a more common engineering problem in the context of urban developments when the client is looking for a strong aesthetics component: an appropriate conceptual design allows to obtain an efficient and elegant structure.
Resumo:
From research groups at the universities of developed countries there is a growing interest in providing solutions to problems of developing countries. In this context we have studied typical problems in many (educational) institutions, such as the lack of technicians who repair the computers, the administration of the machines, and also the difficulty to maintain and configure the old hardware available due to the variety of characteristics of the different machines and the amount of hardware breakdowns and software issues (viruses, administration issues) that the local staff has to face up to with their equipments. We propose a thin client approach that takes into account the human, hardware and software characteristics of developing institutions to provide a complete service for a computer network. The network administration is reduced to the administration of one server only. The maintenance of the machines is simplified and old computers can simulate the running of a powerful computer. Our proposal results in a cheap, simple (from the support point of view) and powerful (in terms of achieved functionalities) design.