945 resultados para Multiple abstraction levels


Relevância:

80.00% 80.00%

Publicador:

Resumo:

La meta de intercambiabilidad de piezas establecida en los sistemas de producción del siglo XIX, es ampliada en el último cuarto del siglo pasado para lograr la capacidad de fabricación de varios tipos de producto en un mismo sistema de manufactura, requerimiento impulsado por la incertidumbre del mercado. Esta incertidumbre conduce a plantear la flexibilidad como característica importante en el sistema de producción. La presente tesis se ubica en el problema de integración del sistema informático (SI) con el equipo de producción (EP) en la búsqueda de una solución que coadyuve a satisfacer los requerimientos de flexibilidad impuestas por las condiciones actuales de mercado. Se describen antecedentes de los sistemas de producción actuales y del concepto de flexibilidad. Se propone una clasificación compacta y práctica de los tipos de flexibilidad relevantes en el problema de integración SI-EP, con la finalidad de ubicar el significado de flexibilidad en el área de interés. Así mismo, las variables a manejar en la solución son clasificadas en cuatro tipos: Medio físico, lenguajes de programación y controlador, naturaleza del equipo y componentes de acoplamiento. Por otra parte, la característica de reusabilidad como un efecto importante y deseable de un sistema flexible, es planteada como meta en la solución propuesta no solo a nivel aplicación del sistema sino también a nivel de reuso de conceptos de diseño. Se propone un esquema de referencia en tres niveles de abstracción, que permita manejar y reutilizar en forma organizada el conocimiento del dominio de aplicación (integración SI-EP), el desarrollo de sistemas de aplicación genérica así como también la aplicación del mismo en un caso particular. Un análisis del concepto de acoplamiento débil (AD) es utilizado como base en la solución propuesta al problema de integración SI-EP. El desarrollo inicia identificando condiciones para la existencia del acoplamiento débil, compensadores para soportar la operación del sistema bajo AD y los efectos que ocasionan en el sistema informático los cambios en el conjunto de equipos de producción. Así mismo, se introducen como componentes principales del acoplamiento los componentes tecnológico, tarea y rol, a utilizar en el análisis de los requerimientos para el desarrollo de una solución de AD entre SI-EP. La estructura de tres niveles del esquema de referencia propuesto surge del análisis del significado de conceptos de referencia comúnmente reportados en la literatura, tales como arquitectura de referencia, modelo de referencia, marco de trabajo, entre otros. Se presenta un análisis de su significado como base para la definición de cada uno de los niveles de la estructura del esquema, pretendiendo con ello evitar la ambigüedad existente debido al uso indistinto de tales conceptos en la literatura revisada. Por otra parte, la relación entre niveles es definida tomando como base la estructura de cuatro capas planteada en el área de modelado de datos. La arquitectura de referencia, implementada en el primer nivel del esquema propuesto es utilizada como base para el desarrollo del modelo de referencia o marco de trabajo para el acoplamiento débil entre el SI y el EP. La solución propuesta es validada en la integración de un sistema informático de coordinación de flujo y procesamiento de pieza con un conjunto variable de equipos de diferentes tipos, naturaleza y fabricantes. En el ejercicio de validación se abordaron diferentes estándares y técnicas comúnmente empleadas como soporte al problema de integración a nivel componente tecnológico, tales como herramientas de cero configuración (ejemplo: plug and play), estándar OPC-UA, colas de mensajes y servicios web, permitiendo así ubicar el apoyo de estas técnicas en el ámbito del componente tecnológico y su relación con los otros componentes de acoplamiento: tarea y rol. ABSTRACT The interchangeability of parts, as a goal of manufacturing systems at the nineteenth century, is extended into the present to achieve the ability to manufacture various types of products in the same manufacturing system, requirement associated with market uncertainty. This uncertainty raises flexibility as an important feature in the production system. This thesis addresses the problem regarding integration of software system (SS) and the set of production equipment (PE); looking for a solution that contributes to satisfy the requirements of flexibility that the current market conditions impose on manufacturing, particularly to the production floor. Antecedents to actual production systems as well as the concept of flexibility are described and analyzed in detail. A practical and compact classification of flexibility types of relevance to the integration SS-EP problem is proposed with the aim to delimit the meaning of flexibility regarding the area of interest. Also, a classification for the variables involved in the integration problem is presented into four types: Physical media, programming and controller languages, equipment nature and coupling components. In addition, the characteristic of reusability that has been seen as an important and desirable effect of a flexible system is taken as a goal in the proposed solution, not only at system implementation level but also at system design level. In this direction, a reference scheme is proposed consisting of three abstraction levels to systematically support management and reuse of domain knowledge (SS-PE), development of a generic system as well as its application in a particular case. The concept of loose coupling is used as a basis in the development of the proposed solution to the problem of integration SS-EP. The first step of the development process consists of an analysis of the loose coupled concept, identifying conditions for its existence, compensators for system operation under loose coupling conditions as well as effects in the software system caused by modification in the set of production equipment. In addition coupling components: technological, task and role are introduced as main components to support the analysis of requirements regarding loose coupling of SS-PE. The three tier structure of the proposed reference scheme emerges from the analysis of reference concepts commonly reported in the literature, such as reference architecture, reference model and framework, among others. An analysis of these concepts is used as a basis for definition of the structure levels of the proposed scheme, trying to avoid the ambiguity due to the indiscriminate use of such concepts in the reviewed literature. In addition, the relation between adjacent levels of the structure is defined based on the four tiers structure commonly used in the data modelling area. The reference architecture is located as the first level in the structure of the proposed reference scheme and it is utilized as a basis for the development of the reference model or loose coupling framework for SS-PE integration. The proposed solution is validated by integrating a software system (process and piece flow coordination system) with a variable set of production equipment including different types, nature and manufacturers of equipment. Furthermore, in this validation exercise, different standards and techniques commonly used have been taken into account to support the issue of technology coupling component, such as tools for zero configuration (i.e. Plug and Play), message queues, OPC-UA standard, and web services. Through this part of the validation exercise, these integration tools are located as a part of the technological component and they are related to the role and task components of coupling.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cities have a major impact on Australian landscapes, especially in coastal regions, to the detriment of native biodiversity. Areas suitable for urban development often coincide with those areas that support high levels of species diversity and endemism. However, there is a paucity of reliable information available to guide urban conservation planning and management, especially regarding the trade-off between investing in protecting and restoring habitat at the landscape level, and investing in programmes to maintain the condition of remnant vegetation at the local (site) level. We review the literature on Australian urban ecology, focusing on urban terrestrial and aquatic vertebrate and invertebrate fauna. We identify four main factors limiting our knowledge of urban fauna: (i) a lack of studies focusing at multiple ecological levels; (ii) a lack of multispecies studies; (iii) an almost total absence of long-term (temporal) studies; and (iv) a need for stronger integration of research outcomes into urban conservation planning and management. We present a set of key principles for the development of a spatially explicit, long-term approach to urban fauna research. This requires an understanding of the importance of local-level habitat quality and condition relative to the composition, configuration and connectivity of habitats within the larger urban landscape. These principles will ultimately strengthen urban fauna management and conservation planning by enabling us to prioritize and allocate limited financial resources to maximize the conservation return.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Much research pursues machine intelligence through better representation of semantics. What is semantics? People in different areas view semantics from different facets although it accompanies interaction through civilization. Some researchers believe that humans have some innate structure in mind for processing semantics. Then, what the structure is like? Some argue that humans evolve a structure for processing semantics through constant learning. Then, how the process is like? Humans have invented various symbol systems to represent semantics. Can semantics be accurately represented? Turing machines are good at processing symbols according to algorithms designed by humans, but they are limited in ability to process semantics and to do active interaction. Super computers and high-speed networks do not help solve this issue as they do not have any semantic worldview and cannot reflect themselves. Can future cyber-society have some semantic images that enable machines and individuals (humans and agents) to reflect themselves and interact with each other with knowing social situation through time? This paper concerns these issues in the context of studying an interactive semantics for the future cyber-society. It firstly distinguishes social semantics from natural semantics, and then explores the interactive semantics in the category of social semantics. Interactive semantics consists of an interactive system and its semantic image, which co-evolve and influence each other. The semantic worldview and interactive semantic base are proposed as the semantic basis of interaction. The process of building and explaining semantic image can be based on an evolving structure incorporating adaptive multi-dimensional classification space and self-organized semantic link network. A semantic lens is proposed to enhance the potential of the structure and help individuals build and retrieve semantic images from different facets, abstraction levels and scales through time.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In different fields a conception of granules is applied both as a group of elements defined by internal properties and as something inseparable whole reflecting external properties. Granular computing may be interpreted in terms of abstraction, generalization, clustering, levels of abstraction, levels of detail, and so on. We have proposed to use multialgebraic systems as a mathematical tool for synthesis and analysis of granules and granule structures. The theorem of necessary and sufficient conditions for multialgebraic systems existence has been proved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This dissertation studied the determinants and consequences of corporate reputation. It explored how firm-, industry-, and country-level factors influence the general public’s assessment of a firm’s reputation and how this reputation assessment impacted the firm’s strategic actions and organizational outcomes. The three empirical essays are grounded on separate theoretical paradigms in strategy, organizational theory, and corporate governance. The first essay used signaling theory to investigate firm-, industry-, and country-level determinants of individual-level corporate reputation assessments. Using a hierarchical linear model, it tested the theory based on individual evaluations of the largest companies across countries. Results indicated that variables at multiple analysis levels simultaneously impact individual level reputation assessments. Interactions were also found between industry- and country-level factors. Results confirmed the multi-level nature of signaling influences on reputation assessments. Building on a stakeholder-power approach to corporate governance, the second essay studied how differences in the power and preferences of three stakeholder groups—shareholders, creditors, and workers—across countries influence the general public’s reputation assessments of corporations. Examining the largest companies across countries, the study found that while the influence of stock market return is stronger in societies where shareholders have more power, social performance has a more significant role in shaping reputation evaluations in societies with stronger labor rights. Unexpectedly, when creditors have greater power, the influence of financial stability on reputation assessment becomes weaker. Exploring the consequences of reputation, the third essay investigated the specific effects of intangible assets on strategic actions and organizational outcomes. Particularly, it individually studied the impacts of acquirer acquisition experience, corporate reputation, and approach toward social responsibilities as well as their combined effect on market reactions to acquisition announcements. Using an event study of acquisition announcements, it confirmed the significant impacts of both action-specific (acquisition experience) and general (reputation and social performance) intangible assets on market expectations of acquisition outcomes. Moreover, the analysis demonstrated that reputation magnifies the impact of acquisition experience on market response to acquisition announcements. In conclusion, this dissertation tried to advance and extend the application of management and organizational theories by explaining the mechanisms underlying antecedents and consequences of corporate reputation.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Over the past few decades, we have been enjoying tremendous benefits thanks to the revolutionary advancement of computing systems, driven mainly by the remarkable semiconductor technology scaling and the increasingly complicated processor architecture. However, the exponentially increased transistor density has directly led to exponentially increased power consumption and dramatically elevated system temperature, which not only adversely impacts the system's cost, performance and reliability, but also increases the leakage and thus the overall power consumption. Today, the power and thermal issues have posed enormous challenges and threaten to slow down the continuous evolvement of computer technology. Effective power/thermal-aware design techniques are urgently demanded, at all design abstraction levels, from the circuit-level, the logic-level, to the architectural-level and the system-level. ^ In this dissertation, we present our research efforts to employ real-time scheduling techniques to solve the resource-constrained power/thermal-aware, design-optimization problems. In our research, we developed a set of simple yet accurate system-level models to capture the processor's thermal dynamic as well as the interdependency of leakage power consumption, temperature, and supply voltage. Based on these models, we investigated the fundamental principles in power/thermal-aware scheduling, and developed real-time scheduling techniques targeting at a variety of design objectives, including peak temperature minimization, overall energy reduction, and performance maximization. ^ The novelty of this work is that we integrate the cutting-edge research on power and thermal at the circuit and architectural-level into a set of accurate yet simplified system-level models, and are able to conduct system-level analysis and design based on these models. The theoretical study in this work serves as a solid foundation for the guidance of the power/thermal-aware scheduling algorithms development in practical computing systems.^

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Stable isotope analysis has emerged as one of the primary means for examining the structure and dynamics of food webs, and numerous analytical approaches are now commonly used in the field. Techniques range from simple, qualitative inferences based on the isotopic niche, to Bayesian mixing models that can be used to characterize food-web structure at multiple hierarchical levels. We provide a comprehensive review of these techniques, and thus a single reference source to help identify the most useful approaches to apply to a given data set. We structure the review around four general questions: (1) what is the trophic position of an organism in a food web?; (2) which resource pools support consumers?; (3) what additional information does relative position of consumers in isotopic space reveal about food-web structure?; and (4) what is the degree of trophic variability at the intrapopulation level? For each general question, we detail different approaches that have been applied, discussing the strengths and weaknesses of each. We conclude with a set of suggestions that transcend individual analytical approaches, and provide guidance for future applications in the field.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

X-ray computed tomography (CT) imaging constitutes one of the most widely used diagnostic tools in radiology today with nearly 85 million CT examinations performed in the U.S in 2011. CT imparts a relatively high amount of radiation dose to the patient compared to other x-ray imaging modalities and as a result of this fact, coupled with its popularity, CT is currently the single largest source of medical radiation exposure to the U.S. population. For this reason, there is a critical need to optimize CT examinations such that the dose is minimized while the quality of the CT images is not degraded. This optimization can be difficult to achieve due to the relationship between dose and image quality. All things being held equal, reducing the dose degrades image quality and can impact the diagnostic value of the CT examination.

A recent push from the medical and scientific community towards using lower doses has spawned new dose reduction technologies such as automatic exposure control (i.e., tube current modulation) and iterative reconstruction algorithms. In theory, these technologies could allow for scanning at reduced doses while maintaining the image quality of the exam at an acceptable level. Therefore, there is a scientific need to establish the dose reduction potential of these new technologies in an objective and rigorous manner. Establishing these dose reduction potentials requires precise and clinically relevant metrics of CT image quality, as well as practical and efficient methodologies to measure such metrics on real CT systems. The currently established methodologies for assessing CT image quality are not appropriate to assess modern CT scanners that have implemented those aforementioned dose reduction technologies.

Thus the purpose of this doctoral project was to develop, assess, and implement new phantoms, image quality metrics, analysis techniques, and modeling tools that are appropriate for image quality assessment of modern clinical CT systems. The project developed image quality assessment methods in the context of three distinct paradigms, (a) uniform phantoms, (b) textured phantoms, and (c) clinical images.

The work in this dissertation used the “task-based” definition of image quality. That is, image quality was broadly defined as the effectiveness by which an image can be used for its intended task. Under this definition, any assessment of image quality requires three components: (1) A well defined imaging task (e.g., detection of subtle lesions), (2) an “observer” to perform the task (e.g., a radiologists or a detection algorithm), and (3) a way to measure the observer’s performance in completing the task at hand (e.g., detection sensitivity/specificity).

First, this task-based image quality paradigm was implemented using a novel multi-sized phantom platform (with uniform background) developed specifically to assess modern CT systems (Mercury Phantom, v3.0, Duke University). A comprehensive evaluation was performed on a state-of-the-art CT system (SOMATOM Definition Force, Siemens Healthcare) in terms of noise, resolution, and detectability as a function of patient size, dose, tube energy (i.e., kVp), automatic exposure control, and reconstruction algorithm (i.e., Filtered Back-Projection– FPB vs Advanced Modeled Iterative Reconstruction– ADMIRE). A mathematical observer model (i.e., computer detection algorithm) was implemented and used as the basis of image quality comparisons. It was found that image quality increased with increasing dose and decreasing phantom size. The CT system exhibited nonlinear noise and resolution properties, especially at very low-doses, large phantom sizes, and for low-contrast objects. Objective image quality metrics generally increased with increasing dose and ADMIRE strength, and with decreasing phantom size. The ADMIRE algorithm could offer comparable image quality at reduced doses or improved image quality at the same dose (increase in detectability index by up to 163% depending on iterative strength). The use of automatic exposure control resulted in more consistent image quality with changing phantom size.

Based on those results, the dose reduction potential of ADMIRE was further assessed specifically for the task of detecting small (<=6 mm) low-contrast (<=20 HU) lesions. A new low-contrast detectability phantom (with uniform background) was designed and fabricated using a multi-material 3D printer. The phantom was imaged at multiple dose levels and images were reconstructed with FBP and ADMIRE. Human perception experiments were performed to measure the detection accuracy from FBP and ADMIRE images. It was found that ADMIRE had equivalent performance to FBP at 56% less dose.

Using the same image data as the previous study, a number of different mathematical observer models were implemented to assess which models would result in image quality metrics that best correlated with human detection performance. The models included naïve simple metrics of image quality such as contrast-to-noise ratio (CNR) and more sophisticated observer models such as the non-prewhitening matched filter observer model family and the channelized Hotelling observer model family. It was found that non-prewhitening matched filter observers and the channelized Hotelling observers both correlated strongly with human performance. Conversely, CNR was found to not correlate strongly with human performance, especially when comparing different reconstruction algorithms.

The uniform background phantoms used in the previous studies provided a good first-order approximation of image quality. However, due to their simplicity and due to the complexity of iterative reconstruction algorithms, it is possible that such phantoms are not fully adequate to assess the clinical impact of iterative algorithms because patient images obviously do not have smooth uniform backgrounds. To test this hypothesis, two textured phantoms (classified as gross texture and fine texture) and a uniform phantom of similar size were built and imaged on a SOMATOM Flash scanner (Siemens Healthcare). Images were reconstructed using FBP and a Sinogram Affirmed Iterative Reconstruction (SAFIRE). Using an image subtraction technique, quantum noise was measured in all images of each phantom. It was found that in FBP, the noise was independent of the background (textured vs uniform). However, for SAFIRE, noise increased by up to 44% in the textured phantoms compared to the uniform phantom. As a result, the noise reduction from SAFIRE was found to be up to 66% in the uniform phantom but as low as 29% in the textured phantoms. Based on this result, it clear that further investigation was needed into to understand the impact that background texture has on image quality when iterative reconstruction algorithms are used.

To further investigate this phenomenon with more realistic textures, two anthropomorphic textured phantoms were designed to mimic lung vasculature and fatty soft tissue texture. The phantoms (along with a corresponding uniform phantom) were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Scans were repeated a total of 50 times in order to get ensemble statistics of the noise. A novel method of estimating the noise power spectrum (NPS) from irregularly shaped ROIs was developed. It was found that SAFIRE images had highly locally non-stationary noise patterns with pixels near edges having higher noise than pixels in more uniform regions. Compared to FBP, SAFIRE images had 60% less noise on average in uniform regions for edge pixels, noise was between 20% higher and 40% lower. The noise texture (i.e., NPS) was also highly dependent on the background texture for SAFIRE. Therefore, it was concluded that quantum noise properties in the uniform phantoms are not representative of those in patients for iterative reconstruction algorithms and texture should be considered when assessing image quality of iterative algorithms.

The move beyond just assessing noise properties in textured phantoms towards assessing detectability, a series of new phantoms were designed specifically to measure low-contrast detectability in the presence of background texture. The textures used were optimized to match the texture in the liver regions actual patient CT images using a genetic algorithm. The so called “Clustured Lumpy Background” texture synthesis framework was used to generate the modeled texture. Three textured phantoms and a corresponding uniform phantom were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Images were reconstructed with FBP and SAFIRE and analyzed using a multi-slice channelized Hotelling observer to measure detectability and the dose reduction potential of SAFIRE based on the uniform and textured phantoms. It was found that at the same dose, the improvement in detectability from SAFIRE (compared to FBP) was higher when measured in a uniform phantom compared to textured phantoms.

The final trajectory of this project aimed at developing methods to mathematically model lesions, as a means to help assess image quality directly from patient images. The mathematical modeling framework is first presented. The models describe a lesion’s morphology in terms of size, shape, contrast, and edge profile as an analytical equation. The models can be voxelized and inserted into patient images to create so-called “hybrid” images. These hybrid images can then be used to assess detectability or estimability with the advantage that the ground truth of the lesion morphology and location is known exactly. Based on this framework, a series of liver lesions, lung nodules, and kidney stones were modeled based on images of real lesions. The lesion models were virtually inserted into patient images to create a database of hybrid images to go along with the original database of real lesion images. ROI images from each database were assessed by radiologists in a blinded fashion to determine the realism of the hybrid images. It was found that the radiologists could not readily distinguish between real and virtual lesion images (area under the ROC curve was 0.55). This study provided evidence that the proposed mathematical lesion modeling framework could produce reasonably realistic lesion images.

Based on that result, two studies were conducted which demonstrated the utility of the lesion models. The first study used the modeling framework as a measurement tool to determine how dose and reconstruction algorithm affected the quantitative analysis of liver lesions, lung nodules, and renal stones in terms of their size, shape, attenuation, edge profile, and texture features. The same database of real lesion images used in the previous study was used for this study. That database contained images of the same patient at 2 dose levels (50% and 100%) along with 3 reconstruction algorithms from a GE 750HD CT system (GE Healthcare). The algorithms in question were FBP, Adaptive Statistical Iterative Reconstruction (ASiR), and Model-Based Iterative Reconstruction (MBIR). A total of 23 quantitative features were extracted from the lesions under each condition. It was found that both dose and reconstruction algorithm had a statistically significant effect on the feature measurements. In particular, radiation dose affected five, three, and four of the 23 features (related to lesion size, conspicuity, and pixel-value distribution) for liver lesions, lung nodules, and renal stones, respectively. MBIR significantly affected 9, 11, and 15 of the 23 features (including size, attenuation, and texture features) for liver lesions, lung nodules, and renal stones, respectively. Lesion texture was not significantly affected by radiation dose.

The second study demonstrating the utility of the lesion modeling framework focused on assessing detectability of very low-contrast liver lesions in abdominal imaging. Specifically, detectability was assessed as a function of dose and reconstruction algorithm. As part of a parallel clinical trial, images from 21 patients were collected at 6 dose levels per patient on a SOMATOM Flash scanner. Subtle liver lesion models (contrast = -15 HU) were inserted into the raw projection data from the patient scans. The projections were then reconstructed with FBP and SAFIRE (strength 5). Also, lesion-less images were reconstructed. Noise, contrast, CNR, and detectability index of an observer model (non-prewhitening matched filter) were assessed. It was found that SAFIRE reduced noise by 52%, reduced contrast by 12%, increased CNR by 87%. and increased detectability index by 65% compared to FBP. Further, a 2AFC human perception experiment was performed to assess the dose reduction potential of SAFIRE, which was found to be 22% compared to the standard of care dose.

In conclusion, this dissertation provides to the scientific community a series of new methodologies, phantoms, analysis techniques, and modeling tools that can be used to rigorously assess image quality from modern CT systems. Specifically, methods to properly evaluate iterative reconstruction have been developed and are expected to aid in the safe clinical implementation of dose reduction technologies.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

During the early Stuart period, England’s return to male monarchal rule resulted in the emergence of a political analogy that understood the authority of the monarch to be rooted in the “natural” authority of the father; consequently, the mother’s authoritative role within the family was repressed. As the literature of the period recognized, however, there would be no family unit for the father to lead without the words and bodies of women to make narratives of dynasty and legitimacy possible. Early modern discourse reveals that the reproductive roles of men and women, and the social hierarchies that grow out of them, are as much a matter of human design as of divine or natural law. Moreover, despite the attempts of James I and Charles I to strengthen royal patriarchal authority, the role of the monarch was repeatedly challenged on stage and in print even prior to the British Civil Wars and the 1649 beheading of Charles I. Texts produced at moments of political crisis reveal how women could uphold the legitimacy of familial and political hierarchies, but they also disclose patriarchy’s limits by representing “natural” male authority as depending in part on women’s discursive control over their bodies. Due to the epistemological instability of the female reproductive body, women play a privileged interpretive role in constructing patriarchal identities. The dearth of definitive knowledge about the female body during this period, and the consequent inability to fix or stabilize somatic meaning, led to the proliferation of differing, and frequently contradictory, depictions of women’s bodies. The female body became a site of contested meaning in early modern discourse, with men and women struggling for dominance, and competitors so diverse as to include kings, midwives, scholars of anatomy, and female religious sectarians. Essentially, this competition came down to a question of where to locate somatic meaning: In the opaque, uncertain bodies of women? In women’s equally uncertain and unreliable words? In the often contradictory claims of various male-authored medical treatises? In the whispered conversations that took place between women behind the closed doors of birthing rooms? My dissertation traces this representational instability through plays by William Shakespeare, John Ford, Thomas Middleton, and William Rowley, as well as in monstrous birth pamphlets, medical treatises, legal documents, histories, satires, and ballads. In these texts, the stories women tell about and through their bodies challenge and often supersede male epistemological control. These stories, which I term female bodily narratives, allow women to participate in defining patriarchal authority at the levels of both the family and the state. After laying out these controversies and instabilities surrounding early modern women’s bodies in my first chapter, my remaining chapters analyze the impact of women’s words on four distinct but overlapping reproductive issues: virginity, pregnancy, birthing room rituals, and paternity. In chapters 2 and 3, I reveal how women construct the inner, unseen “truths” of their reproductive bodies through speech and performance, and in doing so challenge the traditional forms of male authority that depend on these very constructions for coherence. Chapter 2 analyzes virginity in Thomas Middleton and William Rowley’s play The Changeling (1622) and in texts documenting the 1613 Essex divorce, during which Frances Howard, like Beatrice-Joanna in the play, was required to undergo a virginity test. These texts demonstrate that a woman’s ability to feign virginity could allow her to undermine patriarchal authority within the family and the state, even as they reveal how men relied on women to represent their reproductive bodies in socially stabilizing ways. During the British Civil Wars and Interregnum (1642-1660), Parliamentary writers used Howard as an example of how the unruly words and bodies of women could disrupt and transform state politics by influencing court faction; in doing so, they also revealed how female bodily narratives could help recast political historiography. In chapter 3, I investigate depictions of pregnancy in John Ford’s tragedy, ‘Tis Pity She’s a Whore (1633) and in early modern medical treatises from 1604 to 1651. Although medical texts claim to convey definitive knowledge about the female reproductive body, in actuality male knowledge frequently hinged on the ways women chose to interpret the unstable physical indicators of pregnancy. In Ford’s play, Annabella and Putana take advantage of male ignorance in order to conceal Annabella’s incestuous, illegitimate pregnancy from her father and husband, thus raising fears about women’s ability to misrepresent their bodies. Since medical treatises often frame the conception of healthy, legitimate offspring as a matter of national importance, women’s ability to conceal or even terminate their pregnancies could weaken both the patriarchal family and the patriarchal state that the family helped found. Chapters 4 and 5 broaden the socio-political ramifications of women’s words and bodies by demonstrating how female bodily narratives are required to establish paternity and legitimacy, and thus help shape patriarchal authority at multiple social levels. In chapter 4, I study representations of birthing room gossip in Thomas Middleton’s play, A Chaste Maid in Cheapside (1613), and in three Mistris Parliament pamphlets (1648) that satirize parliamentary power. Across these texts, women’s birthing room “gossip” comments on and critiques such issues as men’s behavior towards their wives and children, the proper use of household funds, the finer points of religious ritual, and even the limits of the authority of the monarch. The collective speech of the female-dominated birthing room thus proves central not only to attributing paternity to particular men, but also to the consequent definition and establishment of the political, socio-economic, and domestic roles of patriarchy. Chapter 5 examines anxieties about paternity in William Shakespeare’s The Winter’s Tale (1611) and in early modern monstrous birth pamphlets from 1600 to 1647, in which children born with congenital deformities are explained as God’s punishment for the sexual, religious, and/or political transgressions of their parents or communities. Both the play and the pamphlets explore the formative/deformative power of women’s words and bodies over their offspring, a power that could obscure a father’s connection to his children. However, although the pamphlets attempt to contain and discipline women’s unruly words and bodies with the force of male authority, the play reveals the dangers of male tyranny and the crucial role of maternal authority in reproducing and authenticating dynastic continuity and royal legitimacy. My emphasis on the socio-political impact of women’s self-representation distinguishes my work from that of scholars such as Mary Fissell and Julie Crawford, who claim that early modern beliefs about the female reproductive body influenced textual depictions of major religious and political events, but give little sustained attention to the role female speech plays in these representations. In contrast, my dissertation reveals that in such texts, patriarchal society relies precisely on the words women speak about their own and other women’s bodies. Ultimately, I argue that female bodily narratives were crucial in shaping early modern culture, and they are equally crucial to our critical understanding of sexual and state politics in the literature of the period.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Tämän kandidaatintutkielman tavoitteena on tunnistaa strategiaprosessin eri vaiheet sekä kuvailla, miten prosessi toteutettiin tapausyritys Suominen Oyj:llä. Tavoitteina on myös analysoida, miten Suominen Oyj:n henkilöstö kokee oman suhteensa organisaatioon, sen strategiaan, tavoitteisiin ja arvoihin. Analyysin pohjalta pyritään muodostamaan kuva siitä, kuinka strategian implementoinnissa on onnistuttu. Lisäksi tarkastellaan, eroavatko nämä kokemukset eri maissa tapausyrityksessä. Tutkimuksessa käytetään sekä kvalitatiivisia että kvantitatiivisia tutkimusmenetelmiä. Suominen Oyj:n strategiaprosessista hankitaan informaatiota kahdella laadullisella puolistrukturoidulla teemahaastattelulla. Kvantitatiivisen analyysin kohteena puolestaan on tapausyrityksen yhdeksältäkymmeneltä ylimmässä asemassa työskentelevältä henkilöltä kerätty tilastollinen kyselydata, jota analysoidaan muun muassa Kruskal-Wallisin testillä sekä Spearmanin korrelaatioanalyysillä. Tutkielmassa muodostetaan kuva prosessista, jossa hyvin muotoiltu strategia sekä selkeät tavoitteet luovat pohjan strategian implementoinnille. Onnistuneen implementoinnin keinoja ovat muun muassa kehitysohjelmat, tarkka rekrytointi, tavoitteiden ja palkkioiden sitominen strategiaan, strategian kanssa yhteensopiva organisaatiokulttuuri, useamman eri hierarkiatason osallistuminen ja erilaisten kontrollijärjestelmien sekä implementointityökalujen hyödyntäminen. Havaitaan, että tapausyritys käyttää näistä keinoista useampaa. Yrityksen monikansallisuus vaikuttaa erityisesti organisaatiokulttuuriin sekä sisäiseen viestintään. Tutkielman kvantitatiivisen analyysin tulokset näyttävät, ettei maiden välillä ole tilastollisesti merkittäviä eroja siinä, miten henkilöstö kokee oman suhteensa tapausyritykseen, sen strategiaan, tavoitteisiin ja arvoihin. Kvantitatiivisen analyysin perusteella Suominen Oyj:n strategiaprosessi on onnistunut hyvin. Tutkimuksessa esitellään vielä muutamia kehitysajatuksia yritystä sekä mahdollisia jatkotutkimuksia varten.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Weiss and Isen have provided many supportive comments about the multi-level perspective, but also found limitations. Isen noted the importance of integrating affect, cognition, and motivation. Weiss commented similarly that the model lacked an integrating “thread.” He suggested that, to be truly multilevel, each level should constrain processes at other levels, and also provide guidance for the development of new concepts. Weiss also noted that the focus on biological processes was a strength of the model. I respond by suggesting that these very biological processes may constitute the “missing” thread. To illustrate this, I discuss some of the recent research on emotions in organizational settings, and argue that biology both constrains and guides theory at each level of the model. Based on this proposition, I revisit each of the five levels in the model, to demonstrate how this integration can be accomplished in this fashion. Finally, I address two additional points: aggregation bias, and the possibility of extending the model to include higher levels of industry and region.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Brain derived neurotrophic factor (BDNF) has been related to neuroprotection in a series of central nervous system diseases, although its role in multiple sclerosis (MS) was only partially investigated. In this work, we aimed to evaluate the plasma levels of BDNF from 29 MS patients and 24 control subjects. MS patients had decreased levels of BDNF in comparison with healthy controls. BDNF levels increased significantly after MS relapse. Our results provide some evidence for the involvement of BDNF in the pathogenesis of MS and suggest a role for this neurotrophin during the recovery of acute demyelinating inflammatory lesion. (C) 2009 Elsevier Ireland Ltd. All rights reserved.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The populations of parasites and infectious agents are most of the time structured in complex hierarchy that lies beyond the classical nested design described by Wright's F-statistics (F(IS), F(ST) and F(IT)). In this note we propose a user-friendly step-by-step notice for using recent software (HierFstat) that computes and test fixation indices for any hierarchical structure. We add some tricks and tips for some special data kind (haploid, single locus), some other procedure (bootstrap over loci) and how to handle crossed factors.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Anti-angiogenic therapies are currently in cancer clinical trials, but to date there are no established tests for evaluating the angiogenic status of a patient. We measured 11 circulating angiogenesis-associated molecules in cancer patients before and after local treatment. The purpose of our study was to screen for possible relationships among the different molecules and between individual molecules and tumor burden. We measured VEGF-A, PlGF, SCF, MMP-9, EDB+ -fibronectin, sVEGFR-2, sVEGFR-1, salphaVbeta3, sTie-2, IL-8 and CRP in the blood of 22 healthy volunteers, 17 early breast, 17 early colorectal, and 8 advanced sarcoma/melanoma cancer patients. Breast cancer patients had elevated levels of VEGF-A and sTie-2, colorectal cancer patients of VEGF-A, MMP-9, sTie-2, IL-8 and CRP, and melanoma/sarcoma patients of sVEGFR-1. salphaVbeta3 was decreased in colorectal cancer patients. A correlation between VEGF-A and MMP-9 was found. After tumor removal, MMP-9 and salphaVbeta3 significantly decreased in breast and CRP in colorectal cancer, whereas sVEGFR-1 increased in colorectal cancer patients. In sarcoma/melanoma patients treated regionally with TNF and chemotherapy we observed a rise in VEGF-A, SCF, VEGFR-2, MMP-9, Tie-2 and CRP, a correlation between CRP and IL-8, and a decreased in sVEGFR-1 levels. In conclusion, among all factors measured, only VEGF-A and MMP-9 consistently correlated to each other, elevated CRP levels were associated with tumor burden, whereas sVEGF-R1 increased after tumor removal in colorectal cancer. Treatment with chemotherapy and TNF induced changes consistent with an angiogenic switch. These results warrant a prospective study to compare the effect of surgical tumor removal vs. chemotherapy on some of these markers and to evaluate their prognostic/predictive value.