843 resultados para Realism.


Relevância:

10.00% 10.00%

Publicador:

Resumo:

X-ray computed tomography (CT) imaging constitutes one of the most widely used diagnostic tools in radiology today with nearly 85 million CT examinations performed in the U.S in 2011. CT imparts a relatively high amount of radiation dose to the patient compared to other x-ray imaging modalities and as a result of this fact, coupled with its popularity, CT is currently the single largest source of medical radiation exposure to the U.S. population. For this reason, there is a critical need to optimize CT examinations such that the dose is minimized while the quality of the CT images is not degraded. This optimization can be difficult to achieve due to the relationship between dose and image quality. All things being held equal, reducing the dose degrades image quality and can impact the diagnostic value of the CT examination.

A recent push from the medical and scientific community towards using lower doses has spawned new dose reduction technologies such as automatic exposure control (i.e., tube current modulation) and iterative reconstruction algorithms. In theory, these technologies could allow for scanning at reduced doses while maintaining the image quality of the exam at an acceptable level. Therefore, there is a scientific need to establish the dose reduction potential of these new technologies in an objective and rigorous manner. Establishing these dose reduction potentials requires precise and clinically relevant metrics of CT image quality, as well as practical and efficient methodologies to measure such metrics on real CT systems. The currently established methodologies for assessing CT image quality are not appropriate to assess modern CT scanners that have implemented those aforementioned dose reduction technologies.

Thus the purpose of this doctoral project was to develop, assess, and implement new phantoms, image quality metrics, analysis techniques, and modeling tools that are appropriate for image quality assessment of modern clinical CT systems. The project developed image quality assessment methods in the context of three distinct paradigms, (a) uniform phantoms, (b) textured phantoms, and (c) clinical images.

The work in this dissertation used the “task-based” definition of image quality. That is, image quality was broadly defined as the effectiveness by which an image can be used for its intended task. Under this definition, any assessment of image quality requires three components: (1) A well defined imaging task (e.g., detection of subtle lesions), (2) an “observer” to perform the task (e.g., a radiologists or a detection algorithm), and (3) a way to measure the observer’s performance in completing the task at hand (e.g., detection sensitivity/specificity).

First, this task-based image quality paradigm was implemented using a novel multi-sized phantom platform (with uniform background) developed specifically to assess modern CT systems (Mercury Phantom, v3.0, Duke University). A comprehensive evaluation was performed on a state-of-the-art CT system (SOMATOM Definition Force, Siemens Healthcare) in terms of noise, resolution, and detectability as a function of patient size, dose, tube energy (i.e., kVp), automatic exposure control, and reconstruction algorithm (i.e., Filtered Back-Projection– FPB vs Advanced Modeled Iterative Reconstruction– ADMIRE). A mathematical observer model (i.e., computer detection algorithm) was implemented and used as the basis of image quality comparisons. It was found that image quality increased with increasing dose and decreasing phantom size. The CT system exhibited nonlinear noise and resolution properties, especially at very low-doses, large phantom sizes, and for low-contrast objects. Objective image quality metrics generally increased with increasing dose and ADMIRE strength, and with decreasing phantom size. The ADMIRE algorithm could offer comparable image quality at reduced doses or improved image quality at the same dose (increase in detectability index by up to 163% depending on iterative strength). The use of automatic exposure control resulted in more consistent image quality with changing phantom size.

Based on those results, the dose reduction potential of ADMIRE was further assessed specifically for the task of detecting small (<=6 mm) low-contrast (<=20 HU) lesions. A new low-contrast detectability phantom (with uniform background) was designed and fabricated using a multi-material 3D printer. The phantom was imaged at multiple dose levels and images were reconstructed with FBP and ADMIRE. Human perception experiments were performed to measure the detection accuracy from FBP and ADMIRE images. It was found that ADMIRE had equivalent performance to FBP at 56% less dose.

Using the same image data as the previous study, a number of different mathematical observer models were implemented to assess which models would result in image quality metrics that best correlated with human detection performance. The models included naïve simple metrics of image quality such as contrast-to-noise ratio (CNR) and more sophisticated observer models such as the non-prewhitening matched filter observer model family and the channelized Hotelling observer model family. It was found that non-prewhitening matched filter observers and the channelized Hotelling observers both correlated strongly with human performance. Conversely, CNR was found to not correlate strongly with human performance, especially when comparing different reconstruction algorithms.

The uniform background phantoms used in the previous studies provided a good first-order approximation of image quality. However, due to their simplicity and due to the complexity of iterative reconstruction algorithms, it is possible that such phantoms are not fully adequate to assess the clinical impact of iterative algorithms because patient images obviously do not have smooth uniform backgrounds. To test this hypothesis, two textured phantoms (classified as gross texture and fine texture) and a uniform phantom of similar size were built and imaged on a SOMATOM Flash scanner (Siemens Healthcare). Images were reconstructed using FBP and a Sinogram Affirmed Iterative Reconstruction (SAFIRE). Using an image subtraction technique, quantum noise was measured in all images of each phantom. It was found that in FBP, the noise was independent of the background (textured vs uniform). However, for SAFIRE, noise increased by up to 44% in the textured phantoms compared to the uniform phantom. As a result, the noise reduction from SAFIRE was found to be up to 66% in the uniform phantom but as low as 29% in the textured phantoms. Based on this result, it clear that further investigation was needed into to understand the impact that background texture has on image quality when iterative reconstruction algorithms are used.

To further investigate this phenomenon with more realistic textures, two anthropomorphic textured phantoms were designed to mimic lung vasculature and fatty soft tissue texture. The phantoms (along with a corresponding uniform phantom) were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Scans were repeated a total of 50 times in order to get ensemble statistics of the noise. A novel method of estimating the noise power spectrum (NPS) from irregularly shaped ROIs was developed. It was found that SAFIRE images had highly locally non-stationary noise patterns with pixels near edges having higher noise than pixels in more uniform regions. Compared to FBP, SAFIRE images had 60% less noise on average in uniform regions for edge pixels, noise was between 20% higher and 40% lower. The noise texture (i.e., NPS) was also highly dependent on the background texture for SAFIRE. Therefore, it was concluded that quantum noise properties in the uniform phantoms are not representative of those in patients for iterative reconstruction algorithms and texture should be considered when assessing image quality of iterative algorithms.

The move beyond just assessing noise properties in textured phantoms towards assessing detectability, a series of new phantoms were designed specifically to measure low-contrast detectability in the presence of background texture. The textures used were optimized to match the texture in the liver regions actual patient CT images using a genetic algorithm. The so called “Clustured Lumpy Background” texture synthesis framework was used to generate the modeled texture. Three textured phantoms and a corresponding uniform phantom were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Images were reconstructed with FBP and SAFIRE and analyzed using a multi-slice channelized Hotelling observer to measure detectability and the dose reduction potential of SAFIRE based on the uniform and textured phantoms. It was found that at the same dose, the improvement in detectability from SAFIRE (compared to FBP) was higher when measured in a uniform phantom compared to textured phantoms.

The final trajectory of this project aimed at developing methods to mathematically model lesions, as a means to help assess image quality directly from patient images. The mathematical modeling framework is first presented. The models describe a lesion’s morphology in terms of size, shape, contrast, and edge profile as an analytical equation. The models can be voxelized and inserted into patient images to create so-called “hybrid” images. These hybrid images can then be used to assess detectability or estimability with the advantage that the ground truth of the lesion morphology and location is known exactly. Based on this framework, a series of liver lesions, lung nodules, and kidney stones were modeled based on images of real lesions. The lesion models were virtually inserted into patient images to create a database of hybrid images to go along with the original database of real lesion images. ROI images from each database were assessed by radiologists in a blinded fashion to determine the realism of the hybrid images. It was found that the radiologists could not readily distinguish between real and virtual lesion images (area under the ROC curve was 0.55). This study provided evidence that the proposed mathematical lesion modeling framework could produce reasonably realistic lesion images.

Based on that result, two studies were conducted which demonstrated the utility of the lesion models. The first study used the modeling framework as a measurement tool to determine how dose and reconstruction algorithm affected the quantitative analysis of liver lesions, lung nodules, and renal stones in terms of their size, shape, attenuation, edge profile, and texture features. The same database of real lesion images used in the previous study was used for this study. That database contained images of the same patient at 2 dose levels (50% and 100%) along with 3 reconstruction algorithms from a GE 750HD CT system (GE Healthcare). The algorithms in question were FBP, Adaptive Statistical Iterative Reconstruction (ASiR), and Model-Based Iterative Reconstruction (MBIR). A total of 23 quantitative features were extracted from the lesions under each condition. It was found that both dose and reconstruction algorithm had a statistically significant effect on the feature measurements. In particular, radiation dose affected five, three, and four of the 23 features (related to lesion size, conspicuity, and pixel-value distribution) for liver lesions, lung nodules, and renal stones, respectively. MBIR significantly affected 9, 11, and 15 of the 23 features (including size, attenuation, and texture features) for liver lesions, lung nodules, and renal stones, respectively. Lesion texture was not significantly affected by radiation dose.

The second study demonstrating the utility of the lesion modeling framework focused on assessing detectability of very low-contrast liver lesions in abdominal imaging. Specifically, detectability was assessed as a function of dose and reconstruction algorithm. As part of a parallel clinical trial, images from 21 patients were collected at 6 dose levels per patient on a SOMATOM Flash scanner. Subtle liver lesion models (contrast = -15 HU) were inserted into the raw projection data from the patient scans. The projections were then reconstructed with FBP and SAFIRE (strength 5). Also, lesion-less images were reconstructed. Noise, contrast, CNR, and detectability index of an observer model (non-prewhitening matched filter) were assessed. It was found that SAFIRE reduced noise by 52%, reduced contrast by 12%, increased CNR by 87%. and increased detectability index by 65% compared to FBP. Further, a 2AFC human perception experiment was performed to assess the dose reduction potential of SAFIRE, which was found to be 22% compared to the standard of care dose.

In conclusion, this dissertation provides to the scientific community a series of new methodologies, phantoms, analysis techniques, and modeling tools that can be used to rigorously assess image quality from modern CT systems. Specifically, methods to properly evaluate iterative reconstruction have been developed and are expected to aid in the safe clinical implementation of dose reduction technologies.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

X-ray computed tomography (CT) is a non-invasive medical imaging technique that generates cross-sectional images by acquiring attenuation-based projection measurements at multiple angles. Since its first introduction in the 1970s, substantial technical improvements have led to the expanding use of CT in clinical examinations. CT has become an indispensable imaging modality for the diagnosis of a wide array of diseases in both pediatric and adult populations [1, 2]. Currently, approximately 272 million CT examinations are performed annually worldwide, with nearly 85 million of these in the United States alone [3]. Although this trend has decelerated in recent years, CT usage is still expected to increase mainly due to advanced technologies such as multi-energy [4], photon counting [5], and cone-beam CT [6].

Despite the significant clinical benefits, concerns have been raised regarding the population-based radiation dose associated with CT examinations [7]. From 1980 to 2006, the effective dose from medical diagnostic procedures rose six-fold, with CT contributing to almost half of the total dose from medical exposure [8]. For each patient, the risk associated with a single CT examination is likely to be minimal. However, the relatively large population-based radiation level has led to enormous efforts among the community to manage and optimize the CT dose.

As promoted by the international campaigns Image Gently and Image Wisely, exposure to CT radiation should be appropriate and safe [9, 10]. It is thus a responsibility to optimize the amount of radiation dose for CT examinations. The key for dose optimization is to determine the minimum amount of radiation dose that achieves the targeted image quality [11]. Based on such principle, dose optimization would significantly benefit from effective metrics to characterize radiation dose and image quality for a CT exam. Moreover, if accurate predictions of the radiation dose and image quality were possible before the initiation of the exam, it would be feasible to personalize it by adjusting the scanning parameters to achieve a desired level of image quality. The purpose of this thesis is to design and validate models to quantify patient-specific radiation dose prospectively and task-based image quality. The dual aim of the study is to implement the theoretical models into clinical practice by developing an organ-based dose monitoring system and an image-based noise addition software for protocol optimization.

More specifically, Chapter 3 aims to develop an organ dose-prediction method for CT examinations of the body under constant tube current condition. The study effectively modeled the anatomical diversity and complexity using a large number of patient models with representative age, size, and gender distribution. The dependence of organ dose coefficients on patient size and scanner models was further evaluated. Distinct from prior work, these studies use the largest number of patient models to date with representative age, weight percentile, and body mass index (BMI) range.

With effective quantification of organ dose under constant tube current condition, Chapter 4 aims to extend the organ dose prediction system to tube current modulated (TCM) CT examinations. The prediction, applied to chest and abdominopelvic exams, was achieved by combining a convolution-based estimation technique that quantifies the radiation field, a TCM scheme that emulates modulation profiles from major CT vendors, and a library of computational phantoms with representative sizes, ages, and genders. The prospective quantification model is validated by comparing the predicted organ dose with the dose estimated based on Monte Carlo simulations with TCM function explicitly modeled.

Chapter 5 aims to implement the organ dose-estimation framework in clinical practice to develop an organ dose-monitoring program based on a commercial software (Dose Watch, GE Healthcare, Waukesha, WI). In the first phase of the study we focused on body CT examinations, and so the patient’s major body landmark information was extracted from the patient scout image in order to match clinical patients against a computational phantom in the library. The organ dose coefficients were estimated based on CT protocol and patient size as reported in Chapter 3. The exam CTDIvol, DLP, and TCM profiles were extracted and used to quantify the radiation field using the convolution technique proposed in Chapter 4.

With effective methods to predict and monitor organ dose, Chapters 6 aims to develop and validate improved measurement techniques for image quality assessment. Chapter 6 outlines the method that was developed to assess and predict quantum noise in clinical body CT images. Compared with previous phantom-based studies, this study accurately assessed the quantum noise in clinical images and further validated the correspondence between phantom-based measurements and the expected clinical image quality as a function of patient size and scanner attributes.

Chapter 7 aims to develop a practical strategy to generate hybrid CT images and assess the impact of dose reduction on diagnostic confidence for the diagnosis of acute pancreatitis. The general strategy is (1) to simulate synthetic CT images at multiple reduced-dose levels from clinical datasets using an image-based noise addition technique; (2) to develop quantitative and observer-based methods to validate the realism of simulated low-dose images; (3) to perform multi-reader observer studies on the low-dose image series to assess the impact of dose reduction on the diagnostic confidence for multiple diagnostic tasks; and (4) to determine the dose operating point for clinical CT examinations based on the minimum diagnostic performance to achieve protocol optimization.

Chapter 8 concludes the thesis with a summary of accomplished work and a discussion about future research.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Notre étude porte le western crépusculaire et cherche plus précisément à extraire le « crépusculaire » du genre. L'épithète « crépusculaire », héritée du vocabulaire critique des années 1960 et 1970, définit généralement un nombre relativement restreint d'œuvres dont le récit met en scène des cowboys vieillissants dans un style qui privilégie un réalisme esthétique et psychologique, fréquemment associé à un révisionnisme historique, voire au « western pro-indien », mais qui se démarque par sa propension à filmer des protagonistes fatigués et dépassés par la marche de l'Histoire. Par un détour sur les formes littéraires ayant comme contexte diégétique l’Ouest américain (dime-novel et romans de la frontière), nous effectuons des allers et retours entre les formes épique et romanesque, entre l’Histoire et son mythe, entre le littéraire et le filmique pour mieux saisir la relation dyadique qu’entretient le western avec l’écriture, d’une part monumentale et d’autre part critique, de l’Histoire. Moins intéressée à l’esthétique des images qu’aux aspects narratologiques du film pris comme texte, notre approche tire profit des analyses littéraires pour remettre en cause les classifications étanches qui ont marqué l’évolution du western cinématographique. Nous étudions, à partir des intuitions d’André Bazin au sujet du sur-western, les modulations narratives du western ainsi que l’émergence d’une conscience critique à partir de ses héros mythologiques (notamment le cow-boy). Notre approche est à la fois épistémologique et transhistorique en ce qu’elle cherche à dégager du western crépusculaire un genre au-delà des genres, fondé sur une incitation à la narrativisation crépusculaire de la part du spectateur. Cette dernière, concentrée par une approche deleuzienne de l’image-cristal, renvoie non plus seulement à une conception existentialiste du personnage dans l’Histoire, mais aussi à une mise en relief pointue du hors-cadre du cinéma, moment de clairvoyance à la fois pragmatique et historicisant que nous définissons comme une image-fin, une image chronogénétique relevant de la contemporanéité de ses figures et de leurs auteurs.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

La question du développement en Afrique a été perçue, dès 1962, par l’agronome René Dumont, comme le fait d’une occasion perdue. Mais peu à peu, ce diagnostic lucide a fait l’objet d’une vulgate colportée jusque dans les salons diplomatiques et enceintes officielles du développement international, de Washington à Paris en passant par Londres et Ottawa. Sujette à toutes les complaisances, la critique du développement international est elle-même devenue une « occasion perdue ». Mais en remontant la piste historique, on constate que les séquelles postcoloniales, tant politiques qu’économiques, ont façonné le contexte dans lequel les Africains se sont vus durablement privés de l’occasion de prendre en main leur destin. C’est ainsi que la structure économique extravertie et le poids d’une dette insolvable ont projeté plus profondément encore les pays africains dans la dépendance et la tutelle au monde industrialisé, en particulier à travers les programmes néolibéraux de la Banque mondiale, quelle que soit la mouture sous laquelle ils s’affichent (ajustement structurel ou lutte à la pauvreté). Dans cette veine, le critère de sélectivité dans l’aide publique au développement, mis en avant par l’institution internationale, et qu’adopte notamment le Canada, ouvre la porte aux abus de toutes sortes que commande le réalisme politique.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

OUT OF VIEW is a collection of stories set in the American Southwest about people coping with loss—the death of parents, children, ideals, innocence. The characters in this collection reap or resist lessons of life as they struggle to find their place in the world. In “First Rain,” 15-year-old Tessie struggles with the loss of her father and the demands of her mother as she navigates the rocky terrain of adolescence. In “Monsters,” middle-aged Maury has to choose between a new relationship and protecting the well-being of his 4-year-old ‘daughter.’ The stories are influenced by the Western realism of Maile Meloy and the playful plotting of Ron Carlson. These stories are inspired both by the Sonoran Desert—expansive, sun-soaked, unrepentant—and by the people who live, love, and lose in the interstices between Manifest Destiny and the Reconquista.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Network simulation is an indispensable tool for studying Internet-scale networks due to the heterogeneous structure, immense size and changing properties. It is crucial for network simulators to generate representative traffic, which is necessary for effectively evaluating next-generation network protocols and applications. With network simulation, we can make a distinction between foreground traffic, which is generated by the target applications the researchers intend to study and therefore must be simulated with high fidelity, and background traffic, which represents the network traffic that is generated by other applications and does not require significant accuracy. The background traffic has a significant impact on the foreground traffic, since it competes with the foreground traffic for network resources and therefore can drastically affect the behavior of the applications that produce the foreground traffic. This dissertation aims to provide a solution to meaningfully generate background traffic in three aspects. First is realism. Realistic traffic characterization plays an important role in determining the correct outcome of the simulation studies. This work starts from enhancing an existing fluid background traffic model by removing its two unrealistic assumptions. The improved model can correctly reflect the network conditions in the reverse direction of the data traffic and can reproduce the traffic burstiness observed from measurements. Second is scalability. The trade-off between accuracy and scalability is a constant theme in background traffic modeling. This work presents a fast rate-based TCP (RTCP) traffic model, which originally used analytical models to represent TCP congestion control behavior. This model outperforms other existing traffic models in that it can correctly capture the overall TCP behavior and achieve a speedup of more than two orders of magnitude over the corresponding packet-oriented simulation. Third is network-wide traffic generation. Regardless of how detailed or scalable the models are, they mainly focus on how to generate traffic on one single link, which cannot be extended easily to studies of more complicated network scenarios. This work presents a cluster-based spatio-temporal background traffic generation model that considers spatial and temporal traffic characteristics as well as their correlations. The resulting model can be used effectively for the evaluation work in network studies.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Countering the trend in contemporary ecocriticism to advance realism as an environmentally responsible mode of representation, this essay argues that the anti-realist aesthetics of literary modernism were implicitly “ecological.” In order to make this argument I distinguish between contemporary and modernist ecological culture (both of which I differentiate in turn from ecological science); while the former is concerned primarily with the practical reform characteristic of what we now call “environmentalism,” the latter demanded an all-encompassing reimagination of the relationship between humanity and nature. “Modernist ecology,” as I call it, attempted to envision this change, which would be ontological or metaphysical rather than simply social, through thematically and formally experimental works of art. Its radical vision, suggestive in some ways of today’s “deep” ecology, repudiated modern accounts of nature as a congeries of inert objects to be manipulated by a sovereign subject, and instead foregrounded the chiasmic intertexture of the subject/object relationship. In aesthetic modernism we encounter not “objective” nature, but “nature-being” – a blank substratum beneath the solid contours of what philosopher Kate Soper calls “lay nature” – the revelation of which shatters historical constructions of nature and alone allows for radical alternatives. This essay looks specifically at modernist ecology as it appears in the works of W. B. Yeats, D. H. Lawrence, and Samuel Beckett, detailing their attempts to envision revolutionary new ecologies, but also their struggles with the limited capacity of esoteric modernist art to effect significant ecological change on a collective level.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This dissertation examines the corpse as an object in and of American hardboiled detective fiction written between 1920 and 1950. I deploy several theoretical frames, including narratology, body-as-text theory, object relations theory, and genre theory, in order to demonstrate the significance of objects, symbols, and things primarily in the clever and crafty work of Dashiell Hammett (1894-1961) and Raymond Chandler (1888-1959), but also touching on the writings of their lesser known accomplices. I construct a literary genealogy of American hardboiled detective fiction originating in the writings of Edgar Allan Poe, compare the contributions of classic or Golden Age detective fiction in England, and describe the socio-economic contexts, particularly the predominance of the “pulps,” that gave birth to the realism of the Hardboiled School. Taking seriously Chandler’s obsession with the art of murder, I engage with how authors pre-empt their readers’ knowledge of the tricks of the trade and manipulate their expectations, as well as discuss the characteristics and effect of the inimitable hardboiled style, its sharpshooting language and deadpan humour. Critical scholarship has rarely addressed the body and figure of the corpse, preferring to focus instead on the machinations of the femme fatale, the performance of masculinity, or the prevalence of violence. I cast new light on the world of hardboiled detective fiction by dissecting the corpse as the object that both motivates and de-composes (or rots away from) the narrative that makes it signify. I treat the corpse as an inanimate object, indifferent to representation, that destabilizes the integrity and self-possession, as well as the ratiocination, of the detective who authors the narrative of how the corpse came to be. The corpse is all deceptive and dangerous surface rather than the container of hidden depths of life and meaning that the detective hopes to uncover and reconstruct. I conclude with a chapter that is both critical denouement and creative writing experiment to reveal the self-reflexive (and at times metafictional) dimensions of hardboiled fiction. My dissertation, too, in the manner of hardboiled fiction, hopes to incriminate my readers as much as enlighten them.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this article I explore how the figure of debt illuminates the racial politics of welfare in neoliberal Britain. I begin by giving a reading of the simultaneous unfolding of post-war race politics and the Beveridgean welfare state, and then turn to consider the interpellative appeal of neoliberal debt to minoritiSed subjects who have, in certain respects, been de facto excluded from prevailing models of welfare citizenship. In particular, this article considers the ways in which household debt might, even as it increases social inequality, simultaneously produce ideas about equality and futurity, as well as gesture towards the possibility of post-national forms of identity and belonging. If we are to challenge the lowest-common-denominator logics of ‘capitalist realism it is necessary to develop orientations to the economic that are as convincing as the popular stories that circulate about the operations of the neoliberal marketplace, and which are as meaningful as the social relations they play a part in constituting. Rather than reproduce the racialized model of welfare citizenship that is implicit to the ‘defence’ of the postwar welfare state, I suggest that there are elements of prevailing neoliberal market relations that might themselves serve as a more substantial basis for expressions of racial equality. There is, in other words, something that we can learn from neoliberal debt regimes in order to develop a more egalitarian future-oriented politics of social welfare and economic redistribution.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work examines independence in the Canadian justice system using an approach adapted from new legal realist scholarship called ‘dynamic realism. This approach proposes that issues in law must be considered in relation to their recursive and simultaneous development with historic, social and political events. Such events describe ‘law in action’ and more holistically demonstrate principles like independence, rule of law and access to justice. My dynamic realist analysis of independence in the justice system employs a range methodological tools and approaches from the social sciences, including: historical and historiographical study; public administrative; policy and institutional analysis; an empirical component; as well as constitutional, statutory interpretation and jurisprudential analysis. In my view, principles like independence represent aspirational ideals in law which can be better understood by examining how they manifest in legal culture and in the legal system. This examination focuses on the principle and practice of independence for both lawyers and judges in the justice system, but highlights the independence of the Bar. It considers the inter-relation between lawyer independence and the ongoing refinement of judicial independence in Canadian law. It also considers both independence of the Bar and the Judiciary in the context of the administration of justice, and practically illustrates the interaction between these principles through a case study of a specific aspect of the court system. This work also focuses on recent developments in the principle of Bar independence and its relation to an emerging school of professionalism scholarship in Canada. The work concludes by describing the principle of independence as both conditional and dynamic, but rooted in a unitary concept for both lawyers and judges. In short, independence can be defined as impartiality, neutrality and autonomy of legal decision-makers in the justice system to apply, protect and improve the law for what has become its primary normative purpose: facilitating access to justice. While both independence of the Bar and the Judiciary are required to support access to independent courts, some recent developments suggest the practical interactions between independence and access need to be the subject of further research, to better account for both the principles and the practicalities of the Canadian justice system.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este artículo tiene como objetivo fundamental mostrar cómo el Instituto de Investigaciones y Experiencias Cinematográficas (IIEC) fue, desde finales de los años cuarenta hasta mediados de los años sesenta, el creador de una escuela de maestros de la luz y cómo la enseñanza de estos nuevos métodos de iluminación traspasó las fronteras de las aulas y marcó el inicio del Nuevo Cine Español. Para ello se han revisado las principales fuentes bibliográficas de los que consideramos fueron los primeros maestros de fotografía que dio el IIEC: Juan Julio Baena, Luis Enrique Torán y Luis Cuadrado, y se ha realizado un aná­lisis fílmico de la primera obra audiovisual de cada uno en la que se instaura un ambiente de renovación, un cambio estético y el uso de unas técnicas fotográficas más realistas. Así mismo, hemos observado cómo la nueva docencia de la fotografía fue trasladada a otras personas que también llegaron a ser maestros y no estudiaron en el IIEC. Por tanto, se trata de una escuela generadora de maestros tanto dentro como fuera de sus aulas.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Die neuen Paradigmen der Rechtswissenschaft nötigen uns beim Erwerb juristischen Wissens dazu, uns in eine kreativere und problemorientiertere Dynamik des Lehrens-Lernens einzugliedern. Die neuen Generationen der Rechtsgelehrten sollten darum bemüht sein, die verschiedenen theoretisch-methodologischen Strömungen des Rechts, z.B. des juristischen Realismus, den juristischen Strukturalismus sowie die angewandte juristische Soziologie zu studieren und zu analysieren. Andererseits sollten sie nicht dogmatisch und ausschliesslich den juristischen Positivismus des 19. Jahrhunderts vertreten. Die juristischen Fakultäten und Schulen der Republik Mexiko zeichnen sich dadurch aus, dass sie die traditionelle juristische Lehre bewahren. Darin liegt der Grund, dass es nur wenige Juristen gibt, die sich der juristischen Forschung und der wissenschaftlichen Praxis widmen.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este artículo introduce los conceptos de relaciones de poder analizados por Michel Foucault, haciendo hincapié en el vínculo del ejercicio del poder en las relaciones internacionales. Por ello, se analizan el marco de relaciones que el ejercicio de dicho poder genera, así como las tipologías de relaciones entre Estados, que igualmente genera el poder. El artículo presenta los paradigmas de las relaciones internacionales y su evolución en el tiempo, estableciendo puentes con las consideraciones de Foucault con respecto al poder, la guerra, la represión o las razas. En este sentido, el concepto de guerra de razas se antoja fundamental para identificar dicha noción como la versión más extrema del paradigma realista. Dicha guerra de razas no constituye una categoría que integre necesariamente el concepto biológico, sino que se refiere al proceso de imposición de la identidad del grupo dominante, la cual puede contener aspectos étnicos, socioculturales y políticos. No obstante, ni el realismo ni su visión extrema son los únicos paradigmas existentes, ya que la evolución de los paradigmas alternativos ha conllevado la génesis del reflectivismo como alternativa al realismo.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper deals with the place of narrative, that is, storytelling, in public deliberation. A distinction is made between weak and strong conceptions of narrative. According to the weak one, storytelling is but one rhetorical device among others with which social actors produce and convey meaning. In contrast, the strong conception holds that narrative is necessary to communicate, and argue, about topics such as the human experience of time, collective identities and the moral and ethical validity of values. The upshot of this idea is that storytelling should be a necessary component of any ideal of public deliberation. Contrary to recent work by deliberative theorists, who tend to adopt the weak conception of narrative, the author argues for embracing the strong one. The main contention of this article is that stories not only have a legitimate place in deliberation, but are even necessary to formulate certain arguments in the fi rst place; for instance, arguments drawing on historical experience. This claim, namely that narrative is constitutive of certain arguments, in the sense that, without it, said reasons cannot be articulated, is illustrated by deliberative theory’s own narrative underpinnings. Finally, certain possible objections against the strong conception of narrative are dispelled.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This article focuses on the analysis of the concept of love in the religious philosophy of Pavel Florensky, who shares the ontological approach to the consideration of love with other representatives of Russian religious philosophy (N. berdyaev and S. bulgakov). We pay more careful attention to the understanding of love-άγαπαν by Florensky. We have drawn the conclusion that, in the philosophy of P. Florensky, Love, closely connected with truth and beauty, is considered an ontological basis existence of personality. We develop the ideas of Pavel Florensky, and accordingly assume that it is possible to synthesise love-agape and love-eros around the idea of sacrificial love. Agapelogical and erotical ‘bezels’ of one jewel of love is aspects of united love, which is given by God. this gift of God, the gift of united love, is kept by humans through prayer and deeds of love.