842 resultados para performance-based engineering
Resumo:
In a large number of problems the high dimensionality of the search space, the vast number of variables and the economical constrains limit the ability of classical techniques to reach the optimum of a function, known or unknown. In this thesis we investigate the possibility to combine approaches from advanced statistics and optimization algorithms in such a way to better explore the combinatorial search space and to increase the performance of the approaches. To this purpose we propose two methods: (i) Model Based Ant Colony Design and (ii) Naïve Bayes Ant Colony Optimization. We test the performance of the two proposed solutions on a simulation study and we apply the novel techniques on an appplication in the field of Enzyme Engineering and Design.
Resumo:
To continuously improve the performance of metal-oxide-semiconductor field-effect-transistors (MOSFETs), innovative device architectures, gate stack engineering and mobility enhancement techniques are under investigation. In this framework, new physics-based models for Technology Computer-Aided-Design (TCAD) simulation tools are needed to accurately predict the performance of upcoming nanoscale devices and to provide guidelines for their optimization. In this thesis, advanced physically-based mobility models for ultrathin body (UTB) devices with either planar or vertical architectures such as single-gate silicon-on-insulator (SOI) field-effect transistors (FETs), double-gate FETs, FinFETs and silicon nanowire FETs, integrating strain technology and high-κ gate stacks are presented. The effective mobility of the two-dimensional electron/hole gas in a UTB FETs channel is calculated taking into account its tensorial nature and the quantization effects. All the scattering events relevant for thin silicon films and for high-κ dielectrics and metal gates have been addressed and modeled for UTB FETs on differently oriented substrates. The effects of mechanical stress on (100) and (110) silicon band structures have been modeled for a generic stress configuration. Performance will also derive from heterogeneity, coming from the increasing diversity of functions integrated on complementary metal-oxide-semiconductor (CMOS) platforms. For example, new architectural concepts are of interest not only to extend the FET scaling process, but also to develop innovative sensor applications. Benefiting from properties like large surface-to-volume ratio and extreme sensitivity to surface modifications, silicon-nanowire-based sensors are gaining special attention in research. In this thesis, a comprehensive analysis of the physical effects playing a role in the detection of gas molecules is carried out by TCAD simulations combined with interface characterization techniques. The complex interaction of charge transport in silicon nanowires of different dimensions with interface trap states and remote charges is addressed to correctly reproduce experimental results of recently fabricated gas nanosensors.
Resumo:
Web is constantly evolving, thanks to the 2.0 transition, HTML5 new features and the coming of cloud-computing, the gap between Web and traditional desktop applications is tailing off. Web-apps are more and more widespread and bring several benefits compared to traditional ones. On the other hand reference technologies, JavaScript primarly, are not keeping pace, so a paradim shift is taking place in Web programming, and so many new languages and technologies are coming out. First objective of this thesis is to survey the reference and state-of-art technologies for client-side Web programming focusing in particular on what concerns concurrency and asynchronous programming. Taking into account the problems that affect existing technologies, we finally design simpAL-web, an innovative approach to tackle Web-apps development, based on the Agent-oriented programming abstraction and the simpAL language. == Versione in italiano: Il Web è in continua evoluzione, grazie alla transizione verso il 2.0, alle nuove funzionalità introdotte con HTML5 ed all’avvento del cloud-computing, il divario tra le applicazioni Web e quelle desktop tradizionali va assottigliandosi. Le Web-apps sono sempre più diffuse e presentano diversi vantaggi rispetto a quelle tradizionali. D’altra parte le tecnologie di riferimento, JavaScript in primis, non stanno tenendo il passo, motivo per cui la programmazione Web sta andando incontro ad un cambio di paradigma e nuovi linguaggi e tecnologie stanno spuntando sempre più numerosi. Primo obiettivo di questa tesi è di passare al vaglio le tecnologie di riferimento ed allo stato dell’arte per quel che riguarda la programmmazione Web client-side, porgendo particolare attenzione agli aspetti inerenti la concorrenza e la programmazione asincrona. Considerando i principali problemi di cui soffrono le attuali tecnologie passeremo infine alla progettazione di simpAL-web, un approccio innovativo con cui affrontare lo sviluppo di Web-apps basato sulla programmazione orientata agli Agenti e sul linguaggio simpAL.
Resumo:
It is usual to hear a strange short sentence: «Random is better than...». Why is randomness a good solution to a certain engineering problem? There are many possible answers, and all of them are related to the considered topic. In this thesis I will discuss about two crucial topics that take advantage by randomizing some waveforms involved in signals manipulations. In particular, advantages are guaranteed by shaping the second order statistic of antipodal sequences involved in an intermediate signal processing stages. The first topic is in the area of analog-to-digital conversion, and it is named Compressive Sensing (CS). CS is a novel paradigm in signal processing that tries to merge signal acquisition and compression at the same time. Consequently it allows to direct acquire a signal in a compressed form. In this thesis, after an ample description of the CS methodology and its related architectures, I will present a new approach that tries to achieve high compression by design the second order statistics of a set of additional waveforms involved in the signal acquisition/compression stage. The second topic addressed in this thesis is in the area of communication system, in particular I focused the attention on ultra-wideband (UWB) systems. An option to produce and decode UWB signals is direct-sequence spreading with multiple access based on code division (DS-CDMA). Focusing on this methodology, I will address the coexistence of a DS-CDMA system with a narrowband interferer. To do so, I minimize the joint effect of both multiple access (MAI) and narrowband (NBI) interference on a simple matched filter receiver. I will show that, when spreading sequence statistical properties are suitably designed, performance improvements are possible with respect to a system exploiting chaos-based sequences minimizing MAI only.
Resumo:
Copper(I) halide clusters are recently considered as good candidate for optoelectronic devices such as OLEDs . Although the copper halide clusters, in particular copper iodide, are very well known since the beginning of the 20th century, only in the late ‘70s the interest on these compounds grew dramatically due their particular photophysical behaviour. These complexes are characterized by a dual triplet emission bands, named Cluster Centred (3CC) and Halogen-to-Ligand charge transfer (3XLCT), the intensities of which are strictly related with the temperature. The CC transition, due to the presence of a metallophylic interactions, is prevalent at ambient temperature while the XLCT transition, located preferentially on the ligand part, became more prominent at low temperature. Since these pioneering works, it was easy to understand the photophysical properties of this compounds became more interesting in solid-state respect to solution with an improvement in emission efficiency. In this work we aim to characterize in SS organocopper(I)iodide compounds to valuate the correlation between the molecular crystal structure and the photophysical properties. It is also considered to hike new strategies to synthesize CuI complexes from the wet reactions to the more green solvent free methods. The advantages in using these strategies are evident but, obtain a single crystal suitable for SCXRD analysis from these batches is quite impossible. The structure solution still remains the key point in this research so we tackle this problem solving the structure by X-ray powder diffraction data. When the sample was fully characterized we moved to design and development of the associated OLED-device. Since copper iodide complexes are often insoluble in organic solvents, the high vacuum deposition technique is preferred. A new non-conventional deposition process have also been proposed to avoid the low complex stability in this practice with an in-situ complex formation in a layer-by layer deposition route.
Resumo:
Laser Shock Peening (LSP) is a surface enhancement treatment which induces a significant layer of beneficial compressive residual stresses of up to several mm underneath the surface of metal components in order to improve the detrimental effects of the crack growth behavior rate in it. The aim of this thesis is to predict the crack growth behavior in metallic specimens with one or more stripes which define the compressive residual stress area induced by the Laser Shock Peening treatment. The process was applied as crack retardation stripes perpendicular to the crack propagation direction with the object of slowing down the crack when approaching the peened stripes. The finite element method has been applied to simulate the redistribution of stresses in a cracked model when it is subjected to a tension load and to a compressive residual stress field, and to evaluate the Stress Intensity Factor (SIF) in this condition. Finally, the Afgrow software is used to predict the crack growth behavior of the component following the Laser Shock Peening treatment and to detect the improvement in the fatigue life comparing it to the baseline specimen. An educational internship at the “Research & Technologies Germany – Hamburg” department of AIRBUS helped to achieve knowledge and experience to write this thesis. The main tasks of the thesis are the following: •To up to date Literature Survey related to “Laser Shock Peening in Metallic Structures” •To validate the FE model developed against experimental measurements at coupon level •To develop design of crack growth slowdown in Centered Cracked Tension specimens based on residual stress engineering approach using laser peened strip transversal to the crack path •To evaluate the Stress Intensity Factor values for Centered Cracked Tension specimens after the Laser Shock Peening treatment via Finite Element Analysis •To predict the crack growth behavior in Centered Cracked Tension specimens using as input the SIF values evaluated with the FE simulations •To validate the results by means of experimental tests
Resumo:
The functionalization of substrates through the application of nanostructured coatings allows to create new materials, with enhanced properties. In this work, the development of self-cleaning and antibacterial textiles, through the application of TiO2 and Ag based nanostructured coatings was carried out. The production of TiO2 and Ag functionalized materials was achieved both by the classical dip-padding-curing method and by the innovative electrospinning process to obtain nanofibers doped with nano-TiO2 and nano-Ag. In order to optimize the production of functionalized textiles, the study focused on the comprehension of mechanisms involved in the photocatalytic and antibacterial processes and on the real applicability of the products. In particular, a deep investigation on the relationship between nanosol physicochemical characteristics, nanocoating properties and their performances was accomplished. Self-cleaning textiles with optimized properties were obtained by properly purifying and applying commercial TiO2 nanosol while the studies on the photocatalytic mechanism operating in self-cleaning application demonstrated the strong influence of hydrophilic properties and of interaction surface/radicals on final performance. Moreover, a study about the safety in handling of nano-TiO2 was carried out and risk remediation strategies, based on “safety by design” approach, were developed. In particular, the coating of TiO2 nanoparticles by a SiO2 shell was demonstrated to be the best risk remediation strategy in term of biological response and preserving of photoreactivity. The obtained results were confirmed determining the reactive oxygen species production by a multiple approach. Antibacterial textiles for biotechnological applications were also studied and Ag-coated cotton materials, with significant anti-bacterial properties, were produced. Finally, composite nanofibers were obtained merging biopolymer processing and sol-gel techniques. Indeed, electrospun nanofibers embedded with TiO2 and Ag NPs, starting from aqueous keratin based formulation were produced and the photocatalytic and antibacterial properties were assessed. The results confirmed the capability of electrospun keratin nanofibers matrix to preserve nanoparticle properties.
Resumo:
Self-organising pervasive ecosystems of devices are set to become a major vehicle for delivering infrastructure and end-user services. The inherent complexity of such systems poses new challenges to those who want to dominate it by applying the principles of engineering. The recent growth in number and distribution of devices with decent computational and communicational abilities, that suddenly accelerated with the massive diffusion of smartphones and tablets, is delivering a world with a much higher density of devices in space. Also, communication technologies seem to be focussing on short-range device-to-device (P2P) interactions, with technologies such as Bluetooth and Near-Field Communication gaining greater adoption. Locality and situatedness become key to providing the best possible experience to users, and the classic model of a centralised, enormously powerful server gathering and processing data becomes less and less efficient with device density. Accomplishing complex global tasks without a centralised controller responsible of aggregating data, however, is a challenging task. In particular, there is a local-to-global issue that makes the application of engineering principles challenging at least: designing device-local programs that, through interaction, guarantee a certain global service level. In this thesis, we first analyse the state of the art in coordination systems, then motivate the work by describing the main issues of pre-existing tools and practices and identifying the improvements that would benefit the design of such complex software ecosystems. The contribution can be divided in three main branches. First, we introduce a novel simulation toolchain for pervasive ecosystems, designed for allowing good expressiveness still retaining high performance. Second, we leverage existing coordination models and patterns in order to create new spatial structures. Third, we introduce a novel language, based on the existing ``Field Calculus'' and integrated with the aforementioned toolchain, designed to be usable for practical aggregate programming.
Resumo:
Nowadays the rise of non-recurring engineering (NRE) costs associated with complexity is becoming a major factor in SoC design, limiting both scaling opportunities and the flexibility advantages offered by the integration of complex computational units. The introduction of embedded programmable elements can represent an appealing solution, able both to guarantee the desired flexibility and upgradabilty and to widen the SoC market. In particular embedded FPGA (eFPGA) cores can provide bit-level optimization for those applications which benefits from synthesis, paying on the other side in terms of performance penalties and area overhead with respect to standard cell ASIC implementations. In this scenario this thesis proposes a design methodology for a synthesizable programmable device designed to be embedded in a SoC. A soft-core embedded FPGA (eFPGA) is hence presented and analyzed in terms of the opportunities given by a fully synthesizable approach, following an implementation flow based on Standard-Cell methodology. A key point of the proposed eFPGA template is that it adopts a Multi-Stage Switching Network (MSSN) as the foundation of the programmable interconnects, since it can be efficiently synthesized and optimized through a standard cell based implementation flow, ensuring at the same time an intrinsic congestion-free network topology. The evaluation of the flexibility potentialities of the eFPGA has been performed using different technology libraries (STMicroelectronics CMOS 65nm and BCD9s 0.11μm) through a design space exploration in terms of area-speed-leakage tradeoffs, enabled by the full synthesizability of the template. Since the most relevant disadvantage of the adopted soft approach, compared to a hardcore, is represented by a performance overhead increase, the eFPGA analysis has been made targeting small area budgets. The generation of the configuration bitstream has been obtained thanks to the implementation of a custom CAD flow environment, and has allowed functional verification and performance evaluation through an application-aware analysis.
Resumo:
Patienten, die an Osteosarkom leiden werden derzeit mit intravenös applizierten krebstherapeutischen Mitteln nach Tumorresektion behandelt, was oftmals mit schweren Nebenwirkungen und einem verzögerten Knochenheilungsprozess einhergeht. Darüber hinaus treten vermehrt Rezidive aufgrund von verbleibenden neoplastischen Zellen an der Tumorresektionsstelle auf. Erfolgreiche Knochenregeneration und die Kontrolle von den im Gewebe verbleibenden Krebszellen stellt eine Herausforderung für das Tissue Engineering nach Knochenverlust durch Tumorentfernung dar. In dieser Hinsicht scheint der Einsatz von Hydroxyapatit als Knochenersatzmaterial in Kombination mit Cyclodextrin als Medikamententräger, vielversprechend. Chemotherapeutika können an Biomaterial gebunden und direkt am Tumorbett über einen längeren Zeitraum freigesetzt werden, um verbliebene neoplastische Zellen zu eliminieren. Lokal applizierte Chemotherapie hat diverse Vorteile, einschließlich der direkten zytotoxischen Auswirkung auf lokale Zellen, sowie die Reduzierung schwerer Nebenwirkungen. Diese Studie wurde durchgeführt, um die Funktionsfähigkeit eines solchen Arzneimittelabgabesystems zu bewerten und um Strategien im Bereich des Tissue Engineerings zu entwickeln, die den Knochenheilungsprozess und im speziellen die Vaskularisierung fördern sollen. Die Ergebnisse zeigen, dass nicht nur Krebszellen von der chemotherapeutischen Behandlung betroffen sind. Primäre Endothelzellen wie zum Beispiel HUVEC zeigten eine hohe Sensibilität Cisplatin und Doxorubicin gegenüber. Beide Medikamente lösten in HUVEC ein tumor-unterdrückendes Signal durch die Hochregulation von p53 und p21 aus. Zudem scheint Hypoxie einen krebstherapeutischen Einfluss zu haben, da die Behandlung sensitiver HUVEC mit Hypoxie die Zellen vor Zytotoxizität schützte. Der chemo-protektive Effekt schien deutlich weniger auf Krebszelllinien zu wirken. Diese Resultate könnten eine mögliche chemotherapeutische Strategie darstellen, um den Effekt eines zielgerichteten Medikamenteneinsatzes auf Krebszellen zu verbessern unter gleichzeitiger Schonung gesunder Zellen. Eine erfolgreiche Integration eines Systems, das Arzneimittel abgibt, kombiniert mit einem Biomaterial zur Stabilisierung und Regeneration, könnte gesunden Endothelzellen die Möglichkeit bieten zu proliferieren und Blutgefäße zu bilden, während verbleibende Krebszellen eliminiert werden. Da der Prozess der Knochengeweberemodellierung mit einer starken Beeinträchtigung der Lebensqualität des Patienten einhergeht, ist die Beschleunigung des postoperativen Heilungsprozesses eines der Ziele des Tissue Engineerings. Die Bildung von Blutgefäßen ist unabdingbar für eine erfolgreiche Integration eines Knochentransplantats in das Gewebe. Daher ist ein umfangreich ausgebildetes Blutgefäßsystem für einen verbesserten Heilungsprozess während der klinischen Anwendung wünschenswert. Frühere Experimente zeigen, dass sich die Anwendung von Ko-Kulturen aus humanen primären Osteoblasten (pOB) und humanen outgrowth endothelial cells (OEC) im Hinblick auf die Bildung stabiler gefäßähnlicher Strukturen in vitro, die auch effizient in das mikrovaskuläre System in vivo integriert werden konnten, als erfolgreich erweisen. Dieser Ansatz könnte genutzt werden, um prä-vaskularisierte Konstrukte herzustellen, die den Knochenheilungsprozess nach der Implantation fördern. Zusätzlich repräsentiert das Ko-Kultursystem ein exzellentes in vitro Model, um Faktoren, welche stark in den Prozess der Knochenheilung und Angiogenese eingebunden sind, zu identifizieren und zu analysieren. Es ist bekannt, dass Makrophagen eine maßgebliche Rolle in der inflammatorisch-induzierten Angiogenese spielen. In diesem Zusammenhang hebt diese Studie den positiven Einfluss THP-1 abgeleiteter Makrophagen in Ko-Kultur mit pOB und OEC hervor. Die Ergebnisse zeigten, dass die Anwendung von Makrophagen als inflammatorischer Stimulus im bereits etablierten Ko-Kultursystem zu einer pro-angiogenen Aktivierung der OEC führte, was in einer signifikant erhöhten Bildung blutgefäßähnlicher Strukturen in vitro resultierte. Außerdem zeigte die Analyse von Faktoren, die in der durch Entzündung hervorgerufenen Angiogenese eine wichtige Rolle spielen, eine deutliche Hochregulation von VEGF, inflammatorischer Zytokine und Adhäsionsmoleküle, die letztlich zu einer verstärkten Vaskularisierung beitragen. Diese Resultate werden dem Einfluss von Makrophagen zugeschrieben und könnten zukünftig im Tissue Engineering eingesetzt werden, um den Heilungsprozess zu beschleunigen und damit die klinische Situation von Patienten zu verbessern. Darüber hinaus könnte die Kombination der auf Ko-Kulturen basierenden Ansätze für das Knochen Tissue Engineering mit einem biomaterial-basierenden Arzneimittelabgabesystem zum klinischen Einsatz kommen, der die Eliminierung verbliebener Krebszellen mit der Förderung der Knochenregeneration verbindet.
Resumo:
The optical quality of the human eye mainly depends on the refractive performance of the cornea. The shape of the cornea is a mechanical balance between intraocular pressure and tissue intrinsic stiffness. Several surgical procedures in ophthalmology alter the biomechanics of the cornea to provoke local or global curvature changes for vision correction. Legitimated by the large number of surgical interventions performed every day, the demand for a deeper understanding of corneal biomechanics is rising to improve the safety of procedures and medical devices. The aim of our work is to propose a numerical model of corneal biomechanics, based on the stromal microstructure. Our novel anisotropic constitutive material law features a probabilistic weighting approach to model collagen fiber distribution as observed on human cornea by Xray scattering analysis (Aghamohammadzadeh et. al., Structure, February 2004). Furthermore, collagen cross-linking was explicitly included in the strain energy function. Results showed that the proposed model is able to successfully reproduce both inflation and extensiometry experimental data (Elsheikh et. al., Curr Eye Res, 2007; Elsheikh et. al., Exp Eye Res, May 2008). In addition, the mechanical properties calculated for patients of different age groups (Group A: 65-79 years; Group B: 80-95 years) demonstrate an increased collagen cross-linking, and a decrease in collagen fiber elasticity from younger to older specimen. These findings correspond to what is known about maturing fibrous biological tissue. Since the presented model can handle different loading situations and includes the anisotropic distribution of collagen fibers, it has the potential to simulate clinical procedures involving nonsymmetrical tissue interventions. In the future, such mechanical model can be used to improve surgical planning and the design of next generation ophthalmic devices.
Resumo:
The aim of this study was to assess the performance of two light-emitting diode (LED)- and two laser fluorescence-based devices in detecting occlusal caries in vitro. Ninety-seven permanent molars were assessed twice by two examiners using two LED- (Midwest Caries - MID and VistaProof - VP) and two laser fluorescence-based (DIAGNOdent 2095 - LF and DIAGNOdent pen 2190 - LFpen) devices. After measuring, the teeth were histologically prepared and classified according to lesion extension. At D1 the specificities were 0.76 (LF and LFpen), 0.94 (MID), and 0.70 (VP); the sensitivities were 0.70 (LF), 0.62 (LFpen), 0.31 (MID), and 0.75 (VP). At D(3) threshold the specificities were 0.88 (LF), 0.87 (LFpen), 0.90 (MID), and 0.70 (VP); the sensitivities were 0.63 (LF and LFpen), 0.70 (MID), and 0.96 (VP). Spearman's rank correlations with histology were 0.56 (LF), 0.51 (LFpen), 0.55 (MID), and 0.58 (VP). Inter- and intraexaminer ICC values were high and varied from 0.83 to 0.90. Both LF devices seemed to be useful auxiliary tools to the conventional methods, presenting good reproducibility and better accuracy at D(3) threshold. MID was not able to differentiate sound surfaces from enamel caries and VP still needs improvement on the cut-off limits for its use.
Resumo:
This is the first part of a study investigating a model-based transient calibration process for diesel engines. The motivation is to populate hundreds of parameters (which can be calibrated) in a methodical and optimum manner by using model-based optimization in conjunction with the manual process so that, relative to the manual process used by itself, a significant improvement in transient emissions and fuel consumption and a sizable reduction in calibration time and test cell requirements is achieved. Empirical transient modelling and optimization has been addressed in the second part of this work, while the required data for model training and generalization are the focus of the current work. Transient and steady-state data from a turbocharged multicylinder diesel engine have been examined from a model training perspective. A single-cylinder engine with external air-handling has been used to expand the steady-state data to encompass transient parameter space. Based on comparative model performance and differences in the non-parametric space, primarily driven by a high engine difference between exhaust and intake manifold pressures (ΔP) during transients, it has been recommended that transient emission models should be trained with transient training data. It has been shown that electronic control module (ECM) estimates of transient charge flow and the exhaust gas recirculation (EGR) fraction cannot be accurate at the high engine ΔP frequently encountered during transient operation, and that such estimates do not account for cylinder-to-cylinder variation. The effects of high engine ΔP must therefore be incorporated empirically by using transient data generated from a spectrum of transient calibrations. Specific recommendations on how to choose such calibrations, how many data to acquire, and how to specify transient segments for data acquisition have been made. Methods to process transient data to account for transport delays and sensor lags have been developed. The processed data have then been visualized using statistical means to understand transient emission formation. Two modes of transient opacity formation have been observed and described. The first mode is driven by high engine ΔP and low fresh air flowrates, while the second mode is driven by high engine ΔP and high EGR flowrates. The EGR fraction is inaccurately estimated at both modes, while EGR distribution has been shown to be present but unaccounted for by the ECM. The two modes and associated phenomena are essential to understanding why transient emission models are calibration dependent and furthermore how to choose training data that will result in good model generalization.
Resumo:
Outside of relatively limited crash testing with large trucks, very little is known regarding the performance of traffic barriers subjected to real-world large truck impacts. The purpose of this study was to investigate real-world large truck impacts into traffic barriers to determine barrier crash involvement rates, the impact performance of barriers not specifically designed to redirect large trucks, and the real-world performance of large-truck-specific barriers. Data sources included the Fatality Analysis Reporting System (2000-2009), the General Estimates System (2000-2009) and 155 in-depth large truck-to-barrier crashes from the Large Truck Crash Causation Study. Large truck impacts with a longitudinal barrier were found to comprise 3 percent of all police-reported longitudinal barrier impacts and roughly the same proportion of barrier fatalities. Based on a logistic regression model predicting barrier penetration, large truck barrier penetration risk was found to increase by a factor of 6 for impacts with barriers designed primarily for passenger vehicles. Although large-truck-specific barriers were found to perform better than non-heavy vehicle specific barriers, the penetration rate of these barriers were found to be 17 percent. This penetration rate is especially a concern because the higher test level barriers are designed to protect other road users, not the occupants of the large truck. Surprisingly, barriers not specifically designed for large truck impacts were found to prevent large truck penetration approximately half of the time. This suggests that adding costlier higher test level barriers may not always be warranted, especially on roadways with lower truck volumes.
Resumo:
We present a new approach for corpus-based speech enhancement that significantly improves over a method published by Xiao and Nickel in 2010. Corpus-based enhancement systems do not merely filter an incoming noisy signal, but resynthesize its speech content via an inventory of pre-recorded clean signals. The goal of the procedure is to perceptually improve the sound of speech signals in background noise. The proposed new method modifies Xiao's method in four significant ways. Firstly, it employs a Gaussian mixture model (GMM) instead of a vector quantizer in the phoneme recognition front-end. Secondly, the state decoding of the recognition stage is supported with an uncertainty modeling technique. With the GMM and the uncertainty modeling it is possible to eliminate the need for noise dependent system training. Thirdly, the post-processing of the original method via sinusoidal modeling is replaced with a powerful cepstral smoothing operation. And lastly, due to the improvements of these modifications, it is possible to extend the operational bandwidth of the procedure from 4 kHz to 8 kHz. The performance of the proposed method was evaluated across different noise types and different signal-to-noise ratios. The new method was able to significantly outperform traditional methods, including the one by Xiao and Nickel, in terms of PESQ scores and other objective quality measures. Results of subjective CMOS tests over a smaller set of test samples support our claims.