856 resultados para 0914 Resources Engineering and Extractive Metallurgy


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The rapid growth in high data rate communication systems has introduced new high spectral efficient modulation techniques and standards such as LTE-A (long term evolution-advanced) for 4G (4th generation) systems. These techniques have provided a broader bandwidth but introduced high peak-to-average power ratio (PAR) problem at the high power amplifier (HPA) level of the communication system base transceiver station (BTS). To avoid spectral spreading due to high PAR, stringent requirement on linearity is needed which brings the HPA to operate at large back-off power at the expense of power efficiency. Consequently, high power devices are fundamental in HPAs for high linearity and efficiency. Recent development in wide bandgap power devices, in particular AlGaN/GaN HEMT, has offered higher power level with superior linearity-efficiency trade-off in microwaves communication. For cost-effective HPA design to production cycle, rigorous computer aided design (CAD) AlGaN/GaN HEMT models are essential to reflect real response with increasing power level and channel temperature. Therefore, large-size AlGaN/GaN HEMT large-signal electrothermal modeling procedure is proposed. The HEMT structure analysis, characterization, data processing, model extraction and model implementation phases have been covered in this thesis including trapping and self-heating dispersion accounting for nonlinear drain current collapse. The small-signal model is extracted using the 22-element modeling procedure developed in our department. The intrinsic large-signal model is deeply investigated in conjunction with linearity prediction. The accuracy of the nonlinear drain current has been enhanced through several issues such as trapping and self-heating characterization. Also, the HEMT structure thermal profile has been investigated and corresponding thermal resistance has been extracted through thermal simulation and chuck-controlled temperature pulsed I(V) and static DC measurements. Higher-order equivalent thermal model is extracted and implemented in the HEMT large-signal model to accurately estimate instantaneous channel temperature. Moreover, trapping and self-heating transients has been characterized through transient measurements. The obtained time constants are represented by equivalent sub-circuits and integrated in the nonlinear drain current implementation to account for complex communication signals dynamic prediction. The obtained verification of this table-based large-size large-signal electrothermal model implementation has illustrated high accuracy in terms of output power, gain, efficiency and nonlinearity prediction with respect to standard large-signal test signals.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The global power supply stability is faced to several severe and fundamental threats, in particular steadily increasing power demand, diminishing and degrading fossil and nuclear energy resources, very harmful greenhouse gas emissions, significant energy injustice and a structurally misbalanced ecological footprint. Photovoltaic (PV) power systems are analysed in various aspects focusing on economic and technical considerations of supplemental and substitutional power supply to the constraint conventional power system. To infer the most relevant system approach for PV power plants several solar resources available for PV systems are compared. By combining the different solar resources and respective economics, two major PV systems are identified to be very competitive in almost all regions in the world. The experience curve concept is used as a key technique for the development of scenario assumptions on economic projections for the decade of the 2010s. Main drivers for cost reductions in PV systems are learning and production growth rate, thus several relevant aspects are discussed such as research and development investments, technical PV market potential, different PV technologies and the energetic sustainability of PV. Three major market segments for PV systems are identified: off-grid PV solutions, decentralised small scale on-grid PV systems (several kWp) and large scale PV power plants (tens of MWp). Mainly by application of ‘grid-parity’ and ‘fuel-parity’ concepts per country, local market and conventional power plant basis, the global economic market potential for all major PV system segments is derived. PV power plant hybridization potential of all relevant power technologies and the global power plant structure are analyzed regarding technical, economical and geographical feasibility. Key success criteria for hybrid PV power plants are discussed and comprehensively analysed for all adequate power plant technologies, i.e. oil, gas and coal fired power plants, wind power, solar thermal power (STEG) and hydro power plants. For the 2010s, detailed global demand curves are derived for hybrid PV-Fossil power plants on a per power plant, per country and per fuel type basis. The fundamental technical and economic potentials for hybrid PV-STEG, hybrid PV-Wind and hybrid PV-Hydro power plants are considered. The global resource availability for PV and wind power plants is excellent, thus knowing the competitive or complementary characteristic of hybrid PV-Wind power plants on a local basis is identified as being of utmost relevance. The complementarity of hybrid PV-Wind power plants is confirmed. As a result of that almost no reduction of the global economic PV market potential need to be expected and more complex power system designs on basis of hybrid PV-Wind power plants are feasible. The final target of implementing renewable power technologies into the global power system is a nearly 100% renewable power supply. Besides balancing facilities, storage options are needed, in particular for seasonal power storage. Renewable power methane (RPM) offers respective options. A comprehensive global and local analysis is performed for analysing a hybrid PV-Wind-RPM combined cycle gas turbine power system. Such a power system design might be competitive and could offer solutions for nearly all current energy system constraints including the heating and transportation sector and even the chemical industry. Summing up, hybrid PV power plants become very attractive and PV power systems will very likely evolve together with wind power to the major and final source of energy for mankind.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Die thermische Verarbeitung von Lebensmitteln beeinflusst deren Qualität und ernährungsphysiologischen Eigenschaften. Im Haushalt ist die Überwachung der Temperatur innerhalb des Lebensmittels sehr schwierig. Zudem ist das Wissen über optimale Temperatur- und Zeitparameter für die verschiedenen Speisen oft unzureichend. Die optimale Steuerung der thermischen Zubereitung ist maßgeblich abhängig von der Art des Lebensmittels und der äußeren und inneren Temperatureinwirkung während des Garvorgangs. Das Ziel der Arbeiten war die Entwicklung eines automatischen Backofens, der in der Lage ist, die Art des Lebensmittels zu erkennen und die Temperatur im Inneren des Lebensmittels während des Backens zu errechnen. Die für die Temperaturberechnung benötigten Daten wurden mit mehreren Sensoren erfasst. Hierzu kam ein Infrarotthermometer, ein Infrarotabstandssensor, eine Kamera, ein Temperatursensor und ein Lambdasonde innerhalb des Ofens zum Einsatz. Ferner wurden eine Wägezelle, ein Strom- sowie Spannungs-Sensor und ein Temperatursensor außerhalb des Ofens genutzt. Die während der Aufheizphase aufgenommen Datensätze ermöglichten das Training mehrerer künstlicher neuronaler Netze, die die verschiedenen Lebensmittel in die entsprechenden Kategorien einordnen konnten, um so das optimale Backprogram auszuwählen. Zur Abschätzung der thermische Diffusivität der Nahrung, die von der Zusammensetzung (Kohlenhydrate, Fett, Protein, Wasser) abhängt, wurden mehrere künstliche neuronale Netze trainiert. Mit Ausnahme des Fettanteils der Lebensmittel konnten alle Komponenten durch verschiedene KNNs mit einem Maximum von 8 versteckten Neuronen ausreichend genau abgeschätzt werden um auf deren Grundlage die Temperatur im inneren des Lebensmittels zu berechnen. Die durchgeführte Arbeit zeigt, dass mit Hilfe verschiedenster Sensoren zur direkten beziehungsweise indirekten Messung der äußeren Eigenschaften der Lebensmittel sowie KNNs für die Kategorisierung und Abschätzung der Lebensmittelzusammensetzung die automatische Erkennung und Berechnung der inneren Temperatur von verschiedensten Lebensmitteln möglich ist.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The ongoing depletion of the coastal aquifer in the Gaza strip due to groundwater overexploitation has led to the process of seawater intrusion, which is continually becoming a serious problem in Gaza, as the seawater has further invaded into many sections along the coastal shoreline. As a first step to get a hold on the problem, the artificial neural network (ANN)-model has been applied as a new approach and an attractive tool to study and predict groundwater levels without applying physically based hydrologic parameters, and also for the purpose to improve the understanding of complex groundwater systems and which is able to show the effects of hydrologic, meteorological and anthropogenic impacts on the groundwater conditions. Prediction of the future behaviour of the seawater intrusion process in the Gaza aquifer is thus of crucial importance to safeguard the already scarce groundwater resources in the region. In this study the coupled three-dimensional groundwater flow and density-dependent solute transport model SEAWAT, as implemented in Visual MODFLOW, is applied to the Gaza coastal aquifer system to simulate the location and the dynamics of the saltwater–freshwater interface in the aquifer in the time period 2000-2010. A very good agreement between simulated and observed TDS salinities with a correlation coefficient of 0.902 and 0.883 for both steady-state and transient calibration is obtained. After successful calibration of the solute transport model, simulation of future management scenarios for the Gaza aquifer have been carried out, in order to get a more comprehensive view of the effects of the artificial recharge planned in the Gaza strip for some time on forestall, or even to remedy, the presently existing adverse aquifer conditions, namely, low groundwater heads and high salinity by the end of the target simulation period, year 2040. To that avail, numerous management scenarios schemes are examined to maintain the ground water system and to control the salinity distributions within the target period 2011-2040. In the first, pessimistic scenario, it is assumed that pumping from the aquifer continues to increase in the near future to meet the rising water demand, and that there is not further recharge to the aquifer than what is provided by natural precipitation. The second, optimistic scenario assumes that treated surficial wastewater can be used as a source of additional artificial recharge to the aquifer which, in principle, should not only lead to an increased sustainable yield of the latter, but could, in the best of all cases, revert even some of the adverse present-day conditions in the aquifer, i.e., seawater intrusion. This scenario has been done with three different cases which differ by the locations and the extensions of the injection-fields for the treated wastewater. The results obtained with the first (do-nothing) scenario indicate that there will be ongoing negative impacts on the aquifer, such as a higher propensity for strong seawater intrusion into the Gaza aquifer. This scenario illustrates that, compared with 2010 situation of the baseline model, at the end of simulation period, year 2040, the amount of saltwater intrusion into the coastal aquifer will be increased by about 35 %, whereas the salinity will be increased by 34 %. In contrast, all three cases of the second (artificial recharge) scenario group can partly revert the present seawater intrusion. From the water budget point of view, compared with the first (do nothing) scenario, for year 2040, the water added to the aquifer by artificial recharge will reduces the amount of water entering the aquifer by seawater intrusion by 81, 77and 72 %, for the three recharge cases, respectively. Meanwhile, the salinity in the Gaza aquifer will be decreased by 15, 32 and 26% for the three cases, respectively.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This study was aim to describe the indigenous knowledge of farmers at Nagari Padang laweh Malalo (NPLM) and their adaptability to climate change. Not only the water scarcity is feared, but climate change is also affecting their food security. Local food security can be achieved if biodiversity in their surrounding area is suitable to the local needs. The study was conducted by using Participatory Rural Appraisal (PRA) such as observation and discussion. The combination of in depth interview, life history, semi structure questionnaire, pictures, mapping and expert interviews was implemented. Data was analyzed by using MAXQDA 10 and F4 audio analysis software. The result shows awareness of the people and scarcity of water conditions has allowed the people of NPLM to face this challenge with wisdom. Aia adat (water resources controlled and regulate by custom) is one of their strategies to distribute the water. The general rule is that irrigation will flow from 6 pm – 6 am regularly to all farm land under supervision of kapalo banda. When rains occur, water resources can be used during the day without special supervision. They were used traditional knowledge to manage water resources for their land and daily usage. This study may be helpful for researcher and other farmers in different region to learn encounter water scarcity.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Die zunehmende Vernetzung der Informations- und Kommunikationssysteme führt zu einer weiteren Erhöhung der Komplexität und damit auch zu einer weiteren Zunahme von Sicherheitslücken. Klassische Schutzmechanismen wie Firewall-Systeme und Anti-Malware-Lösungen bieten schon lange keinen Schutz mehr vor Eindringversuchen in IT-Infrastrukturen. Als ein sehr wirkungsvolles Instrument zum Schutz gegenüber Cyber-Attacken haben sich hierbei die Intrusion Detection Systeme (IDS) etabliert. Solche Systeme sammeln und analysieren Informationen von Netzwerkkomponenten und Rechnern, um ungewöhnliches Verhalten und Sicherheitsverletzungen automatisiert festzustellen. Während signatur-basierte Ansätze nur bereits bekannte Angriffsmuster detektieren können, sind anomalie-basierte IDS auch in der Lage, neue bisher unbekannte Angriffe (Zero-Day-Attacks) frühzeitig zu erkennen. Das Kernproblem von Intrusion Detection Systeme besteht jedoch in der optimalen Verarbeitung der gewaltigen Netzdaten und der Entwicklung eines in Echtzeit arbeitenden adaptiven Erkennungsmodells. Um diese Herausforderungen lösen zu können, stellt diese Dissertation ein Framework bereit, das aus zwei Hauptteilen besteht. Der erste Teil, OptiFilter genannt, verwendet ein dynamisches "Queuing Concept", um die zahlreich anfallenden Netzdaten weiter zu verarbeiten, baut fortlaufend Netzverbindungen auf, und exportiert strukturierte Input-Daten für das IDS. Den zweiten Teil stellt ein adaptiver Klassifikator dar, der ein Klassifikator-Modell basierend auf "Enhanced Growing Hierarchical Self Organizing Map" (EGHSOM), ein Modell für Netzwerk Normalzustand (NNB) und ein "Update Model" umfasst. In dem OptiFilter werden Tcpdump und SNMP traps benutzt, um die Netzwerkpakete und Hostereignisse fortlaufend zu aggregieren. Diese aggregierten Netzwerkpackete und Hostereignisse werden weiter analysiert und in Verbindungsvektoren umgewandelt. Zur Verbesserung der Erkennungsrate des adaptiven Klassifikators wird das künstliche neuronale Netz GHSOM intensiv untersucht und wesentlich weiterentwickelt. In dieser Dissertation werden unterschiedliche Ansätze vorgeschlagen und diskutiert. So wird eine classification-confidence margin threshold definiert, um die unbekannten bösartigen Verbindungen aufzudecken, die Stabilität der Wachstumstopologie durch neuartige Ansätze für die Initialisierung der Gewichtvektoren und durch die Stärkung der Winner Neuronen erhöht, und ein selbst-adaptives Verfahren eingeführt, um das Modell ständig aktualisieren zu können. Darüber hinaus besteht die Hauptaufgabe des NNB-Modells in der weiteren Untersuchung der erkannten unbekannten Verbindungen von der EGHSOM und der Überprüfung, ob sie normal sind. Jedoch, ändern sich die Netzverkehrsdaten wegen des Concept drif Phänomens ständig, was in Echtzeit zur Erzeugung nicht stationärer Netzdaten führt. Dieses Phänomen wird von dem Update-Modell besser kontrolliert. Das EGHSOM-Modell kann die neuen Anomalien effektiv erkennen und das NNB-Model passt die Änderungen in Netzdaten optimal an. Bei den experimentellen Untersuchungen hat das Framework erfolgversprechende Ergebnisse gezeigt. Im ersten Experiment wurde das Framework in Offline-Betriebsmodus evaluiert. Der OptiFilter wurde mit offline-, synthetischen- und realistischen Daten ausgewertet. Der adaptive Klassifikator wurde mit dem 10-Fold Cross Validation Verfahren evaluiert, um dessen Genauigkeit abzuschätzen. Im zweiten Experiment wurde das Framework auf einer 1 bis 10 GB Netzwerkstrecke installiert und im Online-Betriebsmodus in Echtzeit ausgewertet. Der OptiFilter hat erfolgreich die gewaltige Menge von Netzdaten in die strukturierten Verbindungsvektoren umgewandelt und der adaptive Klassifikator hat sie präzise klassifiziert. Die Vergleichsstudie zwischen dem entwickelten Framework und anderen bekannten IDS-Ansätzen zeigt, dass der vorgeschlagene IDSFramework alle anderen Ansätze übertrifft. Dies lässt sich auf folgende Kernpunkte zurückführen: Bearbeitung der gesammelten Netzdaten, Erreichung der besten Performanz (wie die Gesamtgenauigkeit), Detektieren unbekannter Verbindungen und Entwicklung des in Echtzeit arbeitenden Erkennungsmodells von Eindringversuchen.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The objective of this study was to develop an internet-based seminar framework applicable for landscape architecture education. This process was accompanied by various aims. The basic expectation was to keep the main characteristics of landscape architecture education also in the online format. On top of that, four further objectives were anticipated: (1) training of competences for virtual team work, (2) fostering intercultural competence, (3) creation of equal opportunities for education through internet-based open access and (4) synergy effects and learning processes across institutional boundaries. This work started with the hypothesis that these four expected advantages would compensate for additional organisational efforts caused by the online delivery of the seminars and thus lead to a sustainable integration of this new learning mode into landscape architecture curricula. This rationale was followed by a presentation of four areas of knowledge to which the seminar development was directly related (1) landscape architecture as a subject and its pedagogy, (2) general learning theories, (3) developments in the ICT sector and (4) wider societal driving forces such as global citizenship and the increase of open educational resources. The research design took the shape of a pedagogical action research cycle. This approach was constructive: The author herself is teaching international landscape architecture students so that the model could directly be applied in practice. Seven online seminars were implemented in the period from 2008 to 2013 and this experience represents the core of this study. The seminars were conducted with varying themes while its pedagogy, organisation and the technological tools remained widely identical. The research design is further based on three levels of observation: (1) the seminar design on the basis of theory and methods from the learning sciences, in particular educational constructivism, (2) the seminar evaluation and (3) the evaluation of the seminars’ long term impact. The seminar model itself basically consists of four elements: (1) the taxonomy of learning objectives, (2) ICT tools and their application and pedagogy, (3) process models and (4) the case study framework. The seminar framework was followed by the presentation of the evaluation findings. The major findings of this study can be summed up as follows: Implementing online seminars across educational and national boundaries was possible both in term of organisation and technology. In particular, a high level of cultural diversity among the seminar participants has definitively been achieved. However, there were also obvious obstacles. These were primarily competing study commitments and incompatible schedules among the students attending from different academic programmes, partly even in different time zones. Both factors had negative impact on the individual and working group performances. With respect to the technical framework it can be concluded that the majority of the participants were able to use the tools either directly without any problem or after overcoming some smaller problems. Also the seminar wiki was intensively used for completing the seminar assignments. However, too less truly collaborative text production was observed which could be improved by changing the requirements for the collaborative task. Two different process models have been applied for guiding the collaboration of the small groups and both were in general successful. However, it needs to be said that even if the students were able to follow the collaborative task and to co-construct and compare case studies, most of them were not able to synthesize the knowledge they had compiled. This means that the area of consideration often remained on the level of the case and further reflections, generalisations and critique were largely missing. This shows that the seminar model needs to find better ways for triggering knowledge building and critical reflection. It was also suggested to have a more differentiated group building strategy in future seminars. A comparison of pre- and post seminar concept maps showed that an increase of factual and conceptual knowledge on the individual level was widely recognizable. Also the evaluation of the case studies (the major seminar output) revealed that the students have undergone developments of both the factual and the conceptual knowledge domain. Also their self-assessment with respect to individual learning development showed that the highest consensus was achieved in the field of subject-specific knowledge. The participants were much more doubtful with regard to the progress of generic competences such as analysis, communication and organisation. However, 50% of the participants confirmed that they perceived individual development on all competence areas the survey had asked for. Have the additional four targets been met? Concerning the competences for working in a virtual team it can be concluded that the vast majority was able to use the internet-based tools and to work with them in a target-oriented way. However, there were obvious differences regarding the intensity and activity of participation, both because of external and personal factors. A very positive aspect is the achievement of a high cultural diversity supporting the participants’ intercultural competence. Learning from group members was obviously a success factor for the working groups. Regarding the possibilities for better accessibility of educational opportunities it became clear that a significant number of participants were not able to go abroad during their studies because of financial or personal reasons. They confirmed that the online seminar was to some extent a compensation for not having been abroad for studying. Inter-institutional learning and synergy was achieved in so far that many teachers from different countries contributed with individual lectures. However, those teachers hardly ever followed more than one session. Therefore, the learning effect remained largely within the seminar learning group. Looking back at the research design it can be said that the pedagogical action research cycle was an appropriate and valuable approach allowing for strong interaction between theory and practice. However, some more external evaluation from peers in particular regarding the participants’ products would have been valuable.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We investigate the effect of the epitaxial structure and the acceptor doping profile on the efficiency droop in InGaN/GaN LEDs by the physics based simulation of experimental internal quantum efficiency (IQE) characteristics. The device geometry is an integral part of our simulation approach. We demonstrate that even for single quantum well LEDs the droop depends critically on the acceptor doping profile. The Auger recombination was found to increase stronger than with the third power of the carrier density and has been found to dominate the droop in the roll over zone of the IQE. The fitted Auger coefficients are in the range of the values predicted by atomistic simulations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The accurate transport of an ion over macroscopic distances represents a challenging control problem due to the different length and time scales that enter and the experimental limitations on the controls that need to be accounted for. Here, we investigate the performance of different control techniques for ion transport in state-of-the-art segmented miniaturized ion traps. We employ numerical optimization of classical trajectories and quantum wavepacket propagation as well as analytical solutions derived from invariant based inverse engineering and geometric optimal control. The applicability of each of the control methods depends on the length and time scales of the transport. Our comprehensive set of tools allows us make a number of observations. We find that accurate shuttling can be performed with operation times below the trap oscillation period. The maximum speed is limited by the maximum acceleration that can be exerted on the ion. When using controls obtained from classical dynamics for wavepacket propagation, wavepacket squeezing is the only quantum effect that comes into play for a large range of trapping parameters. We show that this can be corrected by a compensating force derived from invariant based inverse engineering, without a significant increase in the operation time.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Tunable Optical Sensor Arrays (TOSA) based on Fabry-Pérot (FP) filters, for high quality spectroscopic applications in the visible and near infrared spectral range are investigated within this work. The optical performance of the FP filters is improved by using ion beam sputtered niobium pentoxide (Nb2O5) and silicon dioxide (SiO2) Distributed Bragg Reflectors (DBRs) as mirrors. Due to their high refractive index contrast, only a few alternating pairs of Nb2O5 and SiO2 films can achieve DBRs with high reflectivity in a wide spectral range, while ion beam sputter deposition (IBSD) is utilized due to its ability to produce films with high optical purity. However, IBSD films are highly stressed; resulting in stress induced mirror curvature and suspension bending in the free standing filter suspensions of the MEMS (Micro-Electro-Mechanical Systems) FP filters. Stress induced mirror curvature results in filter transmission line degradation, while suspension bending results in high required filter tuning voltages. Moreover, stress induced suspension bending results in higher order mode filter operation which in turn degrades the optical resolution of the filter. Therefore, the deposition process is optimized to achieve both near zero absorption and low residual stress. High energy ion bombardment during film deposition is utilized to reduce the film density, and hence the film compressive stress. Utilizing this technique, the compressive stress of Nb2O5 is reduced by ~43%, while that for SiO2 is reduced by ~40%. Filters fabricated with stress reduced films show curvatures as low as 100 nm for 70 μm mirrors. To reduce the stress induced bending in the free standing filter suspensions, a stress optimized multi-layer suspension design is presented; with a tensile stressed metal sandwiched between two compressively stressed films. The stress in Physical Vapor Deposited (PVD) metals is therefore characterized for use as filter top-electrode and stress compensating layer. Surface micromachining is used to fabricate tunable FP filters in the visible spectral range using the above mentioned design. The upward bending of the suspensions is reduced from several micrometers to less than 100 nm and 250 nm for two different suspension layer combinations. Mechanical tuning of up to 188 nm is obtained by applying 40 V of actuation voltage. Alternatively, a filter line with transmission of 65.5%, Full Width at Half Maximum (FWHM) of 10.5 nm and a stopband of 170 nm (at an output wavelength of 594 nm) is achieved. Numerical model simulations are also performed to study the validity of the stress optimized suspension design for the near infrared spectral range, wherein membrane displacement and suspension deformation due to material residual stress is studied. Two bandpass filter designs based on quarter-wave and non-quarter-wave layers are presented as integral components of the TOSA. With a filter passband of 135 nm and a broad stopband of over 650 nm, high average filter transmission of 88% is achieved inside the passband, while maximum filter transmission of less than 1.6% outside the passband is achieved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this thesis, a dual mode tunable gas sensor based on intracavity laser absorption spectroscopy (ICLAS) principle is investigated, both, numerically and experimentally. In order to minimize the cost and size of the gas sensor, relative intensity noise (RIN) is implemented as a detection parameter. Investigation is performed to determine the effect of injection current, operating temperature, mode spacing, and cavity length on RIN. It has been found that it is best to operate the gas sensor at smaller mode spacing and near the threshold current or at larger mode spacing and far above the threshold current for the use of RIN as the readout parameter.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This report is intended to shed more light on the ongoing water struggle in Caimanes, a small urban area in the central northern area of Chile, neighbouring Latin America’s biggest tailings dam. Undoubtedly, the water in Caimanes is running out and the conflict between the opponents of the dam and its owner, a multinational copper enterprise, is getting more and more attention by the national and also international media. In the discussion a judgment of the Chilean Supreme Court from last October plays a central role, because it is said to have granted the people from Caimanes their right to water. After a short introduction with some details about Camaines and the tailings from the dam El Mauro, the key points of this judgment shall be outlined. The final part of the report is dedicated to various institutional problems of the Chilean resources law and policy that can become virulent for the water supply and the environmental well-being of many other urban areas in the industrialized north of Chile.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Ein Drittel des weltweiten gesamten Energiebedarfs wird durch Gebäude verbraucht. Um diesen Energiebedarf teilweise zu decken, den erheblichen Energieverbrauch zu reduzieren und weiterhin andere Gebäudefunktionen beizubehalten, ist Gebäudeintegrierte Photovoltaik (BIPV) eine der am besten geeigneten Lösungen für die Gebäudenanwendung. Im Bezug auf eine Vielzahl von Gestalltungsmöglichkeiten, sind die Randbedingungen der BIPV-Anwendungen eindeutig anders im Vergleich zu Standard-PV-Anwendungen, insbesondere bezüglich der Betriebstemperatur. Bisher gab es nicht viele Informationen zu den relevanten thermischen Auswirkungen auf die entsprechenden elektrischen Eigenschaften zusammen mit thermischen und mechanischen relevanten Gebäudenfunktionen. Die meisten Hersteller übernehmen diese Eigenschaften von entsprechenden PV-Modulen und konventionellen Bauprodukten Normen, die zur ungenauen System- und Gebäudeplanungen führen. Deshalb ist die Untersuchung des thermischen Einflusses auf elektrische, thermische sowie mechanische Eigenschaften das Hauptziel der vorliegenden Arbeit. Zunächst wird das Temperatur-Model mit dem Power-Balance-Konzept erstellt. Unter Berücksichtigung der variablen Installationsmöglichkeiten und Konfigurationen des Moduls wird das Model auf Basis dynamischer und stationär Eigenschaften entwickelt. Im Hinblick auf die dynamische Simulation können der Energieertrag und Leistung zusammen mit der thermischen Gebäudesimulation in Echtzeit simuliert werden. Für stationäre Simulationen können die relevanten Gebäudefunktionen von BIPV-Modulen sowohl im Sommer als auch im Winter simuliert werden. Basierend auf unterschiedlichen thermischen und mechanischen Last-Szenarien wurde darüber hinaus das mechanische Model zusammen mit Variationen von Belastungsdauer, Montagesystem und Verkapselungsmaterialien entwickelt. Um die Temperatur- und Mechanik-Modelle zu validieren, wurden die verschiedenen Prüfeinrichtungen zusammen mit neuen Testmethoden entwickelt. Bei Verwendung der Prüfanlage „PV variable mounting system“ und „mechanical testing equipment“ werden zudem die verschiedenen Szenarien von Montagesystemen, Modul-Konfigurationen und mechanischen Belastungen emuliert. Mit der neuen Testmethode „back-bias current concept“ können zum einen die solare Einstrahlung und bestimmte Betriebstemperaturen eingestellt werden. Darüber hinaus wurden mit den eingangs erwähnten validierten Modellen das jeweilige elektrische, thermische und mechanische Verhalten auf andere Konfigurationen bewertet. Zum Abschluss wird die Anwendung von Software-Tools bei PV-Herstellern im Hinblick auf die entsprechenden Modellentwicklungen thematisiert.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The next generations of both biological engineering and computer engineering demand that control be exerted at the molecular level. Creating, characterizing and controlling synthetic biological systems may provide us with the ability to build cells that are capable of a plethora of activities, from computation to synthesizing nanostructures. To develop these systems, we must have a set of tools not only for synthesizing systems, but also designing and simulating them. The BioJADE project provides a comprehensive, extensible design and simulation platform for synthetic biology. BioJADE is a graphical design tool built in Java, utilizing a database back end, and supports a range of simulations using an XML communication protocol. BioJADE currently supports a library of over 100 parts with which it can compile designs into actual DNA, and then generate synthesis instructions to build the physical parts. The BioJADE project contributes several tools to Synthetic Biology. BioJADE in itself is a powerful tool for synthetic biology designers. Additionally, we developed and now make use of a centralized BioBricks repository, which enables the sharing of BioBrick components between researchers, and vastly reduces the barriers to entry for aspiring Synthetic Biologists.