882 resultados para Operating speed
Resumo:
Most of the commercial and financial data are stored in decimal fonn. Recently, support for decimal arithmetic has received increased attention due to the growing importance in financial analysis, banking, tax calculation, currency conversion, insurance, telephone billing and accounting. Performing decimal arithmetic with systems that do not support decimal computations may give a result with representation error, conversion error, and/or rounding error. In this world of precision, such errors are no more tolerable. The errors can be eliminated and better accuracy can be achieved if decimal computations are done using Decimal Floating Point (DFP) units. But the floating-point arithmetic units in today's general-purpose microprocessors are based on the binary number system, and the decimal computations are done using binary arithmetic. Only few common decimal numbers can be exactly represented in Binary Floating Point (BF P). ln many; cases, the law requires that results generated from financial calculations performed on a computer should exactly match with manual calculations. Currently many applications involving fractional decimal data perform decimal computations either in software or with a combination of software and hardware. The performance can be dramatically improved by complete hardware DFP units and this leads to the design of processors that include DF P hardware.VLSI implementations using same modular building blocks can decrease system design and manufacturing cost. A multiplexer realization is a natural choice from the viewpoint of cost and speed.This thesis focuses on the design and synthesis of efficient decimal MAC (Multiply ACeumulate) architecture for high speed decimal processors based on IEEE Standard for Floating-point Arithmetic (IEEE 754-2008). The research goal is to design and synthesize deeimal'MAC architectures to achieve higher performance.Efficient design methods and architectures are developed for a high performance DFP MAC unit as part of this research.
Resumo:
Trawling, despite being heavily energy expensive, still continues to be the most energy expensive fishing method particularly so in View of the export oriented nature of the Indian seafood industry. This study therefore aims at analyzing the efficiency of trawls operation from Cochin, an important fishing center along the southwest coast of India. The analysis is made along two perspectives - economic and technological. Even though technological efficiency complement economic efficiency, in the fishing parlance, parameters like the size composition of the catch, selectivity factors, etc., will have a direct bearing on the technological qualities of the trawl, and which parameters will have a significant impact on the effective exploitation of a fishery stock. Whereas the technological analysis aims at improving the efficiency with regard to the effective utilization of fuel and fishery stocks, economic analysis ascertains the present status of the trawling operations from the commercial angle.
Resumo:
In a sigma-delta analog to digital (A/D) As most of the sigma-delta ADC applications require converter, the most computationally intensive block is decimation filters with linear phase characteristics, the decimation filter and its hardware implementation symmetric Finite Impulse Response (FIR) filters are may require millions of transistors. Since these widely used for implementation. But the number of FIR converters are now targeted for a portable application, filter coefficients will be quite large for implementing a a hardware efficient design is an implicit requirement. narrow band decimation filter. Implementing decimation In this effect, this paper presents a computationally filter in several stages reduces the total number of filter efficient polyphase implementation of non-recursive coefficients, and hence reduces the hardware complexity cascaded integrator comb (CIC) decimators for and power consumption [2]. Sigma-Delta Converters (SDCs). The SDCs are The first stage of decimation filter can be operating at high oversampling frequencies and hence implemented very efficiently using a cascade of integrators require large sampling rate conversions. The filtering and comb filters which do not require multiplication or and rate reduction are performed in several stages to coefficient storage. The remaining filtering is performed reduce hardware complexity and power dissipation. either in single stage or in two stages with more complex The CIC filters are widely adopted as the first stage of FIR or infinite impulse response (IIR) filters according to decimation due to its multiplier free structure. In this the requirements. The amount of passband aliasing or research, the performance of polyphase structure is imaging error can be brought within prescribed bounds by compared with the CICs using recursive and increasing the number of stages in the CIC filter. The non-recursive algorithms in terms of power, speed and width of the passband and the frequency characteristics area. This polyphase implementation offers high speed outside the passband are severely limited. So, CIC filters operation and low power consumption. The polyphase are used to make the transition between high and low implementation of 4th order CIC filter with a sampling rates. Conventional filters operating at low decimation factor of '64' and input word length of sampling rate are used to attain the required transition '4-bits' offers about 70% and 37% of power saving bandwidth and stopband attenuation. compared to the corresponding recursive and Several papers are available in literature that deals non-recursive implementations respectively. The same with different implementations of decimation filter polyphase CIC filter can operate about 7 times faster architecture for sigma-delta ADCs. Hogenauer has than the recursive and about 3.7 times faster than the described the design procedures for decimation and non-recursive CIC filters.
Resumo:
While channel coding is a standard method of improving a system’s energy efficiency in digital communications, its practice does not extend to high-speed links. Increasing demands in network speeds are placing a large burden on the energy efficiency of high-speed links and render the benefit of channel coding for these systems a timely subject. The low error rates of interest and the presence of residual intersymbol interference (ISI) caused by hardware constraints impede the analysis and simulation of coded high-speed links. Focusing on the residual ISI and combined noise as the dominant error mechanisms, this paper analyses error correlation through concepts of error region, channel signature, and correlation distance. This framework provides a deeper insight into joint error behaviours in high-speed links, extends the range of statistical simulation for coded high-speed links, and provides a case against the use of biased Monte Carlo methods in this setting
Resumo:
The country has witnessed tremendous increase in the vehicle population and increased axle loading pattern during the last decade, leaving its road network overstressed and leading to premature failure. The type of deterioration present in the pavement should be considered for determining whether it has a functional or structural deficiency, so that appropriate overlay type and design can be developed. Structural failure arises from the conditions that adversely affect the load carrying capability of the pavement structure. Inadequate thickness, cracking, distortion and disintegration cause structural deficiency. Functional deficiency arises when the pavement does not provide a smooth riding surface and comfort to the user. This can be due to poor surface friction and texture, hydro planning and splash from wheel path, rutting and excess surface distortion such as potholes, corrugation, faulting, blow up, settlement, heaves etc. Functional condition determines the level of service provided by the facility to its users at a particular time and also the Vehicle Operating Costs (VOC), thus influencing the national economy. Prediction of the pavement deterioration is helpful to assess the remaining effective service life (RSL) of the pavement structure on the basis of reduction in performance levels, and apply various alternative designs and rehabilitation strategies with a long range funding requirement for pavement preservation. In addition, they can predict the impact of treatment on the condition of the sections. The infrastructure prediction models can thus be classified into four groups, namely primary response models, structural performance models, functional performance models and damage models. The factors affecting the deterioration of the roads are very complex in nature and vary from place to place. Hence there is need to have a thorough study of the deterioration mechanism under varied climatic zones and soil conditions before arriving at a definite strategy of road improvement. Realizing the need for a detailed study involving all types of roads in the state with varying traffic and soil conditions, the present study has been attempted. This study attempts to identify the parameters that affect the performance of roads and to develop performance models suitable to Kerala conditions. A critical review of the various factors that contribute to the pavement performance has been presented based on the data collected from selected road stretches and also from five corporations of Kerala. These roads represent the urban conditions as well as National Highways, State Highways and Major District Roads in the sub urban and rural conditions. This research work is a pursuit towards a study of the road condition of Kerala with respect to varying soil, traffic and climatic conditions, periodic performance evaluation of selected roads of representative types and development of distress prediction models for roads of Kerala. In order to achieve this aim, the study is focused into 2 parts. The first part deals with the study of the pavement condition and subgrade soil properties of urban roads distributed in 5 Corporations of Kerala; namely Thiruvananthapuram, Kollam, Kochi, Thrissur and Kozhikode. From selected 44 roads, 68 homogeneous sections were studied. The data collected on the functional and structural condition of the surface include pavement distress in terms of cracks, potholes, rutting, raveling and pothole patching. The structural strength of the pavement was measured as rebound deflection using Benkelman Beam deflection studies. In order to collect the details of the pavement layers and find out the subgrade soil properties, trial pits were dug and the in-situ field density was found using the Sand Replacement Method. Laboratory investigations were carried out to find out the subgrade soil properties, soil classification, Atterberg limits, Optimum Moisture Content, Field Moisture Content and 4 days soaked CBR. The relative compaction in the field was also determined. The traffic details were also collected by conducting traffic volume count survey and axle load survey. From the data thus collected, the strength of the pavement was calculated which is a function of the layer coefficient and thickness and is represented as Structural Number (SN). This was further related to the CBR value of the soil and the Modified Structural Number (MSN) was found out. The condition of the pavement was represented in terms of the Pavement Condition Index (PCI) which is a function of the distress of the surface at the time of the investigation and calculated in the present study using deduct value method developed by U S Army Corps of Engineers. The influence of subgrade soil type and pavement condition on the relationship between MSN and rebound deflection was studied using appropriate plots for predominant types of soil and for classified value of Pavement Condition Index. The relationship will be helpful for practicing engineers to design the overlay thickness required for the pavement, without conducting the BBD test. Regression analysis using SPSS was done with various trials to find out the best fit relationship between the rebound deflection and CBR, and other soil properties for Gravel, Sand, Silt & Clay fractions. The second part of the study deals with periodic performance evaluation of selected road stretches representing National Highway (NH), State Highway (SH) and Major District Road (MDR), located in different geographical conditions and with varying traffic. 8 road sections divided into 15 homogeneous sections were selected for the study and 6 sets of continuous periodic data were collected. The periodic data collected include the functional and structural condition in terms of distress (pothole, pothole patch, cracks, rutting and raveling), skid resistance using a portable skid resistance pendulum, surface unevenness using Bump Integrator, texture depth using sand patch method and rebound deflection using Benkelman Beam. Baseline data of the study stretches were collected as one time data. Pavement history was obtained as secondary data. Pavement drainage characteristics were collected in terms of camber or cross slope using camber board (slope meter) for the carriage way and shoulders, availability of longitudinal side drain, presence of valley, terrain condition, soil moisture content, water table data, High Flood Level, rainfall data, land use and cross slope of the adjoining land. These data were used for finding out the drainage condition of the study stretches. Traffic studies were conducted, including classified volume count and axle load studies. From the field data thus collected, the progression of each parameter was plotted for all the study roads; and validated for their accuracy. Structural Number (SN) and Modified Structural Number (MSN) were calculated for the study stretches. Progression of the deflection, distress, unevenness, skid resistance and macro texture of the study roads were evaluated. Since the deterioration of the pavement is a complex phenomena contributed by all the above factors, pavement deterioration models were developed as non linear regression models, using SPSS with the periodic data collected for all the above road stretches. General models were developed for cracking progression, raveling progression, pothole progression and roughness progression using SPSS. A model for construction quality was also developed. Calibration of HDM–4 pavement deterioration models for local conditions was done using the data for Cracking, Raveling, Pothole and Roughness. Validation was done using the data collected in 2013. The application of HDM-4 to compare different maintenance and rehabilitation options were studied considering the deterioration parameters like cracking, pothole and raveling. The alternatives considered for analysis were base alternative with crack sealing and patching, overlay with 40 mm BC using ordinary bitumen, overlay with 40 mm BC using Natural Rubber Modified Bitumen and an overlay of Ultra Thin White Topping. Economic analysis of these options was done considering the Life Cycle Cost (LCC). The average speed that can be obtained by applying these options were also compared. The results were in favour of Ultra Thin White Topping over flexible pavements. Hence, Design Charts were also plotted for estimation of maximum wheel load stresses for different slab thickness under different soil conditions. The design charts showed the maximum stress for a particular slab thickness and different soil conditions incorporating different k values. These charts can be handy for a design engineer. Fuzzy rule based models developed for site specific conditions were compared with regression models developed using SPSS. The Riding Comfort Index (RCI) was calculated and correlated with unevenness to develop a relationship. Relationships were developed between Skid Number and Macro Texture of the pavement. The effort made through this research work will be helpful to highway engineers in understanding the behaviour of flexible pavements in Kerala conditions and for arriving at suitable maintenance and rehabilitation strategies. Key Words: Flexible Pavements – Performance Evaluation – Urban Roads – NH – SH and other roads – Performance Models – Deflection – Riding Comfort Index – Skid Resistance – Texture Depth – Unevenness – Ultra Thin White Topping
Resumo:
Die Maßnahmen zur Förderung der Windenergie in Deutschland haben wichtige Anstöße zur technologischen Weiterentwicklung geliefert und die Grundlagen für den enormen Anlagenzubau geschaffen. Die installierte Windleistung hat heute eine beachtliche Größenordnung erreicht und ein weiteres Wachstum in ähnlichen Dimensionen ist auch für die nächsten Jahre zu erwarten. Die aus Wind erzeugte elektrische Leistung deckt bereits heute in einigen Netzbereichen die Netzlast zu Schwachlastzeiten. Dies zeigt, dass die Windenergie ein nicht mehr zu vernachlässigender Faktor in der elektrischen Energieversorgung geworden ist. Im Rahmen der Kraftwerkseinsatzplanung sind Betrag und Verlauf der Windleistung des folgenden Tages mittlerweile zu wichtigen und zugleich schwierig zu bestimmenden Variablen geworden. Starke Schwankungen und falsche Prognosen der Windstromeinspeisung verursachen zusätzlichen Bedarf an Regel- und Ausgleichsleistung durch die Systemführung. Das im Rahmen dieser Arbeit entwickelte Prognosemodell liefert die zu erwartenden Windleistungen an 16 repräsentativen Windparks bzw. Gruppen von Windparks für bis zu 48 Stunden im Voraus. Aufgrund von prognostizierten Wetterdaten des deutschen Wetterdienstes (DWD) werden die Leistungen der einzelnen Windparks mit Hilfe von künstlichen neuronalen Netzen (KNN) berechnet. Diese Methode hat gegenüber physikalischen Verfahren den Vorteil, dass der komplexe Zusammenhang zwischen Wettergeschehen und Windparkleistung nicht aufwendig analysiert und detailliert mathematisch beschrieben werden muss, sondern anhand von Daten aus der Vergangenheit von den KNN gelernt wird. Das Prognosemodell besteht aus zwei Modulen. Mit dem ersten wird, basierend auf den meteorologischen Vorhersagen des DWD, eine Prognose für den Folgetag erstellt. Das zweite Modul bezieht die online gemessenen Leistungsdaten der repräsentativen Windparks mit ein, um die ursprüngliche Folgetagsprognose zu verbessern und eine sehr genaue Kurzzeitprognose für die nächsten drei bis sechs Stunden zu berechnen. Mit den Ergebnissen der Prognosemodule für die repräsentativen Standorte wird dann über ein Transformationsmodell, dem so genannten Online-Modell, die Gesamteinspeisung in einem größeren Gebiet berechnet. Das Prognoseverfahren hat seine besonderen Vorzüge in der Genauigkeit, den geringen Rechenzeiten und den niedrigen Betriebskosten, da durch die Verwendung des bereits implementierten Online-Modells nur eine geringe Anzahl von Vorhersage- und Messstandorten benötigt wird. Das hier vorgestellte Prognosemodell wurde ursprünglich für die E.ON-Netz GmbH entwickelt und optimiert und ist dort seit Juli 2001 im Einsatz. Es lässt sich jedoch auch leicht an andere Gebiete anpassen. Benötigt werden dazu nur die Messdaten der Leistung ausgewählter repräsentativer Windparks sowie die dazu gehörenden Wettervorhersagen, um die KNN entsprechend zu trainieren.
Resumo:
Ein Luft-Erdwärmetauscher (L-EWT) kommt wegen seines niedrigen Energiebedarfs und möglicher guter Aufwandszahlen als umweltfreundliche Versorgungskomponente für Gebäude in Betracht. Dabei ist besonders vorteilhaft, dass ein L-EWT die Umgebungsluft je nach Jahreszeit vorwärmen oder auch kühlen kann. Dem zufolge sind L-EWT zur Energieeinsparung nicht nur für den Wohnhausbau interessant, sondern auch dort, wo immer noch große Mengen an fossiler Energie für die Raumkühlung benötigt werden, im Büro- und Produktionsgebäudesektor. Der Einsatzbereich eines L-EWT liegt zwischen Volumenströmen von 100 m3/h und mehreren 100.000 m3/h. Aus dieser Bandbreite und den instationären Randbedingungen entstehen erhebliche Schwierigkeiten, allgemeingültige Aussagen über das zu erwartende thermische Systemverhalten aus der Vielzahl möglicher Konstruktionsvarianten zu treffen. Hauptziel dieser Arbeit ist es, auf Basis umfangreicher, mehrjähriger Messungen an einer eigens konzipierten Testanlage und eines speziell angepassten numerischen Rechenmodells, Kennzahlen zu entwickeln, die es ermöglichen, die Betriebseigenschaften eines L-EWT im Planungsalltag zu bestimmen und ein technisch, ökologisch wie ökonomisch effizientes System zu identifizieren. Es werden die Kennzahlen elewt (Aufwandszahl), QV (Netto-Volumenleistung), ME (Meterertrag), sowie die Kombination aus v (Strömungsgeschwindigkeit) und VL (Metervolumenstrom) definiert, die zu wichtigen Informationen führen, mit denen die Qualität von Systemvarianten in der Planungsphase bewertet werden können. Weiterführende Erkenntnisse über die genauere Abschätzung von Bodenkennwerten werden dargestellt. Die hygienische Situation der durch den L-EWT transportierten Luft wird für die warme Jahreszeit, aufgrund auftretender Tauwasserbildung, beschrieben. Aus diesem Grund werden alle relevanten lufthygienischen Parameter in mehreren aufwendigen Messkampagnen erfasst und auf pathogene Wirkungen überprüft. Es wird über Sensitivitätsanalysen gezeigt, welche Fehler bei Annahme falscher Randbedingungen eintreten. Weiterhin werden in dieser Arbeit wesentliche, grundsätzliche Erkenntnisse aufbereitet, die sich aus der Betriebsbeobachtung und der Auswertung der umfangreich vorliegenden Messdaten mehrerer Anlagen ergeben haben und für die praktische Umsetzung und die Betriebsführung bedeutend sind. Hinweise zu Materialeigenschaften und zur Systemwirtschaftlichkeit sind detailliert aufgeführt.
Resumo:
Scanning Probe Microscopy (SPM) has become of fundamental importance for research in area of micro and nano-technology. The continuous progress in these fields requires ultra sensitive measurements at high speed. The imaging speed limitation of the conventional Tapping Mode SPM is due to the actuation time constant of piezotube feedback loop that keeps the tapping amplitude constant. In order to avoid this limit a deflection sensor and an actuator have to be integrated into the cantilever. In this work has been demonstrated the possibility of realisation of piezoresistive cantilever with an embedded actuator. Piezoresistive detection provides a good alternative to the usual optical laser beam deflection technique. In frames of this thesis has been investigated and modelled the piezoresistive effect in bulk silicon (3D case) for both n- and p-type silicon. Moving towards ultra-sensitive measurements it is necessary to realize ultra-thin piezoresistors, which are well localized to the surface, where the stress magnitude is maximal. New physical effects such as quantum confinement which arise due to the scaling of the piezoresistor thickness was taken into account in order to model the piezoresistive effect and its modification in case of ultra-thin piezoresistor (2D case). The two-dimension character of the electron gas in n-type piezoresistors lead up to decreasing of the piezoresistive coefficients with increasing the degree of electron localisation. Moreover for p-type piezoresistors the predicted values of the piezoresistive coefficients are higher in case of localised holes. Additionally, to the integration of the piezoresistive sensor, actuator integrated into the cantilever is considered as fundamental for realisation of fast SPM imaging. Actuation of the beam is achieved thermally by relying on differences in the coefficients of thermal expansion between aluminum and silicon. In addition the aluminum layer forms the heating micro-resistor, which is able to accept heating impulses with frequency up to one megahertz. Such direct oscillating thermally driven bimorph actuator was studied also with respect to the bimorph actuator efficiency. Higher eigenmodes of the cantilever are used in order to increase the operating frequencies. As a result the scanning speed has been increased due to the decreasing of the actuation time constant. The fundamental limits to force sensitivity that are imposed by piezoresistive deflection sensing technique have been discussed. For imaging in ambient conditions the force sensitivity is limited by the thermo-mechanical cantilever noise. Additional noise sources, connected with the piezoresistive detection are negligible.
Resumo:
Excimerlaser sind gepulste Gaslaser, die Laseremission in Form von Linienstrahlung – abhängig von der Gasmischung – im UV erzeugen. Der erste entladungsgepumpte Excimerlaser wurde 1977 von Ischenko demonstriert. Alle kommerziell verfügbaren Excimerlaser sind entladungsgepumpte Systeme. Um eine Inversion der Besetzungsdichte zu erhalten, die notwendig ist, um den Laser zum Anschwingen zu bekommen, muss aufgrund der kurzen Wellenlänge sehr stark gepumpt werden. Diese Pumpleistung muss von einem Impulsleistungsmodul erzeugt werden. Als Schaltelement gebräuchlich sind Thyratrons, Niederdruckschaltröhren, deren Lebensdauer jedoch sehr limitiert ist. Deshalb haben sich seit Mitte der 1990iger Jahre Halbleiterschalter mit Pulskompressionsstufen auch in dieser Anwendung mehr und mehr durchgesetzt. In dieser Arbeit wird versucht, die Pulskompression durch einen direkt schaltenden Halbleiterstapel zu ersetzen und dadurch die Verluste zu reduzieren sowie den Aufwand für diese Pulskompression einzusparen. Zudem kann auch die maximal mögliche Repetitionsrate erhöht werden. Um die Belastung der Bauelemente zu berechnen, wurden für alle Komponenten möglichst einfache, aber leistungsfähige Modelle entwickelt. Da die normalerweise verfügbaren Daten der Bauelemente sich aber auf andere Applikationen beziehen, mussten für alle Bauteile grundlegende Messungen im Zeitbereich der späteren Applikation gemacht werden. Für die nichtlinearen Induktivitäten wurde ein einfaches Testverfahren entwickelt um die Verluste bei sehr hohen Magnetisierungsgeschwindigkeiten zu bestimmen. Diese Messungen sind die Grundlagen für das Modell, das im Wesentlichen eine stromabhängige Induktivität beschreibt. Dieses Modell wurde für den „magnetic assist“ benützt, der die Einschaltverluste in den Halbleitern reduziert. Die Impulskondensatoren wurden ebenfalls mit einem in der Arbeit entwickelten Verfahren nahe den späteren Einsatzparametern vermessen. Dabei zeigte sich, dass die sehr gebräuchlichen Class II Keramikkondensatoren für diese Anwendung nicht geeignet sind. In der Arbeit wurden deshalb Class I Hochspannungs- Vielschicht- Kondensatoren als Speicherbank verwendet, die ein deutlich besseres Verhalten zeigen. Die eingesetzten Halbleiterelemente wurden ebenfalls in einem Testverfahren nahe den späteren Einsatzparametern vermessen. Dabei zeigte sich, dass nur moderne Leistungs-MOSFET´s für diesen Einsatz geeignet sind. Bei den Dioden ergab sich, dass nur Siliziumkarbid (SiC) Schottky Dioden für die Applikation einsetzbar sind. Für die Anwendung sind prinzipiell verschiedene Topologien möglich. Bei näherer Betrachtung zeigt sich jedoch, dass nur die C-C Transfer Anordnung die gewünschten Ergebnisse liefern kann. Diese Topologie wurde realisiert. Sie besteht im Wesentlichen aus einer Speicherbank, die vom Netzteil aufgeladen wird. Aus dieser wird dann die Energie in den Laserkopf über den Schalter transferiert. Aufgrund der hohen Spannungen und Ströme müssen 24 Schaltelemente in Serie und je 4 parallel geschaltet werden. Die Ansteuerung der Schalter wird über hochisolierende „Gate“-Transformatoren erreicht. Es zeigte sich, dass eine sorgfältig ausgelegte dynamische und statische Spannungsteilung für einen sicheren Betrieb notwendig ist. In der Arbeit konnte ein Betrieb mit realer Laserkammer als Last bis 6 kHz realisiert werden, der nur durch die maximal mögliche Repetitionsrate der Laserkammer begrenzt war.
Resumo:
High-speed semiconductor lasers are an integral part in the implemen- tation of high-bit-rate optical communications systems. They are com- pact, rugged, reliable, long-lived, and relatively inexpensive sources of coherent light. Due to the very low attenuation window that exists in the silica based optical fiber at 1.55 μm and the zero dispersion point at 1.3 μm, they have become the mainstay of optical fiber com- munication systems. For the fabrication of lasers with gratings such as, distributed bragg reflector or distributed feedback lasers, etching is the most critical step. Etching defines the lateral dimmensions of the structure which determines the performance of optoelectronic devices. In this thesis studies and experiments were carried out about the exist- ing etching processes for InP and a novel dry etching process was de- veloped. The newly developed process was based on Cl2/CH4/H2/Ar chemistry and resulted in very smooth surfaces and vertical side walls. With this process the grating definition was significantly improved as compared to other technological developments in the respective field. A surface defined grating definition approach is used in this thesis work which does not require any re-growth steps and makes the whole fabrication process simpler and cost effective. Moreover, this grating fabrication process is fully compatible with nano-imprint lithography and can be used for high throughput low-cost manufacturing. With usual etching techniques reported before it is not possible to etch very deep because of aspect ratio dependent etching phenomenon where with increasing etch depth the etch rate slows down resulting in non-vertical side walls and footing effects. Although with our de- veloped process quite vertical side walls were achieved but footing was still a problem. To overcome the challenges related to grating defini- tion and deep etching, a completely new three step gas chopping dry etching process was developed. This was the very first time that a time multiplexed etching process for an InP based material system was demonstrated. The developed gas chopping process showed extra ordinary results including high mask selectivity of 15, moderate etch- ing rate, very vertical side walls and a record high aspect ratio of 41. Both the developed etching processes are completely compatible with nano imprint lithography and can be used for low-cost high-throughput fabrication. A large number of broad area laser, ridge waveguide laser, distributed feedback laser, distributed bragg reflector laser and coupled cavity in- jection grating lasers were fabricated using the developed one step etch- ing process. Very extensive characterization was done to optimize all the important design and fabrication parameters. The devices devel- oped have shown excellent performance with a very high side mode suppression ratio of more than 52 dB, an output power of 17 mW per facet, high efficiency of 0.15 W/A, stable operation over temperature and injected currents and a threshold current as low as 30 mA for almost 1 mm long device. A record high modulation bandwidth of 15 GHz with electron-photon resonance and open eye diagrams for 10 Gbps data transmission were also shown.
Resumo:
The performances of high-speed network communications frequently rest with the distribution of data-stream. In this paper, a dynamic data-stream balancing architecture based on link information is introduced and discussed firstly. Then the algorithms for simultaneously acquiring the passing nodes and links of a path between any two source-destination nodes rapidly, as well as a dynamic data-stream distribution planning are proposed. Some related topics such as data fragment disposal, fair service, etc. are further studied and discussed. Besides, the performance and efficiency of proposed algorithms, especially for fair service and convergence, are evaluated through a demonstration with regard to the rate of bandwidth utilization. Hoping the discussion presented here can be helpful to application developers in selecting an effective strategy for planning the distribution of data-stream.
Resumo:
Resumen tomado de la publicaci??n
Resumo:
From the Brookshear Chapter on Operating Systems
Resumo:
Based on Brookshear textbook (Pearson copyright) files but with additional slides by hcd
Resumo:
Avanzalogistic corresponde a una plataforma multilateral que hace parte de las nuevas tecnologías de información y servicios, por lo cual se considera como una nueva opción para la ciudad de Bogotá, enfocada en dar solución al transporte de mercancía para las pequeñas y medianas empresas de un sector de la ciudad. Teniendo como diferenciación, la asistencia y atención personalizada para este tipo de clientes, donde por medio de la configuración de carga en tiempo real y con la facilidad virtual, podrán recibir el mejor servicio, en el menor tiempo posible y con los mejores precios del mercado. Considerando las necesidades de confiabilidad, seguridad, rapidez y buena atención que exige nuestro segmento objetivo, Avanzalogistic propone un servicio con calidad integral, donde por medio del desempeño enfocado al cliente y el cubrimiento de las necesidades del mismo será potencialmente reconocido como una de las mejores opciones del mercado para las empresas que soliciten transportar su carga de manera rápida y segura, disminuyendo el tiempo de respuesta de 24 horas (tiempo actual de respuesta de la competencia) a instantáneamente (tiempo por Avanzalogistic). Teniendo en cuenta lo anterior, en el proyecto se decide realizar una inversión inicial de $ 135 627 200 pesos colombianos, alcanzando su punto de equilibrio a partir del mes seis del primer año y teniendo un periodo de recuperación representado a partir del segundo año. Avanzalogistic es un proyecto que presenta una rentabilidad del 68,51%, justificada por la baja inversión y los bajos costos de operación, teniendo como principal ventaja la tercerización de los vehículos, además de la gran oportunidad de tener un mercado potencial lo suficientemente amplio que evidencia un ingreso que justifica la utilidad en cada servicio prestado. Siendo así un proyecto no solo viable, sino también con posibilidad de crecimiento y atractivo para invertir.