966 resultados para Processos de back-end
Resumo:
A indústria de semicondutores é um sector em permanente evolução tecnológica. A tendência de miniaturização e de otimização do espaço, a necessidade de produzir circuitos cada vez mais complexos, a tendência para o incremento do número de camadas em cada circuito integrado, são as condições necessárias para que a evolução tecnológica nesta área seja uma constante. Os processos ligados à produção de semicondutores estão também em permanente evolução, dada a pressão efetuada pelas necessidades acima expostas. Os equipamentos necessitam de uma crescente precisão, a qual tem que ser acompanhada de procedimentos rigorosos para que a qualidade atingida tenha sempre o patamar desejado. No entanto, a constante evolução nem sempre permite um adequado levantamento de todas as causas que estão na origem de alguns problemas detetados na fabricação de semicondutores. Este trabalho teve por objetivo efetuar um levantamento dos processos ligados ao fabrico de semicondutores a partir de uma pastilha de silício (wafer) previamente realizada, identificando para cada processo os possíveis defeitos introduzidos pelo mesmo, procurando inventariar as causas possíveis que possam estar na origem desse defeito e realizar procedimentos que permitam criar regras e procedimentos perfeitamente estabelecidos que permitam aprender com os erros e evitar que os mesmos problemas se possam vir a repetir em situações análogas em outros produtos de uma mesma família.
Resumo:
I sistemi di identificazione tramite radiofrequenza hanno acquistato negli ultimi tempi una sempre maggiore importanza per il mondo produttivo, soprattutto nel settore della movimentazione delle merci, evolvendo verso applicazioni di tracciamento sempre più avanzate. La tesi si propone di analizzare in dettaglio la tecnologia RFId, chiarendone gli aspetti fondamentali ed esponendo criticità e punti di forza.
Resumo:
Il web ha cambiato radicalmente le nostre vite. Grazie ad esso, oggi si possono fare cose che solo qualche decennio fa erano pura fantascienza, come ad esempio la telepresenza o gli interventi chirurgici da remoto, ma anche cose più “semplici” come seguire corsi di formazione (anche universitaria), effettuare la spesa, operare con il proprio conto corrente, tutto restando comodamente a casa propria, semplificando così la vita di tutti. Allo stesso tempo il web è stato utilizzato per fini tutt’altro che nobili, ad esempio per commettere crimini informatici, recare danni alla concorrenza, compiere varie forme di truffe ecc. Ogni persona dovrebbe comportarsi in modo corretto e nel pieno rispetto del prossimo, sia sul mondo reale che sul web, ma purtroppo non è sempre così. Per quanto riguarda il mondo del web, sta agli sviluppatori soddisfare le necessità dei propri utenti, assicurandosi però che la propria applicazione non verrà usata per recare qualche tipo di danno a terzi o alla propria infrastruttura. Questa tesi nasce da un idea dei docenti del corso di Programmazione riguardo alla realizzazione di un modulo del sito web del corso che si occupa della correzione automatica di esercizi scritti in linguaggio C dagli studenti del corso, dove per correzione automatica si intende la verifica della correttezza degli esercizi.
Resumo:
Questa tesi tratta dello sviluppo di un progetto, svolto durante il periodo di tirocinio presso SMS.it, azienda specializzata nel settore della telefonia con sede a Bologna. L'azienda in questione ha commissionato, al sottoscritto ed al mio collega Daniele Sciuto, l'implementazione di un'applicazione cross-platform per smartphone, ed il relativo server. L'azienda ci ha fornito le specifiche del progetto, e ci ha seguiti in tutte le fasi del suo sviluppo. L'applicazione è pensata per offrire agli utenti la possibilità di usufruire di tariffe telefoniche agevolate. I vantaggi sono maggiormente apprezzabili nelle chiamate internazionali. Queste tariffe sono possibili grazie agli accordi fra l'azienda e vari operatori di telefonia. Nella primo capitolo di questo elaborato, viene analizzato cosa ci è stato richiesto di realizzare, le specifiche del progetto dateci dall'azienda e quali sono i vincoli ai quali ci si è dovuti attenere. Nella secondo capitolo, viene descritto nel dettaglio la progettazione delle singole funzionalità dell'applicazione, e i rapporti che ci sono fra il front-end ed il back-end. Successivamente, sono analizzate le tecnologie necessarie per la realizzazione e il loro utilizzo nell'applicazione. Come richiestoci dall'azienda, alcuni dettagli implementativi sono stati omessi, per garantire il rispetto del segreto industriale. Nonostante ciò viene comunque fornita una panoramica completa di ciò che è stato realizzato. In ultima analisi è descritta qualitativamente l'applicazione ottenuta, e come aderisca alle specifiche richieste.
Resumo:
L'obiettivo di questa tesi è lo sviluppo di due applicativi per l'azienda EBWorld. Il primo desktop, sviluppato in Java, è utilizzato per gestire la memoria interna di un dispositivo Android collegato al computer, installare l'applicazione mobile sviluppata a seguito ed esportare i progetti, creati dall'utente, durante l'utilizzo dell'applicativo mobile. Il secondo è un applicativo Android, utilizzato per la visualizzazione e l'interazione con dati georiferiti e permette all'utente di creare progetti inserendo ulteriori elementi georiferiti. Per la memorizzazione dei dati sono stati utilizzati diversi file XML e per la comunicazione tra l'applicativo Java ed un dispositivo Android, sono stati utilizzati la libreria JMTP in combinazione con i comandi ADB.
Resumo:
A high-frequency-link micro inverter is proposed with a front-end dual inductor push-pull converter and a grid-connected half-wave cycloconverter. Pulse width modulation is used to control the front-end converter and phase shift modulation is used at the back-end converter to obtain grid synchronized output current. A series resonant circuit and high-frequency transformer are used to interface the front-end and the back-end converters. The operation of the proposed micro-inverter in grid-connected mode is validated using MATLAB/Simpower simulation. Experimental results are provided to further validate the operation.
Resumo:
The future Internet is expected to be composed of a mesh of interoperable web services accessible from all over the web. This approach has not yet caught on since global user?service interaction is still an open issue. This paper states one vision with regard to next-generation front-end Web 2.0 technology that will enable integrated access to services, contents and things in the future Internet. In this paper, we illustrate how front-ends that wrap traditional services and resources can be tailored to the needs of end users, converting end users into prosumers (creators and consumers of service-based applications). To do this, we propose an architecture that end users without programming skills can use to create front-ends, consult catalogues of resources tailored to their needs, easily integrate and coordinate front-ends and create composite applications to orchestrate services in their back-end. The paper includes a case study illustrating that current user-centred web development tools are at a very early stage of evolution. We provide statistical data on how the proposed architecture improves these tools. This paper is based on research conducted by the Service Front End (SFE) Open Alliance initiative.
Resumo:
One of the classic forms of intermediate representation used for communication between compiler front-ends and back-ends are those based on abstract stack machines. It is possible to compile the stack machine instructions into machine code by means of an interpretive code generator, or to simulate the stack machine at runtime using an interpreter. This paper describes an approach intermediate between these two extremes. The front-end for a commercial Modula 2 compiler was ported to the "industry standard PC", and a partially compiling back-end written. The object code runs with the assistance of an interpreter, but may be linked with libraries which are fully compiled. The intent was to provide a programming environment on the PC which is identical to that of the same compilers on 32-bit UNIX machines. This objective has been met, and the compiler is available to educational institutions as free-ware. The design basis of the new compiler is described, and the performance critically evaluated.
Resumo:
Inspection of solder joints has been a critical process in the electronic manufacturing industry to reduce manufacturing cost, improve yield, and ensure product quality and reliability. The solder joint inspection problem is more challenging than many other visual inspections because of the variability in the appearance of solder joints. Although many research works and various techniques have been developed to classify defect in solder joints, these methods have complex systems of illumination for image acquisition and complicated classification algorithms. An important stage of the analysis is to select the right method for the classification. Better inspection technologies are needed to fill the gap between available inspection capabilities and industry systems. This dissertation aims to provide a solution that can overcome some of the limitations of current inspection techniques. This research proposes two inspection steps for automatic solder joint classification system. The “front-end” inspection system includes illumination normalisation, localization and segmentation. The illumination normalisation approach can effectively and efficiently eliminate the effect of uneven illumination while keeping the properties of the processed image. The “back-end” inspection involves the classification of solder joints by using Log Gabor filter and classifier fusion. Five different levels of solder quality with respect to the amount of solder paste have been defined. Log Gabor filter has been demonstrated to achieve high recognition rates and is resistant to misalignment. Further testing demonstrates the advantage of Log Gabor filter over both Discrete Wavelet Transform and Discrete Cosine Transform. Classifier score fusion is analysed for improving recognition rate. Experimental results demonstrate that the proposed system improves performance and robustness in terms of classification rates. This proposed system does not need any special illumination system, and the images are acquired by an ordinary digital camera. In fact, the choice of suitable features allows one to overcome the problem given by the use of non complex illumination systems. The new system proposed in this research can be incorporated in the development of an automated non-contact, non-destructive and low cost solder joint quality inspection system.
Resumo:
Inspection of solder joints has been a critical process in the electronic manufacturing industry to reduce manufacturing cost, improve yield, and ensure product quality and reliability. This paper proposes two inspection modules for an automatic solder joint classification system. The “front-end” inspection system includes illumination normalisation, localisation and segmentation. The “back-end” inspection involves the classification of solder joints using the Log Gabor filter and classifier fusion. Five different levels of solder quality with respect to the amount of solder paste have been defined. The Log Gabor filter has been demonstrated to achieve high recognition rates and is resistant to misalignment. This proposed system does not need any special illumination system, and the images are acquired by an ordinary digital camera. This system could contribute to the development of automated non-contact, non-destructive and low cost solder joint quality inspection systems.
Resumo:
Queensland University of Technology (QUT) was one of the first universities in Australia to establish an institutional repository. Launched in November 2003, the repository (QUT ePrints) uses the EPrints open source repository software (from Southampton) and has enjoyed the benefit of an institutional deposit mandate since January 2004. Currently (April 2012), the repository holds over 36,000 records, including 17,909 open access publications with another 2,434 publications embargoed but with mediated access enabled via the ‘Request a copy’ button which is a feature of the EPrints software. At QUT, the repository is managed by the library.QUT ePrints (http://eprints.qut.edu.au) The repository is embedded into a number of other systems at QUT including the staff profile system and the University’s research information system. It has also been integrated into a number of critical processes related to Government reporting and research assessment. Internally, senior research administrators often look to the repository for information to assist with decision-making and planning. While some statistics could be drawn from the advanced search feature and the existing download statistics feature, they were rarely at the level of granularity or aggregation required. Getting the information from the ‘back end’ of the repository was very time-consuming for the Library staff. In 2011, the Library funded a project to enhance the range of statistics which would be available from the public interface of QUT ePrints. The repository team conducted a series of focus groups and individual interviews to identify and prioritise functionality requirements for a new statistics ‘dashboard’. The participants included a mix research administrators, early career researchers and senior researchers. The repository team identified a number of business criteria (eg extensible, support available, skills required etc) and then gave each a weighting. After considering all the known options available, five software packages (IRStats, ePrintsStats, AWStats, BIRT and Google Urchin/Analytics) were thoroughly evaluated against a list of 69 criteria to determine which would be most suitable. The evaluation revealed that IRStats was the best fit for our requirements. It was deemed capable of meeting 21 out of the 31 high priority criteria. Consequently, IRStats was implemented as the basis for QUT ePrints’ new statistics dashboards which were launched in Open Access Week, October 2011. Statistics dashboards are now available at four levels; whole-of-repository level, organisational unit level, individual author level and individual item level. The data available includes, cumulative total deposits, time series deposits, deposits by item type, % fulltexts, % open access, cumulative downloads, time series downloads, downloads by item type, author ranking, paper ranking (by downloads), downloader geographic location, domains, internal v external downloads, citation data (from Scopus and Web of Science), most popular search terms, non-search referring websites. The data is displayed in charts, maps and table format. The new statistics dashboards are a great success. Feedback received from staff and students has been very positive. Individual researchers have said that they have found the information to be very useful when compiling a track record. It is now very easy for senior administrators (including the Deputy Vice Chancellor-Research) to compare the full-text deposit rates (i.e. mandate compliance rates) across organisational units. This has led to increased ‘encouragement’ from Heads of School and Deans in relation to the provision of full-text versions.
Resumo:
The continuing need for governments to radically improve the delivery of public services has led to a new, holistic government reform strategy labeled “Transformational Government” that strongly emphasizes customer-centricity. Attention has turned to online portals as a cost effective front-end to deliver services and engage customers as well as to the corresponding organizational approaches for the back-end to decouple the service interface from the departmental structures. The research presented in this paper makes three contributions: Firstly, a systematic literature review of approaches to the evaluation of online portal models in the public sector is presented. Secondly, the findings of a usability study comparing the online presences of the Queensland Government, the UK Government and the South Australian Government are reported and the relative strengths and weaknesses of the different approaches are discussed. And thirdly, the limitations of the usability study in the context of a broader “Transformational Government” approach are identified and service bundling is suggested as an innovative solution to further improve online service delivery.
Resumo:
It has long been a concern that the wider uptake of the YAWL environment may have been hindered by the usability issues identified in the current Process Editor. As a consequence, it was decided that the Editor be completely rewritten to address those usability limitations. The result has been the implementation of a new YAWL Process Editor architecture that creates a clear separation between the User Interface component layer and the core processing back end, facilitating the redesign of the default user interface. This new architecture also supports the development of multiple User Interface front ends for specific contexts that take advantage of the core capabilities the new Editor architecture has to offer.
Resumo:
Whole-image descriptors such as GIST have been used successfully for persistent place recognition when combined with temporal filtering or sequential filtering techniques. However, whole-image descriptor localization systems often apply a heuristic rather than a probabilistic approach to place recognition, requiring substantial environmental-specific tuning prior to deployment. In this paper we present a novel online solution that uses statistical approaches to calculate place recognition likelihoods for whole-image descriptors, without requiring either environmental tuning or pre-training. Using a real world benchmark dataset, we show that this method creates distributions appropriate to a specific environment in an online manner. Our method performs comparably to FAB-MAP in raw place recognition performance, and integrates into a state of the art probabilistic mapping system to provide superior performance to whole-image methods that are not based on true probability distributions. The method provides a principled means for combining the powerful change-invariant properties of whole-image descriptors with probabilistic back-end mapping systems without the need for prior training or system tuning.