970 resultados para Black box approach


Relevância:

100.00% 100.00%

Publicador:

Resumo:

A simplified (without phase modulator) scheme of a black box optical regenerator is proposed, where an appropriate nonlinear propagation is used to enhance regeneration. Applying semi-theoretical models the authors optimise and demonstrate feasibility of error-free long distance transmission at 40 Gbit/s.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Artificial Intelligence (AI) and Machine Learning (ML) are novel data analysis techniques providing very accurate prediction results. They are widely adopted in a variety of industries to improve efficiency and decision-making, but they are also being used to develop intelligent systems. Their success grounds upon complex mathematical models, whose decisions and rationale are usually difficult to comprehend for human users to the point of being dubbed as black-boxes. This is particularly relevant in sensitive and highly regulated domains. To mitigate and possibly solve this issue, the Explainable AI (XAI) field became prominent in recent years. XAI consists of models and techniques to enable understanding of the intricated patterns discovered by black-box models. In this thesis, we consider model-agnostic XAI techniques, which can be applied to Tabular data, with a particular focus on the Credit Scoring domain. Special attention is dedicated to the LIME framework, for which we propose several modifications to the vanilla algorithm, in particular: a pair of complementary Stability Indices that accurately measure LIME stability, and the OptiLIME policy which helps the practitioner finding the proper balance among explanations' stability and reliability. We subsequently put forward GLEAMS a model-agnostic surrogate interpretable model which requires to be trained only once, while providing both Local and Global explanations of the black-box model. GLEAMS produces feature attributions and what-if scenarios, from both dataset and model perspective. Eventually, we argue that synthetic data are an emerging trend in AI, being more and more used to train complex models instead of original data. To be able to explain the outcomes of such models, we must guarantee that synthetic data are reliable enough to be able to translate their explanations to real-world individuals. To this end we propose DAISYnt, a suite of tests to measure synthetic tabular data quality and privacy.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Alors que certains mécanismes pourtant jugés cruciaux pour la transformation de la pluie en débit restent peu ou mal compris, le concept de connectivité hydrologique a récemment été proposé pour expliquer pourquoi certains processus sont déclenchés de manière épisodique en fonction des caractéristiques des événements de pluie et de la teneur en eau des sols avant l’événement. L’adoption de ce nouveau concept en hydrologie reste cependant difficile puisqu’il n’y a pas de consensus sur la définition de la connectivité, sa mesure, son intégration dans les modèles hydrologiques et son comportement lors des transferts d’échelles spatiales et temporelles. Le but de ce travail doctoral est donc de préciser la définition, la mesure, l’agrégation et la prédiction des processus liés à la connectivité hydrologique en s’attardant aux questions suivantes : 1) Quel cadre méthodologique adopter pour une étude sur la connectivité hydrologique ?, 2) Comment évaluer le degré de connectivité hydrologique des bassins versants à partir de données de terrain ?, et 3) Dans quelle mesure nos connaissances sur la connectivité hydrologique doivent-elles conduire à la modification des postulats de modélisation hydrologique ? Trois approches d’étude sont différenciées, soit i) une approche de type « boite noire », basée uniquement sur l’exploitation des données de pluie et de débits sans examiner le fonctionnement interne du bassin versant ; ii) une approche de type « boite grise » reposant sur l’étude de données géochimiques ponctuelles illustrant la dynamique interne du bassin versant ; et iii) une approche de type « boite blanche » axée sur l’analyse de patrons spatiaux exhaustifs de la topographie de surface, la topographie de subsurface et l’humidité du sol. Ces trois approches sont ensuite validées expérimentalement dans le bassin versant de l’Hermine (Basses Laurentides, Québec). Quatre types de réponses hydrologiques sont distingués en fonction de leur magnitude et de leur synchronisme, sachant que leur présence relative dépend des conditions antécédentes. Les forts débits enregistrés à l’exutoire du bassin versant sont associés à une contribution accrue de certaines sources de ruissellement, ce qui témoigne d’un lien hydraulique accru et donc d’un fort degré de connectivité hydrologique entre les sources concernées et le cours d’eau. Les aires saturées couvrant des superficies supérieures à 0,85 ha sont jugées critiques pour la genèse de forts débits de crue. La preuve est aussi faite que les propriétés statistiques des patrons d’humidité du sol en milieu forestier tempéré humide sont nettement différentes de celles observées en milieu de prairie tempéré sec, d’où la nécessité d’utiliser des méthodes de calcul différentes pour dériver des métriques spatiales de connectivité dans les deux types de milieux. Enfin, la double existence de sources contributives « linéaires » et « non linéaires » est mise en évidence à l’Hermine. Ces résultats suggèrent la révision de concepts qui sous-tendent l’élaboration et l’exécution des modèles hydrologiques. L’originalité de cette thèse est le fait même de son sujet. En effet, les objectifs de recherche poursuivis sont conformes à la théorie hydrologique renouvelée qui prône l’arrêt des études de particularismes de petite échelle au profit de l’examen des propriétés émergentes des bassins versants telles que la connectivité hydrologique. La contribution majeure de cette thèse consiste ainsi en la proposition d’une définition unifiée de la connectivité, d’un cadre méthodologique, d’approches de mesure sur le terrain, d’outils techniques et de pistes de solution pour la modélisation des systèmes hydrologiques.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this contribution, a system identification procedure of a two-input Wiener model suitable for the analysis of the disturbance behavior of integrated nonlinear circuits is presented. The identified block model is comprised of two linear dynamic and one static nonlinear block, which are determined using an parameterized approach. In order to characterize the linear blocks, an correlation analysis using a white noise input in combination with a model reduction scheme is adopted. After having characterized the linear blocks, from the output spectrum under single tone excitation at each input a linear set of equations will be set up, whose solution gives the coefficients of the nonlinear block. By this data based black box approach, the distortion behavior of a nonlinear circuit under the influence of an interfering signal at an arbitrary input port can be determined. Such an interfering signal can be, for example, an electromagnetic interference signal which conductively couples into the port of consideration. © 2011 Author(s).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

How we get from transistors through to logic gates to ALUs and memory to the stored program and the fetch execute cycle through to machine code and high level languages. Inspired by Tanenbaum's approach in "Structured Computer Organozation"

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The behavior of composed Web services depends on the results of the invoked services; unexpected behavior of one of the invoked services can threat the correct execution of an entire composition. This paper proposes an event-based approach to black-box testing of Web service compositions based on event sequence graphs, which are extended by facilities to deal not only with service behavior under regular circumstances (i.e., where cooperating services are working as expected) but also with their behavior in undesirable situations (i.e., where cooperating services are not working as expected). Furthermore, the approach can be used independently of artifacts (e.g., Business Process Execution Language) or type of composition (orchestration/choreography). A large case study, based on a commercial Web application, demonstrates the feasibility of the approach and analyzes its characteristics. Test generation and execution are supported by dedicated tools. Especially, the use of an enterprise service bus for test execution is noteworthy and differs from other approaches. The results of the case study encourage to suggest that the new approach has the power to detect faults systematically, performing properly even with complex and large compositions. Copyright © 2012 John Wiley & Sons, Ltd.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Modern enterprises work in highly dynamic environment. Thus, the developing of company strategy is of crucial importance. It determines the surviving of the enterprise and its evolution. Adapting the desired management goal in accordance with the environment changes is a complex problem. In the present paper, an approach for solving this problem is suggested. It is based on predictive control philosophy. The enterprise is modelled as a cybernetic system and the future plant response is predicted by a neural network model. The predictions are passed to an optimization routine, which attempts to minimize the quadratic performance criterion.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The complex design and development of a planar multilayer phased array antenna in microstrip technology can be simplified using two commercially available design tools 1) Ansoft Ensemble and 2) HP-EEsof Touchstone. In the approach presented here, Touchstone is used to design RF switches and phase shifters whose scattering parameters are incorporated in Ensemble simulations using its black box tool. Using this approach, Ensemble is able to fully analyze the performance of radiating and beamforming layers of a phased array prior to its manufacturing. This strategy is demonstrated in a design example of a 12-element linearly-polarized circular phased array operating at L band. A comparison between theoretical and experimental results of the array is demonstrated.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The anaerobic ammonium oxidation process is a new process for ammonia removal from wastewater. It is also a new microbial physiology that was previously believed to be impossible. The identification of Candidatus Brocadia anammoxidans and its relatives as the responsible bacteria was only possible with the development of a new experimental approach. That approach is the focus of this paper. The approach is a modernisation of the Winogradsky/Beyerinck strategy of selective enrichment and is based on the introduction of the molecular toolbox and modern bioreactor engineering to microbial ecology. It consists of five steps: (1) postulation of an ecological niche based on thermodynamic considerations and macro-ecological field data; (2) engineering of this niche into a laboratory bioreactor for enrichment culture; (3) black-box physiological characterisation of the enrichment culture as a whole; (4) phylogenetic characterisation of the enriched community using molecular tools; (5) physical separation of the dominant members of the enrichment culture using gradient centrifugation and the identification of the species of interest in accordance with Koch's postulates; (6) verification of the in situ importance of these species in the actual ecosystems. The power of this approach is illustrated with a case study: the identification of the planctomycetes responsible for anaerobic ammonium oxidation. We argue that this was impossible using molecular ecology or conventional 'cultivation based techniques' alone. We suggest that the approach might also be used for the microbiological study of many interesting microbes such as anaerobic methane oxidisers.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Due to usage conditions, hazardous environments or intentional causes, physical and virtual systems are subject to faults in their components, which may affect their overall behaviour. In a ‘black-box’ agent modelled by a set of propositional logic rules, in which just a subset of components is externally visible, such faults may only be recognised by examining some output function of the agent. A (fault-free) model of the agent’s system provides the expected output given some input. If the real output differs from that predicted output, then the system is faulty. However, some faults may only become apparent in the system output when appropriate inputs are given. A number of problems regarding both testing and diagnosis thus arise, such as testing a fault, testing the whole system, finding possible faults and differentiating them to locate the correct one. The corresponding optimisation problems of finding solutions that require minimum resources are also very relevant in industry, as is minimal diagnosis. In this dissertation we use a well established set of benchmark circuits to address such diagnostic related problems and propose and develop models with different logics that we formalise and generalise as much as possible. We also prove that all techniques generalise to agents and to multiple faults. The developed multi-valued logics extend the usual Boolean logic (suitable for faultfree models) by encoding values with some dependency (usually on faults). Such logics thus allow modelling an arbitrary number of diagnostic theories. Each problem is subsequently solved with CLP solvers that we implement and discuss, together with a new efficient search technique that we present. We compare our results with other approaches such as SAT (that require substantial duplication of circuits), showing the effectiveness of constraints over multi-valued logics, and also the adequacy of a general set constraint solver (with special inferences over set functions such as cardinality) on other problems. In addition, for an optimisation problem, we integrate local search with a constructive approach (branch-and-bound) using a variety of logics to improve an existing efficient tool based on SAT and ILP.