907 resultados para Unconstrained minimization


Relevância:

10.00% 10.00%

Publicador:

Resumo:

MSC 2010: 26A33, 70H25, 46F12, 34K37 Dedicated to 80-th birthday of Prof. Rudolf Gorenflo

Relevância:

10.00% 10.00%

Publicador:

Resumo:

2000 Mathematics Subject Classification: 90C25, 68W10, 49M37.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

AMS subject classification: 90C31, 90A09, 49K15, 49L20.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Distributed fibre sensors provide unique capabilities for monitoring large infrastructures with high resolution. Practically, all these sensors are based on some kind of backscattering interaction. A pulsed activating signal is launched on one side of the sensing fibre and the backscattered signal is read as a function of the time of flight of the pulse along the fibre. A key limitation in the measurement range of all these sensors is introduced by fibre attenuation. As the pulse travels along the fibre, the losses in the fibre cause a drop of signal contrast and consequently a growth in the measurement uncertainty. In typical single-mode fibres, attenuation imposes a range limit of less than 30km, for resolutions in the order of 1-2 meters. An interesting improvement in this performance can be considered by using distributed amplification along the fibre [1]. Distributed amplification allows having a more homogeneous signal power along the sensing fibre, which also enables reducing the signal power at the input and therefore avoiding nonlinearities. However, in long structures (≥ 50 km), plain distributed amplification does not perfectly compensate the losses and significant power variations along the fibre are to be expected, leading to inevitable limitations in the measurements. From this perspective, it is simple to understand intuitively that the best possible solution for distributed sensors would be offered by a virtually transparent fibre, i.e. a fibre exhibiting effectively zero attenuation in the spectral region of the pulse. In addition, it can be shown that lossless transmission is the working point that allows the minimization of the amplified spontaneous emission (ASE) noise build-up. © 2011 IEEE.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Many innovations are inspired by past ideas in a nontrivial way. Tracing these origins and identifying scientific branches is crucial for research inspirations. In this paper, we use citation relations to identify the descendant chart, i.e., the family tree of research papers. Unlike other spanning trees that focus on cost or distance minimization, we make use of the nature of citations and identify the most important parent for each publication, leading to a treelike backbone of the citation network. Measures are introduced to validate the backbone as the descendant chart. We show that citation backbones can well characterize the hierarchical and fractal structure of scientific development, and lead to an accurate classification of fields and subfields. © 2011 American Physical Society.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Purpose: Considering the UK's limited capacity for waste disposal (particularly for hazardous/radiological waste) there is growing focus on waste avoidance and minimisation to lower the volumes of waste being sent to disposal. The hazardous nature of some waste can complicate its management and reduction. To address this problem there was a need for a decision making methodology to support managers in the nuclear industry as they identify ways to reduce the production of avoidable hazardous waste. The methodology we developed is called Waste And Sourcematter Analysis (WASAN). A methodology that begins the thought process at the pre-waste creation stage (i.e. Avoid). Design/methodology/ approach: The methodology analyses the source of waste, the production of waste inside the facility, the knock on effects from up/downstream facilities on waste production, and the down-selection of waste minimisation actions/options. WASAN has been applied to case studies with licencees and this paper reports on one such case study - the management of plastic bags in Enriched Uranium Residues Recovery Plant (EURRP) at Springfields (UK) where it was used to analyse the generation of radioactive plastic bag waste. Findings: Plastic bags are used in EURRP as a strategy to contain hazard. Double bagging of materials led to the proliferation of these bags as a waste. The paper reports on the philosophy behind WASAN, the application of the methodology to this problem, the results, and views from managers in EURRP. Originality/value: This paper presents WASAN as a novel methodology for analyzing the minimization of avoidable hazardous waste. This addresses an issue that is important to many industries e.g. where legislation enforces waste minimization, where waste disposal costs encourage waste avoidance, or where plant design can reduce waste. The paper forms part of the HSE Nuclear Installations Inspectorate's desire to work towards greater openness and transparency in its work and the development in its thinking.© Crown Copyright 2011.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A dolgozatban a döntéselméletben fontos szerepet játszó páros összehasonlítás mátrix prioritásvektorának meghatározására új megközelítést alkalmazunk. Az A páros összehasonlítás mátrix és a prioritásvektor által definiált B konzisztens mátrix közötti eltérést a Kullback-Leibler relatív entrópia-függvény segítségével mérjük. Ezen eltérés minimalizálása teljesen kitöltött mátrix esetében konvex programozási feladathoz vezet, nem teljesen kitöltött mátrix esetében pedig egy fixpont problémához. Az eltérésfüggvényt minimalizáló prioritásvektor egyben azzal a tulajdonsággal is rendelkezik, hogy az A mátrix elemeinek összege és a B mátrix elemeinek összege közötti különbség éppen az eltérésfüggvény minimumának az n-szerese, ahol n a feladat mérete. Így az eltérésfüggvény minimumának értéke két szempontból is lehet alkalmas az A mátrix inkonzisztenciájának a mérésére. _____ In this paper we apply a new approach for determining a priority vector for the pairwise comparison matrix which plays an important role in Decision Theory. The divergence between the pairwise comparison matrix A and the consistent matrix B defined by the priority vector is measured with the help of the Kullback-Leibler relative entropy function. The minimization of this divergence leads to a convex program in case of a complete matrix, leads to a fixed-point problem in case of an incomplete matrix. The priority vector minimizing the divergence also has the property that the difference of the sums of elements of the matrix A and the matrix B is n times the minimum of the divergence function where n is the dimension of the problem. Thus we developed two reasons for considering the value of the minimum of the divergence as a measure of inconsistency of the matrix A.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The purpose of this research was to examine from a syntactic and narrative structure perspective two narrative summary types: a summary with a length constraint and an unconstrained summary. In addition, this research served to develop a multidimensional theory of narrative comprehension.^ College freshmen read two short stories written by written by Sake and were asked to write a constrained summary for one text and an unconstrained summary for the other text. Following this the subjects completed a metacognitive questionnaire. The summaries were analyzed to examine transitivity features and narrative structure features. The metacognitive questionnaires were examined to extract information about plot structure, differences between one and two episode stories, and to gain insight into the strategies used by subjects in producing both summary types.^ A Paired t-test conducted on the data found that there was a significant transitivity feature mean difference between a constrained summary and an unconstrained summary indicating that the number of transitivity features produced from each summary type were task dependent.^ Chi-square tests conducted on the data found that there were proportional differences in usage between plot features and thematic abstract units in an unconstrained summary and a constrained summary indicating that plot features and thematic abstract units produced from each summary type were task dependent.^ Qualitative analyses indicated that setting, goal, and resolution are typical within plot organization, there are summary production differences between one and two episode narratives, and subjects do not seem to be aware of summary production strategies.^ The results of this research have implications for comprehension and writing instruction. ^

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Due to the powerful nature of confession evidence, it is imperative that we investigate the factors that affect the likelihood of obtaining true and false confessions. Previous research has been conducted with a paradigm limited to the study of false confessions to an act of negligence, thereby limiting the generalizability of the findings. The first goal of the current study was to introduce a novel paradigm involving a more serious, intentional act that can be used in the study of both true and false confessions. The second goal was to explore the effects of two police interrogation tactics, minimization and an offer of leniency, on true and false confession rates. ^ Three hundred and thirty-four undergraduates at a large southeastern university were recruited to participate in a study on problem-solving and decision-making. During the course of the laboratory experiment, participants were induced to intentionally break or not break an experimental rule, an act that was characterized as “cheating.” All participants (i.e., both innocent and guilty) were later accused of the act and interrogated. For half of the participants, the interrogator used minimization tactics, which involved downplaying the seriousness of the offense, expressing sympathy, and providing face-saving excuses, in order to encourage the participant to confess. An offer of leniency was also manipulated in which half the participants were offered a “deal” that involved the option of confessing and accepting a known punishment or not confessing and facing the threat of harsher punishment. Results indicated that guilty persons were more likely to confess than innocent persons, and that the use of minimization and an explicit offer of leniency increased both the true and false confession rates. Furthermore, a cumulative effect of techniques was observed, such that the diagnosticity of the interrogation (the ratio of true confessions to false confessions) decreased as the number of techniques used increased. Taken together, the results suggest that caution should be used when implementing these techniques in the interrogation room. ^

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Since the 1950s, the theory of deterministic and nondeterministic finite automata (DFAs and NFAs, respectively) has been a cornerstone of theoretical computer science. In this dissertation, our main object of study is minimal NFAs. In contrast with minimal DFAs, minimal NFAs are computationally challenging: first, there can be more than one minimal NFA recognizing a given language; second, the problem of converting an NFA to a minimal equivalent NFA is NP-hard, even for NFAs over a unary alphabet. Our study is based on the development of two main theories, inductive bases and partials, which in combination form the foundation for an incremental algorithm, ibas, to find minimal NFAs. An inductive basis is a collection of languages with the property that it can generate (through union) each of the left quotients of its elements. We prove a fundamental characterization theorem which says that a language can be recognized by an n-state NFA if and only if it can be generated by an n-element inductive basis. A partial is an incompletely-specified language. We say that an NFA recognizes a partial if its language extends the partial, meaning that the NFA’s behavior is unconstrained on unspecified strings; it follows that a minimal NFA for a partial is also minimal for its language. We therefore direct our attention to minimal NFAs recognizing a given partial. Combining inductive bases and partials, we generalize our characterization theorem, showing that a partial can be recognized by an n-state NFA if and only if it can be generated by an n-element partial inductive basis. We apply our theory to develop and implement ibas, an incremental algorithm that finds minimal partial inductive bases generating a given partial. In the case of unary languages, ibas can often find minimal NFAs of up to 10 states in about an hour of computing time; with brute-force search this would require many trillions of years.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

To achieve the goal of sustainable development, the building energy system was evaluated from both the first and second law of thermodynamics point of view. The relationship between exergy destruction and sustainable development were discussed at first, followed by the description of the resource abundance model, the life cycle analysis model and the economic investment effectiveness model. By combining the forgoing models, a new sustainable index was proposed. Several green building case studies in U.S. and China were presented. The influences of building function, geographic location, climate pattern, the regional energy structure, and the technology improvement potential of renewable energy in the future were discussed. The building’s envelope, HVAC system, on-site renewable energy system life cycle analysis from energy, exergy, environmental and economic perspective were compared. It was found that climate pattern had a dramatic influence on the life cycle investment effectiveness of the building envelope. The building HVAC system energy performance was much better than its exergy performance. To further increase the exergy efficiency, renewable energy rather than fossil fuel should be used as the primary energy. A building life cycle cost and exergy consumption regression model was set up. The optimal building insulation level could be affected by either cost minimization or exergy consumption minimization approach. The exergy approach would cause better insulation than cost approach. The influence of energy price on the system selection strategy was discussed. Two photovoltaics (PV) systems—stand alone and grid tied system were compared by the life cycle assessment method. The superiority of the latter one was quite obvious. The analysis also showed that during its life span PV technology was less attractive economically because the electricity price in U.S. and China did not fully reflect the environmental burden associated with it. However if future energy price surges and PV system cost reductions were considered, the technology could be very promising for sustainable buildings in the future.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Over the past few decades, we have been enjoying tremendous benefits thanks to the revolutionary advancement of computing systems, driven mainly by the remarkable semiconductor technology scaling and the increasingly complicated processor architecture. However, the exponentially increased transistor density has directly led to exponentially increased power consumption and dramatically elevated system temperature, which not only adversely impacts the system's cost, performance and reliability, but also increases the leakage and thus the overall power consumption. Today, the power and thermal issues have posed enormous challenges and threaten to slow down the continuous evolvement of computer technology. Effective power/thermal-aware design techniques are urgently demanded, at all design abstraction levels, from the circuit-level, the logic-level, to the architectural-level and the system-level. ^ In this dissertation, we present our research efforts to employ real-time scheduling techniques to solve the resource-constrained power/thermal-aware, design-optimization problems. In our research, we developed a set of simple yet accurate system-level models to capture the processor's thermal dynamic as well as the interdependency of leakage power consumption, temperature, and supply voltage. Based on these models, we investigated the fundamental principles in power/thermal-aware scheduling, and developed real-time scheduling techniques targeting at a variety of design objectives, including peak temperature minimization, overall energy reduction, and performance maximization. ^ The novelty of this work is that we integrate the cutting-edge research on power and thermal at the circuit and architectural-level into a set of accurate yet simplified system-level models, and are able to conduct system-level analysis and design based on these models. The theoretical study in this work serves as a solid foundation for the guidance of the power/thermal-aware scheduling algorithms development in practical computing systems.^

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Current technology permits connecting local networks via high-bandwidth telephone lines. Central coordinator nodes may use Intelligent Networks to manage data flow over dialed data lines, e.g. ISDN, and to establish connections between LANs. This dissertation focuses on cost minimization and on establishing operational policies for query distribution over heterogeneous, geographically distributed databases. Based on our study of query distribution strategies, public network tariff policies, and database interface standards we propose methods for communication cost estimation, strategies for the reduction of bandwidth allocation, and guidelines for central to node communication protocols. Our conclusion is that dialed data lines offer a cost effective alternative for the implementation of distributed database query systems, and that existing commercial software may be adapted to support query processing in heterogeneous distributed database systems. ^

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Since the 1950s, the theory of deterministic and nondeterministic finite automata (DFAs and NFAs, respectively) has been a cornerstone of theoretical computer science. In this dissertation, our main object of study is minimal NFAs. In contrast with minimal DFAs, minimal NFAs are computationally challenging: first, there can be more than one minimal NFA recognizing a given language; second, the problem of converting an NFA to a minimal equivalent NFA is NP-hard, even for NFAs over a unary alphabet. Our study is based on the development of two main theories, inductive bases and partials, which in combination form the foundation for an incremental algorithm, ibas, to find minimal NFAs. An inductive basis is a collection of languages with the property that it can generate (through union) each of the left quotients of its elements. We prove a fundamental characterization theorem which says that a language can be recognized by an n-state NFA if and only if it can be generated by an n-element inductive basis. A partial is an incompletely-specified language. We say that an NFA recognizes a partial if its language extends the partial, meaning that the NFA's behavior is unconstrained on unspecified strings; it follows that a minimal NFA for a partial is also minimal for its language. We therefore direct our attention to minimal NFAs recognizing a given partial. Combining inductive bases and partials, we generalize our characterization theorem, showing that a partial can be recognized by an n-state NFA if and only if it can be generated by an n-element partial inductive basis. We apply our theory to develop and implement ibas, an incremental algorithm that finds minimal partial inductive bases generating a given partial. In the case of unary languages, ibas can often find minimal NFAs of up to 10 states in about an hour of computing time; with brute-force search this would require many trillions of years.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

To achieve the goal of sustainable development, the building energy system was evaluated from both the first and second law of thermodynamics point of view. The relationship between exergy destruction and sustainable development were discussed at first, followed by the description of the resource abundance model, the life cycle analysis model and the economic investment effectiveness model. By combining the forgoing models, a new sustainable index was proposed. Several green building case studies in U.S. and China were presented. The influences of building function, geographic location, climate pattern, the regional energy structure, and the technology improvement potential of renewable energy in the future were discussed. The building’s envelope, HVAC system, on-site renewable energy system life cycle analysis from energy, exergy, environmental and economic perspective were compared. It was found that climate pattern had a dramatic influence on the life cycle investment effectiveness of the building envelope. The building HVAC system energy performance was much better than its exergy performance. To further increase the exergy efficiency, renewable energy rather than fossil fuel should be used as the primary energy. A building life cycle cost and exergy consumption regression model was set up. The optimal building insulation level could be affected by either cost minimization or exergy consumption minimization approach. The exergy approach would cause better insulation than cost approach. The influence of energy price on the system selection strategy was discussed. Two photovoltaics (PV) systems – stand alone and grid tied system were compared by the life cycle assessment method. The superiority of the latter one was quite obvious. The analysis also showed that during its life span PV technology was less attractive economically because the electricity price in U.S. and China did not fully reflect the environmental burden associated with it. However if future energy price surges and PV system cost reductions were considered, the technology could be very promising for sustainable buildings in the future.