856 resultados para Unified Formulation
Resumo:
The diversity in the way cloud providers o↵er their services, give their SLAs, present their QoS, or support di↵erent technologies, makes very difficult the portability and interoperability of cloud applications, and favours the well-known vendor lock-in problem. We propose a model to describe cloud applications and the required resources in an agnostic, and providers- and resources-independent way, in which individual application modules, and entire applications, may be re-deployed using different services without modification. To support this model, and after the proposal of a variety of cross-cloud application management tools by different authors, we propose going one step further in the unification of cloud services with a management approach in which IaaS and PaaS services are integrated into a unified interface. We provide support for deploying applications whose components are distributed on different cloud providers, indistinctly using IaaS and PaaS services.
Resumo:
Process systems design, operation and synthesis problems under uncertainty can readily be formulated as two-stage stochastic mixed-integer linear and nonlinear (nonconvex) programming (MILP and MINLP) problems. These problems, with a scenario based formulation, lead to large-scale MILPs/MINLPs that are well structured. The first part of the thesis proposes a new finitely convergent cross decomposition method (CD), where Benders decomposition (BD) and Dantzig-Wolfe decomposition (DWD) are combined in a unified framework to improve the solution of scenario based two-stage stochastic MILPs. This method alternates between DWD iterations and BD iterations, where DWD restricted master problems and BD primal problems yield a sequence of upper bounds, and BD relaxed master problems yield a sequence of lower bounds. A variant of CD, which includes multiple columns per iteration of DW restricted master problem and multiple cuts per iteration of BD relaxed master problem, called multicolumn-multicut CD is then developed to improve solution time. Finally, an extended cross decomposition method (ECD) for solving two-stage stochastic programs with risk constraints is proposed. In this approach, a CD approach at the first level and DWD at a second level is used to solve the original problem to optimality. ECD has a computational advantage over a bilevel decomposition strategy or solving the monolith problem using an MILP solver. The second part of the thesis develops a joint decomposition approach combining Lagrangian decomposition (LD) and generalized Benders decomposition (GBD), to efficiently solve stochastic mixed-integer nonlinear nonconvex programming problems to global optimality, without the need for explicit branch and bound search. In this approach, LD subproblems and GBD subproblems are systematically solved in a single framework. The relaxed master problem obtained from the reformulation of the original problem, is solved only when necessary. A convexification of the relaxed master problem and a domain reduction procedure are integrated into the decomposition framework to improve solution efficiency. Using case studies taken from renewable resource and fossil-fuel based application in process systems engineering, it can be seen that these novel decomposition approaches have significant benefit over classical decomposition methods and state-of-the-art MILP/MINLP global optimization solvers.
Développement des bétons autoplaçants à faible teneur en poudre, Éco-BAP: formulation et performance
Resumo:
Abstract : Although concrete is a relatively green material, the astronomical volume of concrete produced worldwide annually places the concrete construction sector among the noticeable contributors to the global warming. The most polluting constituent of concrete is cement due to its production process which releases, on average, 0.83 kg CO[subscript 2] per kg of cement. Self-consolidating concrete (SCC), a type of concrete that can fill in the formwork without external vibration, is a technology that can offer a solution to the sustainability issues of concrete industry. However, all of the workability requirements of SCC originate from a higher powder content (compared to conventional concrete) which can increase both the cost of construction and the environmental impact of SCC for some applications. Ecological SCC, Eco-SCC, is a recent development combing the advantages of SCC and a significantly lower powder content. The maximum powder content of this concrete, intended for building and commercial construction, is limited to 315 kg/m[superscript 3]. Nevertheless, designing Eco-SCC can be challenging since a delicate balance between different ingredients of this concrete is required to secure a satisfactory mixture. In this Ph.D. program, the principal objective is to develop a systematic design method to produce Eco-SCC. Since the particle lattice effect (PLE) is a key parameter to design stable Eco-SCC mixtures and is not well understood, in the first phase of this research, this phenomenon is studied. The focus in this phase is on the effect of particle-size distribution (PSD) on the PLE and stability of model mixtures as well as SCC. In the second phase, the design protocol is developed, and the properties of obtained Eco-SCC mixtures in both fresh and hardened states are evaluated. Since the assessment of robustness is crucial for successful production of concrete on large-scale, in the final phase of this work, the robustness of one the best-performing mixtures of Phase II is examined. It was found that increasing the volume fraction of a stable size-class results in an increase in the stability of that class, which in turn contributes to a higher PLE of the granular skeleton and better stability of the system. It was shown that a continuous PSD in which the volume fraction of each size class is larger than the consecutive coarser class can increase the PLE. Using such PSD was shown to allow for a substantial increase in the fluidity of SCC mixture without compromising the segregation resistance. An index to predict the segregation potential of a suspension of particles in a yield stress fluid was proposed. In the second phase of the dissertation, a five-step design method for Eco-SCC was established. The design protocol started with the determination of powder and water contents followed by the optimization of sand and coarse aggregate volume fractions according to an ideal PSD model (Funk and Dinger). The powder composition was optimized in the third step to minimize the water demand while securing adequate performance in the hardened state. The superplasticizer (SP) content of the mixtures was determined in next step. The last step dealt with the assessment of the global warming potential of the formulated Eco-SCC mixtures. The optimized Eco-SCC mixtures met all the requirements of self-consolidation in the fresh state. The 28-day compressive strength of such mixtures complied with the target range of 25 to 35 MPa. In addition, the mixtures showed sufficient performance in terms of drying shrinkage, electrical resistivity, and frost durability for the intended applications. The eco-performance of the developed mixtures was satisfactory as well. It was demonstrated in the last phase that the robustness of Eco-SCC is generally good with regards to water content variations and coarse aggregate characteristics alterations. Special attention must be paid to the dosage of SP during batching.
Resumo:
The idea of spacecraft formations, flying in tight configurations with maximum baselines of a few hundred meters in low-Earth orbits, has generated widespread interest over the last several years. Nevertheless, controlling the movement of spacecraft in formation poses difficulties, such as in-orbit high-computing demand and collision avoidance capabilities, which escalate as the number of units in the formation is increased and complicated nonlinear effects are imposed to the dynamics, together with uncertainty which may arise from the lack of knowledge of system parameters. These requirements have led to the need of reliable linear and nonlinear controllers in terms of relative and absolute dynamics. The objective of this thesis is, therefore, to introduce new control methods to allow spacecraft in formation, with circular/elliptical reference orbits, to efficiently execute safe autonomous manoeuvres. These controllers distinguish from the bulk of literature in that they merge guidance laws never applied before to spacecraft formation flying and collision avoidance capacities into a single control strategy. For this purpose, three control schemes are presented: linear optimal regulation, linear optimal estimation and adaptive nonlinear control. In general terms, the proposed control approaches command the dynamical performance of one or several followers with respect to a leader to asymptotically track a time-varying nominal trajectory (TVNT), while the threat of collision between the followers is reduced by repelling accelerations obtained from the collision avoidance scheme during the periods of closest proximity. Linear optimal regulation is achieved through a Riccati-based tracking controller. Within this control strategy, the controller provides guidance and tracking toward a desired TVNT, optimizing fuel consumption by Riccati procedure using a non-infinite cost function defined in terms of the desired TVNT, while repelling accelerations generated from the CAS will ensure evasive actions between the elements of the formation. The relative dynamics model, suitable for circular and eccentric low-Earth reference orbits, is based on the Tschauner and Hempel equations, and includes a control input and a nonlinear term corresponding to the CAS repelling accelerations. Linear optimal estimation is built on the forward-in-time separation principle. This controller encompasses two stages: regulation and estimation. The first stage requires the design of a full state feedback controller using the state vector reconstructed by means of the estimator. The second stage requires the design of an additional dynamical system, the estimator, to obtain the states which cannot be measured in order to approximately reconstruct the full state vector. Then, the separation principle states that an observer built for a known input can also be used to estimate the state of the system and to generate the control input. This allows the design of the observer and the feedback independently, by exploiting the advantages of linear quadratic regulator theory, in order to estimate the states of a dynamical system with model and sensor uncertainty. The relative dynamics is described with the linear system used in the previous controller, with a control input and nonlinearities entering via the repelling accelerations from the CAS during collision avoidance events. Moreover, sensor uncertainty is added to the control process by considering carrier-phase differential GPS (CDGPS) velocity measurement error. An adaptive control law capable of delivering superior closed-loop performance when compared to the certainty-equivalence (CE) adaptive controllers is finally presented. A novel noncertainty-equivalence controller based on the Immersion and Invariance paradigm for close-manoeuvring spacecraft formation flying in both circular and elliptical low-Earth reference orbits is introduced. The proposed control scheme achieves stabilization by immersing the plant dynamics into a target dynamical system (or manifold) that captures the desired dynamical behaviour. They key feature of this methodology is the addition of a new term to the classical certainty-equivalence control approach that, in conjunction with the parameter update law, is designed to achieve adaptive stabilization. This parameter has the ultimate task of shaping the manifold into which the adaptive system is immersed. The performance of the controller is proven stable via a Lyapunov-based analysis and Barbalat’s lemma. In order to evaluate the design of the controllers, test cases based on the physical and orbital features of the Prototype Research Instruments and Space Mission Technology Advancement (PRISMA) are implemented, extending the number of elements in the formation into scenarios with reconfigurations and on-orbit position switching in elliptical low-Earth reference orbits. An extensive analysis and comparison of the performance of the controllers in terms of total Δv and fuel consumption, with and without the effects of the CAS, is presented. These results show that the three proposed controllers allow the followers to asymptotically track the desired nominal trajectory and, additionally, those simulations including CAS show an effective decrease of collision risk during the performance of the manoeuvre.
Resumo:
With a new finite strain anisotropic framework, we introduce a unified approach for constitutive model- ing and delamination of composites. We describe a finite-strain semi-implicit integration algorithm and the application to assumed-strain hexahedra. In a laminate composite, the laminae are modeled by an anisotropic Kirchhoff/Saint-Venant material and the interfaces are modeled by the exponential cohesive law with intrinsic characteristic length and the criterion by Benzeggagh and Kenane for the equivalent fracture toughness. For the element formulation, a weighted least-squares algorithm is used to calculate the mixed strain. Löwdin frames are used to model orthotropic materials without the added task of per- forming a polar decomposition or empirical frames. To assess the validity of our proposals and inspect step and mesh size dependence, a least-squares based hexahedral element is implemented and tested in depth in both deformation and delamination examples.
Resumo:
In this study, a three-dimensional (3D) non-ordinary state-based peridynamics (NOSB-PD) formulation for thermomechanical brittle and ductile fracture is presented. The Johnson–Cook (JC) constitutive and damage model is used to taken into account plastic hardening, thermal softening and fracture. The for- mulation is validated by considering two benchmark examples: 1) The Taylor-bar impact and 2) the Kalthoff– Winkler tests. The results show good agreements between the numerical simulations and the experimental results.
Resumo:
2008
Resumo:
This paper considers the question, ‘what is co-creative media, and why is it a useful idea in social media research’? The term ‘co-creative media’ is now used by Creative Industries researchers at QUT to describe their digital storytelling practices. Digital storytelling is a set of collaborative digital media production techniques that have been used to facilitate social participation in numerous Australian and international contexts. Digital storytelling has been adapted by Creative Industries researchers at QUT as a platform for researching the potential of vernacular creativity in a variety of contexts, including social inclusion of marginalized and disadvantaged groups; inclusion in public histories of narratives that might be overlooked; and articulation of voices that otherwise remain silent in the formulation of social and economic development strategies. The adaption of digital storytelling to different contexts has been shaped by the reflexive, recursive, and pragmatic requirements of action research. Amongst other things, this activity draws attention to the agency of researchers in facilitating these kinds of participatory media processes and outcomes. This discussion serves to problematise concepts of participatory media by introducing the term ‘co-creative media’ and differentiating these from other social media production practices.
Resumo:
For certain continuum problems, it is desirable and beneficial to combine two different methods together in order to exploit their advantages while evading their disadvantages. In this paper, a bridging transition algorithm is developed for the combination of the meshfree method (MM) with the finite element method (FEM). In this coupled method, the meshfree method is used in the sub-domain where the MM is required to obtain high accuracy, and the finite element method is employed in other sub-domains where FEM is required to improve the computational efficiency. The MM domain and the FEM domain are connected by a transition (bridging) region. A modified variational formulation and the Lagrange multiplier method are used to ensure the compatibility of displacements and their gradients. To improve the computational efficiency and reduce the meshing cost in the transition region, regularly distributed transition particles, which are independent of either the meshfree nodes or the FE nodes, can be inserted into the transition region. The newly developed coupled method is applied to the stress analysis of 2D solids and structures in order to investigate its’ performance and study parameters. Numerical results show that the present coupled method is convergent, accurate and stable. The coupled method has a promising potential for practical applications, because it can take advantages of both the meshfree method and FEM when overcome their shortcomings.
Resumo:
We present a method for topological SLAM that specifically targets loop closing for edge-ordered graphs. Instead of using a heuristic approach to accept or reject loop closing, we propose a probabilistically grounded multi-hypothesis technique that relies on the incremental construction of a map/state hypothesis tree. Loop closing is introduced automatically within the tree expansion, and likely hypotheses are chosen based on their posterior probability after a sequence of sensor measurements. Careful pruning of the hypothesis tree keeps the growing number of hypotheses under control and a recursive formulation reduces storage and computational costs. Experiments are used to validate the approach.
Resumo:
The service-orientation paradigm has not only become prevalent in the software systems domain in recent years, but is also increasingly applied on the business level to restructure organisational capabilities. In this paper, we present the results of an extensive literature review of 30 approaches related to service identification and analysis for both domains. Based on the consolidation of a superset of comparison criteria for service-oriented methodologies found in related literature, we compare and evaluate the different characteristics of service engineering methods with a focus on service analysis. Although a close business and IT alignment is regarded as one of the core beneficial promises of service-orientation, our analysis suggests that there is a lack of unified, comprehensive methodology for service identification and analysis integrating and addressing both domains. Thus, we discuss how our results can inform directions for future research in this area.
Resumo:
Objectives The objectives of this project were two-fold: • Assess the ease with which current architectural CAD systems supported the use ofparametric descriptions in defining building shape, engineering system performance and cost at the early stages of building design; • Assess the feasibility of implementing a software decision support system that allowed designers to trade-off the characteristics and configuration of various engineering systems to move towards a “global optimum” rather than considering each system in isolation and expecting humans to weigh up all of the costs and benefits. The first stage of the project consisted of using four different CAD systems to define building shells (envelopes) with different usages. These models were then exported into a shared database using the IFC information exchange specifications. The second stage involved the implementation of small computer programs that were able to estimate relevant system parameters based on performance requirements and the constraints imposed by the other systems. These are presented in a unified user interface that extracts the appropriate building shape parameters from the shared database Note that the term parametric in this context refers to the relationships among and between all elements of the building model - not just geometric associations - which will enable the desired coordination.
Resumo:
Reinforced concrete structures are susceptible to a variety of deterioration mechanisms due to creep and shrinkage, alkali-silica reaction (ASR), carbonation, and corrosion of the reinforcement. The deterioration problems can affect the integrity and load carrying capacity of the structure. Substantial research has been dedicated to these various mechanisms aiming to identify the causes, reactions, accelerants, retardants and consequences. This has improved our understanding of the long-term behaviour of reinforced concrete structures. However, the strengthening of reinforced concrete structures for durability has to date been mainly undertaken after expert assessment of field data followed by the development of a scheme to both terminate continuing degradation, by separating the structure from the environment, and strengthening the structure. The process does not include any significant consideration of the residual load-bearing capacity of the structure and the highly variable nature of estimates of such remaining capacity. Development of performance curves for deteriorating bridge structures has not been attempted due to the difficulty in developing a model when the input parameters have an extremely large variability. This paper presents a framework developed for an asset management system which assesses residual capacity and identifies the most appropriate rehabilitation method for a given reinforced concrete structure exposed to aggressive environments. In developing the framework, several industry consultation sessions have been conducted to identify input data required, research methodology and output knowledge base. Capturing expert opinion in a useable knowledge base requires development of a rule based formulation, which can subsequently be used to model the reliability of the performance curve of a reinforced concrete structure exposed to a given environment.
Resumo:
This paper introduces fast algorithms for performing group operations on twisted Edwards curves, pushing the recent speed limits of Elliptic Curve Cryptography (ECC) forward in a wide range of applications. Notably, the new addition algorithm uses for suitably selected curve constants. In comparison, the fastest point addition algorithms for (twisted) Edwards curves stated in the literature use . It is also shown that the new addition algorithm can be implemented with four processors dropping the effective cost to . This implies an effective speed increase by the full factor of 4 over the sequential case. Our results allow faster implementation of elliptic curve scalar multiplication. In addition, the new point addition algorithm can be used to provide a natural protection from side channel attacks based on simple power analysis (SPA).