800 resultados para Flexible housing
Resumo:
Electronic energy transfer (EET) rate constants between a naphthalene donor and anthracene acceptor in [ZnL4a](ClO4)(2) and [ZnL4b](ClO4)(2) were determined by time-resolved fluorescence where L-4a and L-4b are the trans and cis isomers of 6-((anthracen-9-yl-methyl)amino)-6,13-dimethyl-13-((naphthalen-1-yl-methyl)amino)-1,4,8,11-tetraazacyclotetradecane, respectively. These isomers differ in the relative disposition of the appended chromophores with respect to the macrocyclic plane. The trans isomer has an energy transfer rate constant (k(EET)) of 8.7 x 10(8) s(-1), whereas that of the cis isomer is significantly faster (2.3 x 10(9) s(-1)). Molecular modeling was used to determine the likely distribution of conformations in CH3CN solution for these complexes in an attempt to identify any distance or orientation dependency that may account for the differing rate constants observed. The calculated conformational distributions together with analysis by H-1 NMR for the [ZnL4a](2+) trans complex in the common trans-III N-based isomer gave a calculated Forster rate constant close to that observed experimentally. For the [ZnL4b](2+) cis complex, the experimentally determined rate constant may be attributed to a combination of trans-Ill and trans-I N-based isomeric forms of the complex in solution.
Resumo:
Management systems standards (MSSs) have developed in an unprecedented manner in the last few years. These MSS cover a wide array of different disciplines, aims and activities of organisations. Also, organisations are populated with an enormous diversity of independent management systems (MSs). An integrated management system (IMS) tends to integrate some or all components of the business. Maximising their integration in one coherent and efficient MS is increasingly a strategic priority and constitutes an opportunity for businesses to be more competitive and consequently, promote its sustainable success. Those organisations that are quicker and more efficient in their integration and continuous improvement will have a competitive advantage in obtaining sustainable value in our global and competitive business world. Several scholars have proposed various theoretical approaches regarding the integration of management sub-systems, leading to the conclusion that there is no common practice for all organisations as they encompass different characteristics. One other author shows that several tangible and intangible gains for organisations, as well as to their internal and external stakeholders, are achieved with the integration of the individual standardised MSs. The purpose of this work was to conceive a model, Flexible, Integrator and Lean for IMSs, according to ISO 9001 for quality; ISO 14001 for environment and OHSAS 18001 for occupational health and safety (IMS–QES), that can be adapted and progressively assimilate other MSs, such as, SA 8000/ISO 26000 for social accountability, ISO 31000 for risk management and ISO/IEC 27001 for information security management, among others. The IMS–QES model was designed in the real environment of an industrial Portuguese small and medium enterprise, that over the years has been adopting, gradually, in whole or in part, individual MSSs. The developed model is based on a preliminary investigation conducted through a questionnaire. The strategy and research methods have taken into consideration the case study. Among the main findings of the survey we highlight: the creation of added value for the business through the elimination of several organisational wastes; the integrated management of the sustainability components; the elimination of conflicts between independent MS; dialogue with the main stakeholders and commitment to their ongoing satisfaction and increased contribution to the company’s competitiveness; and greater valorisation and motivation of employees as a result of the expansion of their skill base, actions and responsibilities, with their consequent empowerment. A set of key performance indicators (KPIs) constitute the support, in a perspective of business excellence, to the follow up of the organisation’s progress towards the vision and achievement of the defined objectives in the context of each component of the IMS model. The conceived model had many phases and the one presented in this work is the last required for the integration of quality, environment, safety and others individual standardised MSs. Globally, the investigation results, by themselves, justified and prioritised the conception of an IMS–QES model, to be implemented at the company where the investigation was conducted, but also a generic model of an IMS, which may be more flexible, integrator and lean as possible, potentiating the efficiency, added value both in the present and, fundamentally, for future.
Resumo:
Urban regeneration is more and more a “universal issue” and a crucial factor in the new trends of urban planning. It is no longer only an area of study and research; it became part of new urban and housing policies. Urban regeneration involves complex decisions as a consequence of the multiple dimensions of the problems that include special technical requirements, safety concerns, socio-economic, environmental, aesthetic, and political impacts, among others. This multi-dimensional nature of urban regeneration projects and their large capital investments justify the development and use of state-of-the-art decision support methodologies to assist decision makers. This research focuses on the development of a multi-attribute approach for the evaluation of building conservation status in urban regeneration projects, thus supporting decision makers in their analysis of the problem and in the definition of strategies and priorities of intervention. The methods presented can be embedded into a Geographical Information System for visualization of results. A real-world case study was used to test the methodology, whose results are also presented.
Resumo:
One of the most efficient approaches to generate the side information (SI) in distributed video codecs is through motion compensated frame interpolation where the current frame is estimated based on past and future reference frames. However, this approach leads to significant spatial and temporal variations in the correlation noise between the source at the encoder and the SI at the decoder. In such scenario, it would be useful to design an architecture where the SI can be more robustly generated at the block level, avoiding the creation of SI frame regions with lower correlation, largely responsible for some coding efficiency losses. In this paper, a flexible framework to generate SI at the block level in two modes is presented: while the first mode corresponds to a motion compensated interpolation (MCI) technique, the second mode corresponds to a motion compensated quality enhancement (MCQE) technique where a low quality Intra block sent by the encoder is used to generate the SI by doing motion estimation with the help of the reference frames. The novel MCQE mode can be overall advantageous from the rate-distortion point of view, even if some rate has to be invested in the low quality Intra coding blocks, for blocks where the MCI produces SI with lower correlation. The overall solution is evaluated in terms of RD performance with improvements up to 2 dB, especially for high motion video sequences and long Group of Pictures (GOP) sizes.
Resumo:
A new high throughput and scalable architecture for unified transform coding in H.264/AVC is proposed in this paper. Such flexible structure is capable of computing all the 4x4 and 2x2 transforms for Ultra High Definition Video (UHDV) applications (4320x7680@ 30fps) in real-time and with low hardware cost. These significantly high performance levels were proven with the implementation of several different configurations of the proposed structure using both FPGA and ASIC 90 nm technologies. In addition, such experimental evaluation also demonstrated the high area efficiency of theproposed architecture, which in terms of Data Throughput per Unit of Area (DTUA) is at least 1.5 times more efficient than its more prominent related designs(1).
Resumo:
OBJECTIVE: To examine whether any impairments in health and social lives can be found under different kinds of flexible working hours, and whether such effects are related to specific characteristics of these working hours. METHODS: Two studies - a company based survey (N=660) and an internet survey (N=528) - have been conducted. The first one was a questionnaire study (paper and pencil) on employees working under some 'typical' kinds of different flexible working time arrangements in different companies and different occupational fields (health care, manufacturing, retail, administration, call centres). The second study was an internet-based survey, using an adaptation of the questionnaire from the first study. RESULTS: The results of both studies consistently show that high variability of working hours is associated with increased impairments in health and well-being and this is especially true if this variability is company controlled. These effects are less pronounced if variability is self-controlled; however, autonomy does not compensate the effects of variability. CONCLUSIONS: Recommendations for an appropriate design of flexible working hours should be developed in order to minimize any impairing effects on health and psychosocial well-being; these recommendations should include - besides allowing for discretion in controlling one's (flexible) working hours - that variability in flexible working hours should be kept low (or at least moderate), even if this variability is self-controlled.
Resumo:
This paper presents a distributed model predictive control (DMPC) for indoor thermal comfort that simultaneously optimizes the consumption of a limited shared energy resource. The control objective of each subsystem is to minimize the heating/cooling energy cost while maintaining the indoor temperature and used power inside bounds. In a distributed coordinated environment, the control uses multiple dynamically decoupled agents (one for each subsystem/house) aiming to achieve satisfaction of coupling constraints. According to the hourly power demand profile, each house assigns a priority level that indicates how much is willing to bid in auction for consume the limited clean resource. This procedure allows the bidding value vary hourly and consequently, the agents order to access to the clean energy also varies. Despite of power constraints, all houses have also thermal comfort constraints that must be fulfilled. The system is simulated with several houses in a distributed environment.
Resumo:
OBJECTIVE: To assess the relationship between housing conditions and low birthweight and preterm low birthweight among low-income women. METHODS: A case-control study was conducted with post-partum women living in the city of Rio de Janeiro, Southeast Brazil, in 2003-2005. Two groups of cases, low birthweight (n=96) and preterm low birthweight infants (n=68), were compared against normal weight term controls (n=393). Housing conditions were categorized into three levels: adequate, inadequate, and highly inadequate. Covariates included sociodemographic and anthropometric characteristics, risk behaviors, violence, anxiety, satisfaction during pregnancy, obstetric history and prenatal care. RESULTS: Poor housing conditions was independently associated with low birthweight (inadequate - OR 2.3 [1.1;4.6]; highly inadequate - OR 7.6 [2.1;27.6]) and preterm low birthweight (inadequate - OR 2.2 [1.1;4.3]; highly inadequate - OR 7.6 [2.4;23.9]) and factors associated with outcomes were inadequate prenatal care and previous preterm birth. Low income and low maternal body mass index remained associated with low birthweight. CONCLUSIONS: Poor housing conditions were associated with low birthweight and preterm low birthweight.
Resumo:
Replication is a proven concept for increasing the availability of distributed systems. However, actively replicating every software component in distributed embedded systems may not be a feasible approach. Not only the available resources are often limited, but also the imposed overhead could significantly degrade the system's performance. The paper proposes heuristics to dynamically determine which components to replicate based on their significance to the system as a whole, its consequent number of passive replicas, and where to place those replicas in the network. The results show that the proposed heuristics achieve a reasonably higher system's availability than static offline decisions when lower replication ratios are imposed due to resource or cost limitations. The paper introduces a novel approach to coordinate the activation of passive replicas in interdependent distributed environments. The proposed distributed coordination model reduces the complexity of the needed interactions among nodes and is faster to converge to a globally acceptable solution than a traditional centralised approach.
Resumo:
Replication is a proven concept for increasing the availability of distributed systems. However, actively replicating every software component in distributed embedded systems may not be a feasible approach. Not only the available resources are often limited, but also the imposed overhead could significantly degrade the system’s performance. This paper proposes heuristics to dynamically determine which components to replicate based on their significance to the system as a whole, its consequent number of passive replicas, and where to place those replicas in the network. The activation of passive replicas is coordinated through a fast convergence protocol that reduces the complexity of the needed interactions among nodes until a new collective global service solution is determined.
Resumo:
Within the European project R-Fieldbus (http://www.hurray.isep.ipp.pt/activities/rfieldbus/), an industrial manufacturing field trial was developed. This field trial was conceived as a demonstration test bed for the technologies developed during the project. Because the R-Fieldbus field trial included prototype hardware devices, the purpose of this equipment changed and since the conclusion of the project, several new technologies also emerged, therefore an update of the field trial was required. This document describes an update of the manufacturing field trial. The purpose of this update, the changes and improvements introduced are described in the document. Additionally, this document also provides a reliable source of documentation for the equipment, configuration and software components of the manufacturing field trial.
Resumo:
Solar cells on lightweight and flexible substrates have advantages over glass-or wafer-based photovoltaic devices in both terrestrial and space applications. Here, we report on development of amorphous silicon thin film photovoltaic modules fabricated at maximum deposition temperature of 150 degrees C on 100 mu m thick polyethylene-naphtalate plastic films. Each module of 10 cm x 10 cm area consists of 72 a-Si:H n-i-p rectangular structures with transparent conducting oxide top electrodes with Al fingers and metal back electrodes deposited through the shadow masks. Individual structures are connected in series forming eight rows with connection ports provided for external blocking diodes. The design optimization and device performance analysis are performed using a developed SPICE model.
Resumo:
Cloud SLAs compensate customers with credits when average availability drops below certain levels. This is too inflexible because consumers lose non-measurable amounts of performance being only compensated later, in next charging cycles. We propose to schedule virtual machines (VMs), driven by range-based non-linear reductions of utility, different for classes of users and across different ranges of resource allocations: partial utility. This customer-defined metric, allows providers transferring resources between VMs in meaningful and economically efficient ways. We define a comprehensive cost model incorporating partial utility given by clients to a certain level of degradation, when VMs are allocated in overcommitted environments (Public, Private, Community Clouds). CloudSim was extended to support our scheduling model. Several simulation scenarios with synthetic and real workloads are presented, using datacenters with different dimensions regarding the number of servers and computational capacity. We show the partial utility-driven driven scheduling allows more VMs to be allocated. It brings benefits to providers, regarding revenue and resource utilization, allowing for more revenue per resource allocated and scaling well with the size of datacenters when comparing with an utility-oblivious redistribution of resources. Regarding clients, their workloads’ execution time is also improved, by incorporating an SLA-based redistribution of their VM’s computational power.
Resumo:
Data analytic applications are characterized by large data sets that are subject to a series of processing phases. Some of these phases are executed sequentially but others can be executed concurrently or in parallel on clusters, grids or clouds. The MapReduce programming model has been applied to process large data sets in cluster and cloud environments. For developing an application using MapReduce there is a need to install/configure/access specific frameworks such as Apache Hadoop or Elastic MapReduce in Amazon Cloud. It would be desirable to provide more flexibility in adjusting such configurations according to the application characteristics. Furthermore the composition of the multiple phases of a data analytic application requires the specification of all the phases and their orchestration. The original MapReduce model and environment lacks flexible support for such configuration and composition. Recognizing that scientific workflows have been successfully applied to modeling complex applications, this paper describes our experiments on implementing MapReduce as subworkflows in the AWARD framework (Autonomic Workflow Activities Reconfigurable and Dynamic). A text mining data analytic application is modeled as a complex workflow with multiple phases, where individual workflow nodes support MapReduce computations. As in typical MapReduce environments, the end user only needs to define the application algorithms for input data processing and for the map and reduce functions. In the paper we present experimental results when using the AWARD framework to execute MapReduce workflows deployed over multiple Amazon EC2 (Elastic Compute Cloud) instances.
Resumo:
In order to provide a more flexible learning environment in physics, the developed projectile launch apparatus enables students to determine the acceleration of gravity and the dependence of a set of parameters in the projectile movement. This apparatus is remotely operated and accessed via web, by first scheduling an access time slot. This machine has a number of configuration parameters that support different learning scenarios with different complexities.