959 resultados para efficient algorithms
Resumo:
The driving forces for current research of flame retardants are increased fire safety in combination with flame retardant formulations that fulfill the criteria of sustainable production and products. In recent years, important questions about the environmental safety of antimony, and in particular, brominated flame retardants have been raised. As a consequence of this, the current doctoral thesis work describes efforts to develop new halogen-free flame retardants that are based on various radical generators and phosphorous compounds. The investigation was first focused on compounds that are capable of generating alkyl radicals in order to study their role on flame retardancy of polypropylene. The family of azoalkanes was selected as the cleanest and most convenient source of free alkyl radicals. Therefore, a number of symmetrical and unsymmetrical azoalkanes of the general formula R-N=N-R’ were prepared. The experimental results show that in the series of different sized azocycloalkanes the flame retardant efficacy decreased in the following order: R = R´= cyclohexyl > cyclopentyl > cyclobutyl > cyclooctanyl > cyclododecanyl. However, in the series of aliphatic azoalkanes compounds, the efficacy decreased as followed: R = R´= n-alkyl > tert-butyl > tert-octyl. The most striking difference in flame retardant efficacy was observed in thick polypropylene plaques of 1 mm, e.g. azocyclohexane (AZO) had a much better flame retardant performance than did the commercial reference FR (Flamestab® NOR116) in thick PP sections. In addition, some of the prepared azoalkane flame retardants e.g. 4’4- bis(cyclohexylazocyclohexyl) methane (BISAZO) exhibited non-burning dripping behavior. Extrusion coating experiments of flame retarded low density polyethylene (LDPE) onto a standard machine finished Kraft paper were carried out in order to investigate the potential of azoalkanes in multilayer facings. The results show that azocyclohexane (AZO) and 4’4-bis (cyclohexylazocyclohexyl) methane (BISAZO) can significantly improve the flame retardant properties of low density polyethylene coated paper already at 0.5 wt.% loadings, provided that the maximum extrusion temperature of 260 oC is not exceeded and coating weight is kept low at 13 g/m2. In addition, various triazene-based flame retardants (RN1=N2-N3R’R’’) were prepared. For example, polypropylene samples containing a very low concentration of only 0.5 wt.% of bis- 4’4’-(3’3’-dimethyltriazene) diphenyl ether and other triazenes passed the DIN 4102-1 test with B2 classification. It is noteworthy that no burning dripping could be detected and the average burning times were very short with exceptionally low weight losses. Therefore, triazene compounds constitute a new and interesting family of radical generators for flame retarding of polymeric materials. The high flame retardant potential of triazenes can be attributed to their ability to generate various types of radicals during their thermal decomposition. According to thermogravimetric analysis/Fourier transform infrared spectroscopy/MS analysis, triazene units are homolytically cleaved into various aminyl, resonance-stabilized aryl radicals, and different CH fragments with simultaneous evolution of elemental nitrogen. Furthermore, the potential of thirteen aliphatic, aromatic, thiuram and heterocyclic substituted organic disulfide derivatives of the general formula R-S-S-R’ as a new group of halogen-free flame retardants for polypropylene films have been investigated. According to the DIN 4102- 1 standard ignitibility test, for the first time it has been demonstrated that many of the disulfides alone can effectively provide flame retardancy and self-extinguishing properties to polypropylene films at already very low concentrations of 0.5 wt.%. For the disulfide family, the highest FR activity was recorded for 5’5’-dithiobis (2-nitrobenzoic acid). Very low values for burning length (53 mm) and burning time (10 s) reflect significantly increased fire retardant performance of this disulfide compared to other compounds in this series as well as to Flamestab® NOR116. Finally, two new, phosphorus-based flame retardants were synthesized: P’P-diphenyl phosphinic hydrazide (PAH) and melamine phenyl phosphonate (MPhP). The DIN 4102-1 test and the more stringent UL94 vertical burning test (UL94 V) were used to assess the formulations ability to extinguish a flame once ignited. A very strong synergistic effect with azoalkanes was found, i.e. in combination with these radical generators even UL94 V0 rate could be obtained.
Resumo:
Gasification of biomass is an efficient method process to produce liquid fuels, heat and electricity. It is interesting especially for the Nordic countries, where raw material for the processes is readily available. The thermal reactions of light hydrocarbons are a major challenge for industrial applications. At elevated temperatures, light hydrocarbons react spontaneously to form higher molecular weight compounds. In this thesis, this phenomenon was studied by literature survey, experimental work and modeling effort. The literature survey revealed that the change in tar composition is likely caused by the kinetic entropy. The role of the surface material is deemed to be an important factor in the reactivity of the system. The experimental results were in accordance with previous publications on the subject. The novelty of the experimental work lies in the used time interval for measurements combined with an industrially relevant temperature interval. The aspects which are covered in the modeling include screening of possible numerical approaches, testing of optimization methods and kinetic modelling. No significant numerical issues were observed, so the used calculation routines are adequate for the task. Evolutionary algorithms gave a better performance combined with better fit than the conventional iterative methods such as Simplex and Levenberg-Marquardt methods. Three models were fitted on experimental data. The LLNL model was used as a reference model to which two other models were compared. A compact model which included all the observed species was developed. The parameter estimation performed on that model gave slightly impaired fit to experimental data than LLNL model, but the difference was barely significant. The third tested model concentrated on the decomposition of hydrocarbons and included a theoretical description of the formation of carbon layer on the reactor walls. The fit to experimental data was extremely good. Based on the simulation results and literature findings, it is likely that the surface coverage of carbonaceous deposits is a major factor in thermal reactions.
Resumo:
To determine the effects of combined therapy of gliclazide and bedtime insulin on glycemic control and C-peptide secretion, we studied 25 patients with type 2 diabetes and sulfonylurea secondary failure, aged 56.8 ± 8.3 years, with a duration of diabetes of 10.6 ± 6.6 years, fasting plasma glucose of 277.3 ± 64.6 mg/dl and a body mass index of 27.4 ± 4.8 kg/m². Patients were submitted to three therapeutic regimens lasting 2 months each: 320 mg gliclazide (phase 1), 320 mg gliclazide and bedtime NPH insulin (phase 2), and insulin (phase 3). At the end of each period, glycemic and C-peptide curves in response to a mixed meal were determined. During combined therapy, there was a decrease in all glycemic curve values (P<0.01). Twelve patients (48%) reached fasting plasma glucose <140 mg/dl with a significant weight gain of 64.8 kg (43.1-98.8) vs 66.7 kg (42.8-101.4) (P<0.05), with no increase in C-peptide secretion or decrease in HbA1. C-Peptide glucose score (C-peptide/glucose x 100) increased from 0.9 (0.2-2.1) to 1.3 (0.2-4.7) during combined therapy (P<0.01). Despite a 50% increase in insulin doses in phase 3 (12 U (9-30) vs 18 U (11-60); P<0.01) only 3 patients who responded to combined therapy maintained fasting plasma glucose <140 mg/dl (P<0.02). A tendency to a higher absolute increase in C-peptide (0.99 (0.15-2.5) vs 0.6 (0-2.15); P = 0.08) and C-peptide incremental area (2.47 (0.22-6.2) vs 1.2 (0-3.35); P = 0.07) was observed among responders. We conclude that combined therapy resulted in a better glucose response to a mixed meal than insulin alone and should be tried in type 2 diabetic patients before starting insulin monotherapy, despite difficulties in predicting the response.
Resumo:
This thesis considers optimization problems arising in printed circuit board assembly. Especially, the case in which the electronic components of a single circuit board are placed using a single placement machine is studied. Although there is a large number of different placement machines, the use of collect-and-place -type gantry machines is discussed because of their flexibility and increasing popularity in the industry. Instead of solving the entire control optimization problem of a collect-andplace machine with a single application, the problem is divided into multiple subproblems because of its hard combinatorial nature. This dividing technique is called hierarchical decomposition. All the subproblems of the one PCB - one machine -context are described, classified and reviewed. The derived subproblems are then either solved with exact methods or new heuristic algorithms are developed and applied. The exact methods include, for example, a greedy algorithm and a solution based on dynamic programming. Some of the proposed heuristics contain constructive parts while others utilize local search or are based on frequency calculations. For the heuristics, it is made sure with comprehensive experimental tests that they are applicable and feasible. A number of quality functions will be proposed for evaluation and applied to the subproblems. In the experimental tests, artificially generated data from Markov-models and data from real-world PCB production are used. The thesis consists of an introduction and of five publications where the developed and used solution methods are described in their full detail. For all the problems stated in this thesis, the methods proposed are efficient enough to be used in the PCB assembly production in practice and are readily applicable in the PCB manufacturing industry.
Resumo:
The aim of this thesis was to create a process for all multi-site ramp-up (MSRU) projects in the case company in order to have simultaneous ramp-ups early in the market. The research was done through case study in one company and semi-structured interviews. There are already processes, which are now in use in MSRU-cases. Interviews of 20 ramp-up specialists revealed topics to be improved. Those were project team set up, roles and responsibilities and recommended project organization, communication, product change management practices, competence and know how transfer practices and support model. More R&D support and involvement is needed in MSRU-projects. DCM’s role is very important in the MSRU-projects among PMT-team; he should be the business owner of the project. Recommendation is that product programs could take care of the product and repair training of new products in volume factories. R&D’s participation in competence transfers is essential important in MSRU-projects. Communication in projects could be shared through special intranet commune. Blogging and tweeting could be considered in the communication plan. If hundreds of change notes are open in ramp-up phase, it should be considered not to approve the product into volume ramp-up. PMTs’ supports are also important and MSRU-projects should be planned, budgeted and executed together. Finally a new MSRU-process is presented in this thesis to be used in all MSRU-projects.
Resumo:
Due to various advantages such as flexibility, scalability and updatability, software intensive systems are increasingly embedded in everyday life. The constantly growing number of functions executed by these systems requires a high level of performance from the underlying platform. The main approach to incrementing performance has been the increase of operating frequency of a chip. However, this has led to the problem of power dissipation, which has shifted the focus of research to parallel and distributed computing. Parallel many-core platforms can provide the required level of computational power along with low power consumption. On the one hand, this enables parallel execution of highly intensive applications. With their computational power, these platforms are likely to be used in various application domains: from home use electronics (e.g., video processing) to complex critical control systems. On the other hand, the utilization of the resources has to be efficient in terms of performance and power consumption. However, the high level of on-chip integration results in the increase of the probability of various faults and creation of hotspots leading to thermal problems. Additionally, radiation, which is frequent in space but becomes an issue also at the ground level, can cause transient faults. This can eventually induce a faulty execution of applications. Therefore, it is crucial to develop methods that enable efficient as well as resilient execution of applications. The main objective of the thesis is to propose an approach to design agentbased systems for many-core platforms in a rigorous manner. When designing such a system, we explore and integrate various dynamic reconfiguration mechanisms into agents functionality. The use of these mechanisms enhances resilience of the underlying platform whilst maintaining performance at an acceptable level. The design of the system proceeds according to a formal refinement approach which allows us to ensure correct behaviour of the system with respect to postulated properties. To enable analysis of the proposed system in terms of area overhead as well as performance, we explore an approach, where the developed rigorous models are transformed into a high-level implementation language. Specifically, we investigate methods for deriving fault-free implementations from these models into, e.g., a hardware description language, namely VHDL.
Resumo:
We compared the cost-benefit of two algorithms, recently proposed by the Centers for Disease Control and Prevention, USA, with the conventional one, the most appropriate for the diagnosis of hepatitis C virus (HCV) infection in the Brazilian population. Serum samples were obtained from 517 ELISA-positive or -inconclusive blood donors who had returned to Fundação Pró-Sangue/Hemocentro de São Paulo to confirm previous results. Algorithm A was based on signal-to-cut-off (s/co) ratio of ELISA anti-HCV samples that show s/co ratio ³95% concordance with immunoblot (IB) positivity. For algorithm B, reflex nucleic acid amplification testing by PCR was required for ELISA-positive or -inconclusive samples and IB for PCR-negative samples. For algorithm C, all positive or inconclusive ELISA samples were submitted to IB. We observed a similar rate of positive results with the three algorithms: 287, 287, and 285 for A, B, and C, respectively, and 283 were concordant with one another. Indeterminate results from algorithms A and C were elucidated by PCR (expanded algorithm) which detected two more positive samples. The estimated cost of algorithms A and B was US$21,299.39 and US$32,397.40, respectively, which were 43.5 and 14.0% more economic than C (US$37,673.79). The cost can vary according to the technique used. We conclude that both algorithms A and B are suitable for diagnosing HCV infection in the Brazilian population. Furthermore, algorithm A is the more practical and economical one since it requires supplemental tests for only 54% of the samples. Algorithm B provides early information about the presence of viremia.
Resumo:
Our objective is to evaluate the accuracy of three algorithms in differentiating the origins of outflow tract ventricular arrhythmias (OTVAs). This study involved 110 consecutive patients with OTVAs for whom a standard 12-lead surface electrocardiogram (ECG) showed typical left bundle branch block morphology with an inferior axis. All the ECG tracings were retrospectively analyzed using the following three recently published ECG algorithms: 1) the transitional zone (TZ) index, 2) the V2 transition ratio, and 3) V2 R wave duration and R/S wave amplitude indices. Considering all patients, the V2 transition ratio had the highest sensitivity (92.3%), while the R wave duration and R/S wave amplitude indices in V2 had the highest specificity (93.9%). The latter finding had a maximal area under the ROC curve of 0.925. In patients with left ventricular (LV) rotation, the V2 transition ratio had the highest sensitivity (94.1%), while the R wave duration and R/S wave amplitude indices in V2 had the highest specificity (87.5%). The former finding had a maximal area under the ROC curve of 0.892. All three published ECG algorithms are effective in differentiating the origin of OTVAs, while the V2 transition ratio, and the V2 R wave duration and R/S wave amplitude indices are the most sensitive and specific algorithms, respectively. Amongst all of the patients, the V2 R wave duration and R/S wave amplitude algorithm had the maximal area under the ROC curve, but in patients with LV rotation the V2 transition ratio algorithm had the maximum area under the ROC curve.
Resumo:
Recent advances in Information and Communication Technology (ICT), especially those related to the Internet of Things (IoT), are facilitating smart regions. Among many services that a smart region can offer, remote health monitoring is a typical application of IoT paradigm. It offers the ability to continuously monitor and collect health-related data from a person, and transmit the data to a remote entity (for example, a healthcare service provider) for further processing and knowledge extraction. An IoT-based remote health monitoring system can be beneficial in rural areas belonging to the smart region where people have limited access to regular healthcare services. The same system can be beneficial in urban areas where hospitals can be overcrowded and where it may take substantial time to avail healthcare. However, this system may generate a large amount of data. In order to realize an efficient IoT-based remote health monitoring system, it is imperative to study the network communication needs of such a system; in particular the bandwidth requirements and the volume of generated data. The thesis studies a commercial product for remote health monitoring in Skellefteå, Sweden. Based on the results obtained via the commercial product, the thesis identified the key network-related requirements of a typical remote health monitoring system in terms of real-time event update, bandwidth requirements and data generation. Furthermore, the thesis has proposed an architecture called IReHMo - an IoT-based remote health monitoring architecture. This architecture allows users to incorporate several types of IoT devices to extend the sensing capabilities of the system. Using IReHMo, several IoT communication protocols such as HTTP, MQTT and CoAP has been evaluated and compared against each other. Results showed that CoAP is the most efficient protocol to transmit small size healthcare data to the remote servers. The combination of IReHMo and CoAP significantly reduced the required bandwidth as well as the volume of generated data (up to 56 percent) compared to the commercial product. Finally, the thesis conducted a scalability analysis, to determine the feasibility of deploying the combination of IReHMo and CoAP in large numbers in regions in north Sweden.
Resumo:
Many industrial applications need object recognition and tracking capabilities. The algorithms developed for those purposes are computationally expensive. Yet ,real time performance, high accuracy and small power consumption are essential measures of the system. When all these requirements are combined, hardware acceleration of these algorithms becomes a feasible solution. The purpose of this study is to analyze the current state of these hardware acceleration solutions, which algorithms have been implemented in hardware and what modifications have been done in order to adapt these algorithms to hardware.
Resumo:
Simplification of highly detailed CAD models is an important step when CAD models are visualized or by other means utilized in augmented reality applications. Without simplification, CAD models may cause severe processing and storage is- sues especially in mobile devices. In addition, simplified models may have other advantages like better visual clarity or improved reliability when used for visual pose tracking. The geometry of CAD models is invariably presented in form of a 3D mesh. In this paper, we survey mesh simplification algorithms in general and focus especially to algorithms that can be used to simplify CAD models. We test some commonly known algorithms with real world CAD data and characterize some new CAD related simplification algorithms that have not been surveyed in previous mesh simplification reviews.
Resumo:
Energy drinks are becoming popular in Brazil and in the world due to their stimulant properties. Caffeine is present in energy drinks with the aim of stimulating the central nervous system and intensifying brain activity. On the other hand, the ingestion of high doses of caffeine can cause undesirable symptoms such as anxiety and tachycardia. Therefore, it is necessary to monitor the caffeine content added to energy drinks to guarantee that the levels in the final product are in accordance with the labeling and within the legislation limits. The goal of this work was to validate a fast, efficient, and low-cost method for the determination of caffeine in energy drinks by micellar electrokinetic chromatography (MEKC). A total of seven brands were analyzed, each in three lots. The electrolyte was prepared with 50 mmol.L-1 of sodium dodecyl sulfate (SDS) and 10 mmol.L-1 of sodium carbonate (pH 11.0). The mean concentration of caffeine ranged from 122.8 to 318.6 mg.L-1. None of the brands had caffeine levels above the maximum limit. Considering the interval of confidence (95%), 72% of the samples had less caffeine than the amount informed on the product label.
Resumo:
This study aimed to evaluate the efficiency of natural biocides, brown and green propolis, for the control of bacterial contamination in the production of sugarcane spirit. The treatments consisted of brown and green propolis extracts, ampicillin, and a control and were assessed at the beginning and end of harvest season in ten fermentation cycles. In the microbiological analyses, the lactic acid bacteria were quantified in the inoculum before and after the treatment with biocides, and the viability of yeast cells during fermentation was evaluated. The levels of acids, glycerol, total residual reducing sugars, and ethanol were analyzed for the wine resulting from each fermentation cycle. A reduction in the number of bacterial contaminants in the inoculum in the treatments with the natural biocides was observed, but it did not affect the viability of yeast cells. The control of the contaminants led to the production of higher levels of ethanol and reduced acidity in the wine produced. The results of the use of brown and green propolis to control the growth microorganisms in the fermentation of sugarcane spirit can be of great importance for using alternative strategies to synthetic antibacterials in fermentation processes including other distilled beverage or spirits.
Resumo:
Recent developments in power electronics technology have made it possible to develop competitive and reliable low-voltage DC (LVDC) distribution networks. Further, islanded microgrids—isolated small-scale localized distribution networks— have been proposed to reliably supply power using distributed generations. However, islanded operations face many issues such as power quality, voltage regulation, network stability, and protection. In this thesis, an energy management system (EMS) that ensures efficient energy and power balancing and voltage regulation has been proposed for an LVDC island network utilizing solar panels for electricity production and lead-acid batteries for energy storage. The EMS uses the master/slave method with robust communication infrastructure to control the production, storage, and loads. The logical basis for the EMS operations has been established by proposing functionalities of the network components as well as by defining appropriate operation modes that encompass all situations. During loss-of-powersupply periods, load prioritizations and disconnections are employed to maintain the power supply to at least some loads. The proposed EMS ensures optimal energy balance in the network. A sizing method based on discrete-event simulations has also been proposed to obtain reliable capacities of the photovoltaic array and battery. In addition, an algorithm to determine the number of hours of electric power supply that can be guaranteed to the customers at any given location has been developed. The successful performances of all the proposed algorithms have been demonstrated by simulations.
Resumo:
This research attempted to address the question of the role of explicit algorithms and episodic contexts in the acquisition of computational procedures for regrouping in subtraction. Three groups of students having difficulty learning to subtract with regrouping were taught procedures for doing so through either an explicit algorithm, an episodic content or an examples approach. It was hypothesized that the use of an explicit algorithm represented in a flow chart format would facilitate the acquisition and retention of specific procedural steps relative to the other two conditions. On the other hand, the use of paragraph stories to create episodic content was expected to facilitate the retrieval of algorithms, particularly in a mixed presentation format. The subjects were tested on similar, near, and far transfer questions over a four-day period. Near and far transfer algorithms were also introduced on Day Two. The results suggested that both explicit and episodic context facilitate performance on questions requiring subtraction with regrouping. However, the differential effects of these two approaches on near and far transfer questions were not as easy to identify. Explicit algorithms may facilitate the acquisition of specific procedural steps while at the same time inhibiting the application of such steps to transfer questions. Similarly, the value of episodic context in cuing the retrieval of an algorithm may be limited by the ability of a subject to identify and classify a new question as an exemplar of a particular episodically deflned problem type or category. The implications of these findings in relation to the procedures employed in the teaching of Mathematics to students with learning problems are discussed in detail.