971 resultados para Divisor of Zero
Resumo:
In drawing a conclusion for this study, care must be taken in generalizing findings since the population of students and teachers investigated were limited to certain levels in the different schools and countries. This study recognized some complexity of the factors underlying the status of school gardening instruction and activities in Germany, Nigeria and the U.S. as inadequate time for decision-making in the process of gardening, motivation of teachers and students. This was seen as the major impediments that influenced the status of gardening in the three countries. However, these factors were considered to have affected students’ mode of participation in the school gardening projects. This research finding suggests that the promotion and encouragement of students in gardening activities will promote vegetable production and increasing the numbers of practical farmers. Gardening has the potential to create opportunities for learning in an environment where children are able to experience nature first hand and to use the shared experience for communication (Bowker & Tearle, 2007). Therefore, the need for students to be encouraged to participate in gardening programs as the benefit will not only reduce the rate of obesity currently spreading among youths, but will contribute to the improve knowledge on science subjects. To build a network between community, parents and schools, a parent’s community approach should be used as the curriculum. The community approach will tighten the link between schools; community members, parents, teachers and students. This will help facilitate a better gardening projects implementation. Through a close collaboration, teachers and students will be able to identify issues affecting communities and undertake action learning in collaboration with community organizations to assess community needs and plan the implementation strategies as parents are part of the community. The sense of efficacy is a central factor in motivational and learning processes that govern educational improvement, standard and performance on complex tasks of both teachers and students. Dedication and willingness are the major stimulator and achievement of a project. Through a stimulator and provision of incentives and facilities, schools can achieve the best in project development. Teachers and principals should be aware that students are the lever for achieving the set goals in schools. Failure to understand what students need will result in achieving zero result. Therefore, it is advised that schools focus more on how to lure students to work through proper collaboration with the parents and community members. Principals and teachers should identify areas where students need to be corrected, helping them to correct the problem will enable them be committed in the schools’ programs.
Resumo:
High-speed semiconductor lasers are an integral part in the implemen- tation of high-bit-rate optical communications systems. They are com- pact, rugged, reliable, long-lived, and relatively inexpensive sources of coherent light. Due to the very low attenuation window that exists in the silica based optical fiber at 1.55 μm and the zero dispersion point at 1.3 μm, they have become the mainstay of optical fiber com- munication systems. For the fabrication of lasers with gratings such as, distributed bragg reflector or distributed feedback lasers, etching is the most critical step. Etching defines the lateral dimmensions of the structure which determines the performance of optoelectronic devices. In this thesis studies and experiments were carried out about the exist- ing etching processes for InP and a novel dry etching process was de- veloped. The newly developed process was based on Cl2/CH4/H2/Ar chemistry and resulted in very smooth surfaces and vertical side walls. With this process the grating definition was significantly improved as compared to other technological developments in the respective field. A surface defined grating definition approach is used in this thesis work which does not require any re-growth steps and makes the whole fabrication process simpler and cost effective. Moreover, this grating fabrication process is fully compatible with nano-imprint lithography and can be used for high throughput low-cost manufacturing. With usual etching techniques reported before it is not possible to etch very deep because of aspect ratio dependent etching phenomenon where with increasing etch depth the etch rate slows down resulting in non-vertical side walls and footing effects. Although with our de- veloped process quite vertical side walls were achieved but footing was still a problem. To overcome the challenges related to grating defini- tion and deep etching, a completely new three step gas chopping dry etching process was developed. This was the very first time that a time multiplexed etching process for an InP based material system was demonstrated. The developed gas chopping process showed extra ordinary results including high mask selectivity of 15, moderate etch- ing rate, very vertical side walls and a record high aspect ratio of 41. Both the developed etching processes are completely compatible with nano imprint lithography and can be used for low-cost high-throughput fabrication. A large number of broad area laser, ridge waveguide laser, distributed feedback laser, distributed bragg reflector laser and coupled cavity in- jection grating lasers were fabricated using the developed one step etch- ing process. Very extensive characterization was done to optimize all the important design and fabrication parameters. The devices devel- oped have shown excellent performance with a very high side mode suppression ratio of more than 52 dB, an output power of 17 mW per facet, high efficiency of 0.15 W/A, stable operation over temperature and injected currents and a threshold current as low as 30 mA for almost 1 mm long device. A record high modulation bandwidth of 15 GHz with electron-photon resonance and open eye diagrams for 10 Gbps data transmission were also shown.
Resumo:
This article aims to analyse the progress taken by Brazil towards the accomplishment of a sustainable development, mainly highlighting the success of the Zero Hunger social programme in achieving the elimination of starvation. As one of the essential pillars to the sustainable development, the State’s social performance demands actions directed to the elimination of extreme poverty and hunger, establishing a basis for equitable growth. It is necessary to clarify the juridical and constitutional framework of Brazil, aiming to emphasise the Zero Hunger importance in achieving the Brazilian Republic’s goals and reaching an international pattern on sustainable development. Also, it will be stressed that the international order influences Brazilian domestic law and is one of the main aspects for the Zero Hunger program development.
Resumo:
Background: The most common application of imputation is to infer genotypes of a high-density panel of markers on animals that are genotyped for a low-density panel. However, the increase in accuracy of genomic predictions resulting from an increase in the number of markers tends to reach a plateau beyond a certain density. Another application of imputation is to increase the size of the training set with un-genotyped animals. This strategy can be particularly successful when a set of closely related individuals are genotyped. ----- Methods: Imputation on completely un-genotyped dams was performed using known genotypes from the sire of each dam, one offspring and the offspring’s sire. Two methods were applied based on either allele or haplotype frequencies to infer genotypes at ambiguous loci. Results of these methods and of two available software packages were compared. Quality of imputation under different population structures was assessed. The impact of using imputed dams to enlarge training sets on the accuracy of genomic predictions was evaluated for different populations, heritabilities and sizes of training sets. ----- Results: Imputation accuracy ranged from 0.52 to 0.93 depending on the population structure and the method used. The method that used allele frequencies performed better than the method based on haplotype frequencies. Accuracy of imputation was higher for populations with higher levels of linkage disequilibrium and with larger proportions of markers with more extreme allele frequencies. Inclusion of imputed dams in the training set increased the accuracy of genomic predictions. Gains in accuracy ranged from close to zero to 37.14%, depending on the simulated scenario. Generally, the larger the accuracy already obtained with the genotyped training set, the lower the increase in accuracy achieved by adding imputed dams. ----- Conclusions: Whenever a reference population resembling the family configuration considered here is available, imputation can be used to achieve an extra increase in accuracy of genomic predictions by enlarging the training set with completely un-genotyped dams. This strategy was shown to be particularly useful for populations with lower levels of linkage disequilibrium, for genomic selection on traits with low heritability, and for species or breeds for which the size of the reference population is limited.
Resumo:
Tunable Optical Sensor Arrays (TOSA) based on Fabry-Pérot (FP) filters, for high quality spectroscopic applications in the visible and near infrared spectral range are investigated within this work. The optical performance of the FP filters is improved by using ion beam sputtered niobium pentoxide (Nb2O5) and silicon dioxide (SiO2) Distributed Bragg Reflectors (DBRs) as mirrors. Due to their high refractive index contrast, only a few alternating pairs of Nb2O5 and SiO2 films can achieve DBRs with high reflectivity in a wide spectral range, while ion beam sputter deposition (IBSD) is utilized due to its ability to produce films with high optical purity. However, IBSD films are highly stressed; resulting in stress induced mirror curvature and suspension bending in the free standing filter suspensions of the MEMS (Micro-Electro-Mechanical Systems) FP filters. Stress induced mirror curvature results in filter transmission line degradation, while suspension bending results in high required filter tuning voltages. Moreover, stress induced suspension bending results in higher order mode filter operation which in turn degrades the optical resolution of the filter. Therefore, the deposition process is optimized to achieve both near zero absorption and low residual stress. High energy ion bombardment during film deposition is utilized to reduce the film density, and hence the film compressive stress. Utilizing this technique, the compressive stress of Nb2O5 is reduced by ~43%, while that for SiO2 is reduced by ~40%. Filters fabricated with stress reduced films show curvatures as low as 100 nm for 70 μm mirrors. To reduce the stress induced bending in the free standing filter suspensions, a stress optimized multi-layer suspension design is presented; with a tensile stressed metal sandwiched between two compressively stressed films. The stress in Physical Vapor Deposited (PVD) metals is therefore characterized for use as filter top-electrode and stress compensating layer. Surface micromachining is used to fabricate tunable FP filters in the visible spectral range using the above mentioned design. The upward bending of the suspensions is reduced from several micrometers to less than 100 nm and 250 nm for two different suspension layer combinations. Mechanical tuning of up to 188 nm is obtained by applying 40 V of actuation voltage. Alternatively, a filter line with transmission of 65.5%, Full Width at Half Maximum (FWHM) of 10.5 nm and a stopband of 170 nm (at an output wavelength of 594 nm) is achieved. Numerical model simulations are also performed to study the validity of the stress optimized suspension design for the near infrared spectral range, wherein membrane displacement and suspension deformation due to material residual stress is studied. Two bandpass filter designs based on quarter-wave and non-quarter-wave layers are presented as integral components of the TOSA. With a filter passband of 135 nm and a broad stopband of over 650 nm, high average filter transmission of 88% is achieved inside the passband, while maximum filter transmission of less than 1.6% outside the passband is achieved.
Resumo:
This thesis addresses the problem of developing automatic grasping capabilities for robotic hands. Using a 2-jointed and a 4-jointed nmodel of the hand, we establish the geometric conditions necessary for achieving form closure grasps of cylindrical objects. We then define and show how to construct the grasping pre-image for quasi-static (friction dominated) and zero-G (inertia dominated) motions for sensorless and sensor-driven grasps with and without arm motions. While the approach does not rely on detailed modeling, it is computationally inexpensive, reliable, and easy to implement. Example behaviors were successfully implemented on the Salisbury hand and on a planar 2-fingered, 4 degree-of-freedom hand.
Resumo:
The dynamic power requirement of CMOS circuits is rapidly becoming a major concern in the design of personal information systems and large computers. In this work we present a number of new CMOS logic families, Charge Recovery Logic (CRL) as well as the much improved Split-Level Charge Recovery Logic (SCRL), within which the transfer of charge between the nodes occurs quasistatically. Operating quasistatically, these logic families have an energy dissipation that drops linearly with operating frequency, i.e., their power consumption drops quadratically with operating frequency as opposed to the linear drop of conventional CMOS. The circuit techniques in these new families rely on constructing an explicitly reversible pipelined logic gate, where the information necessary to recover the energy used to compute a value is provided by computing its logical inverse. Information necessary to uncompute the inverse is available from the subsequent inverse logic stage. We demonstrate the low energy operation of SCRL by presenting the results from the testing of the first fully quasistatic 8 x 8 multiplier chip (SCRL-1) employing SCRL circuit techniques.
Resumo:
In this paper we consider the problem of approximating a function belonging to some funtion space Φ by a linear comination of n translates of a given function G. Ussing a lemma by Jones (1990) and Barron (1991) we show that it is possible to define function spaces and functions G for which the rate of convergence to zero of the erro is 0(1/n) in any number of dimensions. The apparent avoidance of the "curse of dimensionality" is due to the fact that these function spaces are more and more constrained as the dimension increases. Examples include spaces of the Sobolev tpe, in which the number of weak derivatives is required to be larger than the number of dimensions. We give results both for approximation in the L2 norm and in the Lc norm. The interesting feature of these results is that, thanks to the constructive nature of Jones" and Barron"s lemma, an iterative procedure is defined that can achieve this rate.
Resumo:
This paper considers a connection between the deterministic and noisy behavior of nonlinear networks. Specifically, a particular bridge circuit is examined which has two possibly nonlinear energy storage elements. By proper choice of the constitutive relations for the network elements, the deterministic terminal behavior reduces to that of a single linear resistor. This reduction of the deterministic terminal behavior, in which a natural frequency of a linear circuit does not appear in the driving-point impedance, has been shown in classical circuit theory books (e.g. [1, 2]). The paper shows that, in addition to the reduction of the deterministic behavior, the thermal noise at the terminals of the network, arising from the usual Nyquist-Johnson noise model associated with each resistor in the network, is also exactly that of a single linear resistor. While this result for the linear time-invariant (LTI) case is a direct consequence of a well-known result for RLC circuits, the nonlinear result is novel. We show that the terminal noise current is precisely that predicted by the Nyquist-Johnson model for R if the driving voltage is zero or constant, but not if the driving voltage is time-dependent or the inductor and capacitor are time-varying
Resumo:
As stated in Aitchison (1986), a proper study of relative variation in a compositional data set should be based on logratios, and dealing with logratios excludes dealing with zeros. Nevertheless, it is clear that zero observations might be present in real data sets, either because the corresponding part is completely absent –essential zeros– or because it is below detection limit –rounded zeros. Because the second kind of zeros is usually understood as “a trace too small to measure”, it seems reasonable to replace them by a suitable small value, and this has been the traditional approach. As stated, e.g. by Tauber (1999) and by Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000), the principal problem in compositional data analysis is related to rounded zeros. One should be careful to use a replacement strategy that does not seriously distort the general structure of the data. In particular, the covariance structure of the involved parts –and thus the metric properties– should be preserved, as otherwise further analysis on subpopulations could be misleading. Following this point of view, a non-parametric imputation method is introduced in Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000). This method is analyzed in depth by Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2003) where it is shown that the theoretical drawbacks of the additive zero replacement method proposed in Aitchison (1986) can be overcome using a new multiplicative approach on the non-zero parts of a composition. The new approach has reasonable properties from a compositional point of view. In particular, it is “natural” in the sense that it recovers the “true” composition if replacement values are identical to the missing values, and it is coherent with the basic operations on the simplex. This coherence implies that the covariance structure of subcompositions with no zeros is preserved. As a generalization of the multiplicative replacement, in the same paper a substitution method for missing values on compositional data sets is introduced
Resumo:
There is almost not a case in exploration geology, where the studied data doesn’t includes below detection limits and/or zero values, and since most of the geological data responds to lognormal distributions, these “zero data” represent a mathematical challenge for the interpretation. We need to start by recognizing that there are zero values in geology. For example the amount of quartz in a foyaite (nepheline syenite) is zero, since quartz cannot co-exists with nepheline. Another common essential zero is a North azimuth, however we can always change that zero for the value of 360°. These are known as “Essential zeros”, but what can we do with “Rounded zeros” that are the result of below the detection limit of the equipment? Amalgamation, e.g. adding Na2O and K2O, as total alkalis is a solution, but sometimes we need to differentiate between a sodic and a potassic alteration. Pre-classification into groups requires a good knowledge of the distribution of the data and the geochemical characteristics of the groups which is not always available. Considering the zero values equal to the limit of detection of the used equipment will generate spurious distributions, especially in ternary diagrams. Same situation will occur if we replace the zero values by a small amount using non-parametric or parametric techniques (imputation). The method that we are proposing takes into consideration the well known relationships between some elements. For example, in copper porphyry deposits, there is always a good direct correlation between the copper values and the molybdenum ones, but while copper will always be above the limit of detection, many of the molybdenum values will be “rounded zeros”. So, we will take the lower quartile of the real molybdenum values and establish a regression equation with copper, and then we will estimate the “rounded” zero values of molybdenum by their corresponding copper values. The method could be applied to any type of data, provided we establish first their correlation dependency. One of the main advantages of this method is that we do not obtain a fixed value for the “rounded zeros”, but one that depends on the value of the other variable. Key words: compositional data analysis, treatment of zeros, essential zeros, rounded zeros, correlation dependency
Resumo:
The log-ratio methodology makes available powerful tools for analyzing compositional data. Nevertheless, the use of this methodology is only possible for those data sets without null values. Consequently, in those data sets where the zeros are present, a previous treatment becomes necessary. Last advances in the treatment of compositional zeros have been centered especially in the zeros of structural nature and in the rounded zeros. These tools do not contemplate the particular case of count compositional data sets with null values. In this work we deal with \count zeros" and we introduce a treatment based on a mixed Bayesian-multiplicative estimation. We use the Dirichlet probability distribution as a prior and we estimate the posterior probabilities. Then we apply a multiplicative modi¯cation for the non-zero values. We present a case study where this new methodology is applied. Key words: count data, multiplicative replacement, composition, log-ratio analysis
Resumo:
The object of this experience is to offer the students the opportunity to take part in the construction of a pedagogic strategy centred on the ludic, for the promotion of the integral health and the prevention of the disease with an educational community; directed to supporting and qualifying the well-being so much individually as group. The project is designed to five years, about interdisciplinary character (Speech Therapy, Medicine, Psychology, Nursery, Occupational Therapy), interinstitutional (Universidad del Rosario, Universidad de San Buenaventura y Universidad de Cundinamarca) and intersectorial (Education and Health). It considers the different actors of the educational community and school and the home as propitious scenes for the strengthening potential, beside being the fundamental spaces for the construction of knowledges and learnings concerning the integral health. To achieve the target, one has come constructing from the second semester of 2003, one pedagogic strategy centred on the ludic and the creativity, from which they are planned, they develop and evaluate the actions of promotion of skills, values, behaviors and attitudes in the care of the health and the prevention of disease, orientated to the early, opportune and effective detection of risk factors and problematic of the development that they affect the integral health. The above mentioned strategy raises a so called scene Bienestarópolis: A healthy world for conquering, centred on prominent figures, spaces and elements that alternate between the fantasy and the reality to facilitate the approximation, the interiorización and the appropriation of the integral health. Across this one, the children motivated by the adults enter an imaginary world in that theirs desires, knowledges and attitudes are the axis of his development. Since Vigotsky raises it, in the game the child realizes actions in order to adapt to the world that surrounds it acquiring skills for the learning. The actions of the project have involved 410 children and 25 teachers, of the degrees Zero, The First and The Second of basic primary; 90 parents of family, and an average of 40 students and 8 teachers of the already mentioned disciplines.
Resumo:
In the finite field (FF) treatment of vibrational polarizabilities and hyperpolarizabilities, the field-free Eckart conditions must be enforced in order to prevent molecular reorientation during geometry optimization. These conditions are implemented for the first time. Our procedure facilities identification of field-induced internal coordinates that make the major contribution to the vibrational properties. Using only two of these coordinates, quantitative accuracy for nuclear relaxation polarizabilities and hyperpolarizabilities is achieved in π-conjugated systems. From these two coordinates a single most efficient natural conjugation coordinate (NCC) can be extracted. The limitations of this one coordinate approach are discussed. It is shown that the Eckart conditions can lead to an isotope effect that is comparable to the isotope effect on zero-point vibrational averaging, but with a different mass-dependence
Resumo:
To obtain a state-of-the-art benchmark potential energy surface (PES) for the archetypal oxidative addition of the methane C-H bond to the palladium atom, we have explored this PES using a hierarchical series of ab initio methods (Hartree-Fock, second-order Møller-Plesset perturbation theory, fourth-order Møller-Plesset perturbation theory with single, double and quadruple excitations, coupled cluster theory with single and double excitations (CCSD), and with triple excitations treated perturbatively [CCSD(T)]) and hybrid density functional theory using the B3LYP functional, in combination with a hierarchical series of ten Gaussian-type basis sets, up to g polarization. Relativistic effects are taken into account either through a relativistic effective core potential for palladium or through a full four-component all-electron approach. Counterpoise corrected relative energies of stationary points are converged to within 0.1-0.2 kcal/mol as a function of the basis-set size. Our best estimate of kinetic and thermodynamic parameters is -8.1 (-8.3) kcal/mol for the formation of the reactant complex, 5.8 (3.1) kcal/mol for the activation energy relative to the separate reactants, and 0.8 (-1.2) kcal/mol for the reaction energy (zero-point vibrational energy-corrected values in parentheses). This agrees well with available experimental data. Our work highlights the importance of sufficient higher angular momentum polarization functions, f and g, for correctly describing metal-d-electron correlation and, thus, for obtaining reliable relative energies. We show that standard basis sets, such as LANL2DZ+ 1f for palladium, are not sufficiently polarized for this purpose and lead to erroneous CCSD(T) results. B3LYP is associated with smaller basis set superposition errors and shows faster convergence with basis-set size but yields relative energies (in particular, a reaction barrier) that are ca. 3.5 kcal/mol higher than the corresponding CCSD(T) values