960 resultados para Probabilidade de default
Resumo:
Information and Communication Technology (ICT) is becoming increasingly central to many people’s lives, making it possible to be connected in any place at any time, be unceasingly and instantly informed, and benefit from greater economic and educational opportunities. With all the benefits afforded by these new-found capabilities, however, come potential drawbacks. A plethora of new PCs, laptops, tablets, smartphones, Bluetooth, the internet, Wi-Fi (the list goes on) expect us to know or be able to guess, what, where and when to connect, click, double-click, tap, flick, scroll, in order to realise these benefits, and to have the physical and cognitive capability to do all these things. One of the groups most affected by this increase in high-demand technology is older people. They do not understand and use technology in the same way that younger generations do, because they grew up in the simpler electro-mechanical era and embedded that particular model of the world in their minds. Any consequential difficulty in familiarising themselves with modern ICT and effectively applying it to their needs can also be exacerbated by age-related changes in vision, motor control and cognitive functioning. Such challenges lead to digital exclusion. Much has been written about this topic over the years, usually by academics from the area of inclusive product design. The issue is complex and it is fair to say that no one researcher has the whole picture. It is difficult to understand and adequately address the issue of digital exclusion among the older generation without looking across disciplines and at industry’s and government’s understanding, motivation and efforts toward resolving this important problem. To do otherwise is to risk misunderstanding the true impact that ICT has and could have on people’s lives across all generations. In this European year of Active Ageing and Solidarity between Generations and as the British government is moving forward with its Digital by Default initiative as part of a wider objective to make ICT accessible to as many people as possible by 2015, the Engineering Design Centre (EDC) at the University of Cambridge collaborated with BT to produce a book of thought pieces to address, and where appropriate redress, these important and long-standing issues. “Ageing, Adaption and Accessibility: Time for the Inclusive Revolution!” brings together opinions and insights from twenty one prominent thought leaders from government, industry and academia regarding the problems, opportunities and strategies for combating digital exclusion among senior citizens. The contributing experts were selected as individuals, rather than representatives of organisations, to provide the broadest possible range of perspectives. They are renowned in their respective fields and their opinions are formed not only from their own work, but also from the contributions of others in their area. Their views were elicited through conversations conducted by the editors of this book who then drafted the thought pieces to be edited and approved by the experts. We hope that this unique collection of thought pieces will give you a broader perspective on ageing, people’s adaption to the ever changing world of technology and insights into better ways of designing digital devices and services for the older population.
Resumo:
Ideally, one would like to perform image search using an intuitive and friendly approach. Many existing image search engines, however, present users with sets of images arranged in some default order on the screen, typically the relevance to a query, only. While this certainly has its advantages, arguably, a more flexible and intuitive way would be to sort images into arbitrary structures such as grids, hierarchies, or spheres so that images that are visually or semantically alike are placed together. This paper focuses on designing such a navigation system for image browsers. This is a challenging task because arbitrary layout structure makes it difficult - if not impossible - to compute cross-similarities between images and structure coordinates, the main ingredient of traditional layouting approaches. For this reason, we resort to a recently developed machine learning technique: kernelized sorting. It is a general technique for matching pairs of objects from different domains without requiring cross-domain similarity measures and hence elegantly allows sorting images into arbitrary structures. Moreover, we extend it so that some images can be preselected for instance forming the tip of the hierarchy allowing to subsequently navigate through the search results in the lower levels in an intuitive way. Copyright 2010 ACM.
Resumo:
Biofuels are increasingly promoted worldwide as a means for reducing greenhouse gas (GHG) emissions from transport. However, current regulatory frameworks and most academic life cycle analyses adopt a deterministic approach in determining the GHG intensities of biofuels and thus ignore the inherent risk associated with biofuel production. This study aims to develop a transparent stochastic method for evaluating UK biofuels that determines both the magnitude and uncertainty of GHG intensity on the basis of current industry practices. Using wheat ethanol as a case study, we show that the GHG intensity could span a range of 40-110 gCO2e MJ-1 when land use change (LUC) emissions and various sources of uncertainty are taken into account, as compared with a regulatory default value of 44 gCO2e MJ-1. This suggests that the current deterministic regulatory framework underestimates wheat ethanol GHG intensity and thus may not be effective in evaluating transport fuels. Uncertainties in determining the GHG intensity of UK wheat ethanol include limitations of available data at a localized scale, and significant scientific uncertainty of parameters such as soil N2O and LUC emissions. Biofuel polices should be robust enough to incorporate the currently irreducible uncertainties and flexible enough to be readily revised when better science is available. © 2013 IOP Publishing Ltd.
Resumo:
In this work, we performed an evaluation of decay heat power of advanced, fast spectrum, lead and molten salt-cooled reactors, with flexible conversion ratio. The decay heat power was calculated using the BGCore computer code, which explicitly tracks over 1700 isotopes in the fuel throughout its burnup and subsequent decay. In the first stage, the capability of the BGCore code to accurately predict the decay heat power was verified by performing a benchmark calculation for a typical UO2 fuel in a Pressurized Water Reactor environment against the (ANSI/ANS-5.1-2005, "Decay Heat Power in Light Water Reactors," American National Standard) standard. Very good agreement (within 5%) between the two methods was obtained. Once BGCore calculation capabilities were verified, we calculated decay power for fast reactors with different coolants and conversion ratios, for which no standard procedure is currently available. Notable differences were observed for the decay power of the advanced reactor as compared with the conventional UO2 LWR. The importance of the observed differences was demonstrated by performing a simulation of a Station Blackout transient with the RELAP5 computer code for a lead-cooled fast reactor. The simulation was performed twice: using the code-default ANS-79 decay heat curve and using the curve calculated specifically for the studied core by BGCore code. The differences in the decay heat power resulted in failure to meet maximum cladding temperature limit criteria by ∼100 °C in the latter case, while in the transient simulation with the ANS-79 decay heat curve, all safety limits were satisfied. The results of this study show that the design of new reactor safety systems must be based on decay power curves specific to each individual case in order to assure the desired performance of these systems. © 2009 Elsevier B.V. All rights reserved.
Resumo:
Transmission Volume Phase Holographic Grating (VPHG) is adopted as spectral element in the real-time Optical Channel Performance Monitor (OCPM), which is in dire need in the Dense Wavelength -division-multiplexing(DATDM) system. And the tolerance of incident angle, which can be fully determined by two angles: 6 and (p, is finally inferred in this paper. Commonly, the default setting is that the incident plane is perpendicular to the fringes when the incident angle is mentioned. Now the situation out of the vertical is discussed. By combining the theoretic analysis of VPHG with its use in OCPM and changing 6 and (0 precisely in the computation and experiment, the two physical quantities which can fully specify the performance of VPHG the diffraction efficiency and the resolution, are analyzed. The results show that the diffraction efficiency varies greatly with the change of 6 or (p. But from the view of the whole C-band, only the peak diffraction efficiency drifts to another wavelength. As for the resolution, it deteriorates more rapidly than diffraction efficiency with the change of (p, while more slowly with the change of theta. Only if \phi\less than or equal to+/-1degrees and alpha(B) -0.5 less than or equal to theta less than or equal to alpha(B) + 0.5, the performance of the VPHG would be good enough to be used in OCPM system.
Resumo:
The simulating wave nearshore (SWAN) wave model has been widely used in coastal areas, lakes and estuaries. However, we found a poor agreement between modeling results and measurements in analyzing the chosen four typical cases when we used the default parameters of the source function formulas of the SWAN to make wave simulation for the Bohai Sea. Also, it was found that at the same wind process the simulated results of two wind generation expressions (Komen, Janssen) demonstrated a large difference. Further study showed that the proportionality coefficient alpha in linear growth term of wave growth source function plays an unperceived role in the process of wave development. Based on experiments and analysis, we thought that the coefficient alpha should change rather than be a constant. Therefore, the coefficient alpha changing with the variation of friction velocity U (*) was introduced into the linear growth term of wave growth source function. Four weather processes were adopted to validate the improvement in the linear growth term. The results from the improved coefficient alpha agree much better with the measurements than those from the default constant coefficient alpha. Furthermore, the large differences of results between Komen wind generation expression and Janssen wind generation expression were eliminated. We also experimented with the four weather processes to test the new white-capping mechanisms based on the cumulative steepness method. It was found that the parameters of the new white-capping mechanisms are not suitable for the Bohai Sea, but Alkyon's white-capping mechanisms can be applicable to the Bohai Sea after amendments, demonstrating that this improvement of parameter alpha can improve the simulated results of the Bohai Sea.
Resumo:
Seismic exploration is the main tools of exploration for petroleum. as the society needs more petroleum and the level of exploration is going up, the exploration in the area of complex geology construction is the main task in oil industry, so the seismic prestack depth migration appeared, it has good ability for complex construction imaging. Its result depends on the velocity model strongly. So for seismic prestack depth migration has become the main research area. In this thesis the difference in seismic prestack depth migration between our country and the abroad has been analyzed in system. the tomographical method with no layer velocity model, the residual curve velocity analysical method based on velocity model and the deleting method in pre-processing have been developed. In the thesis, the tomographysical method in velocity analysis is been analyzed at first. It characterized with perfection in theory and diffculity in application. This method use the picked first arrivial, compare the difference between the picked first arrival and the calculated arrival in theory velocity model, and then anti-projected the difference along the ray path to get the new velocity model. This method only has the hypothesis of high frequency, no other hypothesis. So it is very effective and has high efficiency. But this method has default still. The picking of first arrival is difficult in the prestack data. The reasons are the ratio of signal to noise is very low and many other event cross each other in prestack data. These phenomenon appear strongly in the complex geology construction area. Based on these a new tomophysical methos in velocity analysis with no layer velocity model is been developed. The aim is to solve the picking problem. It do not need picking the event time contiunely. You can picking in random depending on the reliability. This methos not only need the pick time as the routine tomographysical mehtod, but also the slope of event. In this methos we use the high slope analysis method to improve the precision of picking. In addition we also make research on the residual curve velocity analysis and find that its application is not good and the efficiency is low. The reasons is that the hypothesis is rigid and it is a local optimizing method, it can solve seismic velocity problem in the area with laterical strong velocity variation. A new method is developed to improve the precision of velocity model building . So far the pattern of seismic prestack depth migration is the same as it aborad. Before the work of velocity building the original seismic data must been corrected on a datum plane, and then to make the prestack depth migration work. As we know the successful example is in Mexico bay. It characterized with the simple surface layer construction, the pre-precessing is very simple and its precision is very high. But in our country the main seismic work is in land, the surface layer is very complex, in some area the error of pre-precessing is big, it affect the velocity building. So based on this a new method is developed to delete the per-precessing error and improve the precision of velocity model building. Our main work is, (1) developing a effective tomographical velocity building method with no layer velocity model. (2) a new high resolution slope analysis method is developed. (3) developing a global optimized residual curve velocity buliding method based on velocity model. (4) a effective method of deleting the pre-precessing error is developing. All the method as listed above has been ceritified by the theorical calculation and the actual seismic data.
Resumo:
Eight experiments tested how object array structure and learning location influenced the establishing and utilization of self-to-object and object-to-object spatial representations in locomotion and reorientation. In Experiment 1 to 4, participants learned either at the periphery of or amidst regular or irregular object array, and then pointed to objects while blindfolded in three conditions: before turning (baseline), after rotating 240 degrees (updating), and after disorientation (disorientation). In Experiment 5 to 8, participants received instruction to keep track of self-to-object or object-to-object spatial representations before rotation. In each condition, the configuration error, which means the standard deviation of the means per target object of the signed pointing errors, was calculated as the index of the fidelity of representation used in each condition. Results indicate that participants form both self-to-object and object-to-object spatial representations after learning an object-array. Object-array structure influences the selection of representation during updating. By default, object-to-object spatial representation is updated when people learned the regular object-array structure, and self-to-object spatial representation is updated when people learned the irregular object array. But people could also update the other representation when they are required to do so. The fidelity of representations will confine this kind of “switch”. People could only “switch” from a low fidelity representation to a high fidelity representation or between two representations of similar fidelity. They couldn’t “switch” from a high fidelity representation to a low fidelity representation. Leaning location might influence the fidelity of representations. When people learned at the periphery of object array, they could acquire both self-to-object and object-to-object spatial representations of high fidelity. But when people learned amidst the object array, they could only acquire self-to-object spatial representation of high fidelity, and the fidelity of object-to-object spatial representation was low.
Resumo:
2007
Resumo:
O objetivo deste documento é mostrar o potencial da integração de um sistema de informações geográficas (SIG) com um modelo de probabilidade, usando a distribuição de Poisson, para espacializar variáveis discretas. Modelos estatísticos são ferramentas importantes no estudo de variáveis ambientais, principalmente com a crescente importância da valoração do capital ambiental. A distribuição do Poisson é um bom modelo estatístico para manejo de variáveis discretas, pois mostra seu comportamento. Um passo posterior seria saber como essas variáveis se comportam no espaço, mostrando sua distribuição espacial. Nesse caso, os sistemas de informações geográficas (SIG) são bastante eficientes (Miranda, 2005). Para testar o uso de ambas as ferramentas e mostrar sua eficiência, este trabalho traz uma implementação específica usando uma variável ambiental discreta, secas mensais.
Resumo:
O objetivo deste documento é apresentar o potencial da integração de um sistema de informações geográficas (SIG) com um modelo estocástico, a distribuição de probabilidade Gama, para espacializar variáveis contínuas. Aplica-se um estudo de caso para a precipitação no Estado de São Paulo, usando uma série temporal de quinze anos com dados de precipitação diária, de janeiro de 1978 a dezembro de 1992.
Resumo:
Este trabalho teve como objetivo avaliar o crescimento e a producao da borracha de clones seringueira [Havea brasiliensis ( Wild. Ex Adr. De Juss.) Muell. Arg.] sob diferentes sistemas de sangria, em condição de Cerrado dos Municipios de Barro Alto Goianesia no Estado de Goais. O plantio foi feito em fevereiro de 1992, no espacamento de 8,0 x 2,5 m (500 plantas/ha), em talhoes de 8 a 10 hectares para cada um dos clones RRIM 600, GT 1, PB 217, PB 235, PR 107 e PR 255 os quais receberam as mesmas praticas de manejo. Aos oito anos de idade, foram feitas as seguintes avaliacoes: estande final; circunferencia do caule a 1,20 m do solo; porcentagens de plantas aptas a sangria; producao de borracha acumulada na caneca pesada mensalmente; incidencia de seca de painel. A producao foi avaliada em nove sistemas de sangria em meia espiral (1/2 S), praticados cinco dias por semana (5d/7) e 10 meses ao ano (10m/12), variando na frequência de sangria (d/4 e d/7 = a cada 4 e 7 dias), a concentracao de Ethephon (ET 0,25%, 2,5%,3,3% e 5,0%) e sua frequencia de aplicacao durante o periodo chuvoso ( a cada 22, 28 e 35 dias), como segue: 1) 1/2S, d/7, ET 2,5% a cada 22 dias; 2)1/2S, d/7, ET 2,5% a cada 30 dias (referncia); 3) 1/2S. d/4, ET 2,5% a cada 30 dias; 4) 1/2S, d/7, ET 3,3% a cada 22 dias; 5)1/2S, d/7, ET 3,3% a cada 30 dias; 6) 1/2S. d/7, ET 5,0% a cada 22 dias 7) 1/2S, d/7, ET 5,0% a cada 30 dias; 8) 1/2S, d/7, ET 5,0% a cada 35 dias; 9) 1/2S, d/7, ET 0,25% (pulverizando 10 ml por painel) a cada 22 dias. Nos sistemas 1 a 8, o Ethephon foi pincelado ( 1mL) na canaleta de corte e ate 2 cm acima dela (Pa e La). O delineamento experimental foi de blocos ao acaso, com quatro repeticoes de 10 plantas poe parcela. Cada clone constitui um experimento separado, sendo os resultados de producao acumulada anual submetidos a analise de variancia e, nos caso de significancia, as medias dos sistemas foram comparadas pelo teste Tukey, ao nível de 5% de probabilidade. Nao foi constatada qualquer incidencia de seca de painel e os resultados possibilitaram as seguintes conclusoes para as condicoes da regiao: 1) o sistema 1/2S, d/7, ET 2,5% a cada 30 dias e o mais indicado par a sangria dos clones PR 255, PR 107, PB 235, PB 217 e GT 1; 2) o sistema 1/2S, d/7, ET 3,3% a cada 30 dias e o mais indicado para a sangria do clone RRIM 600; 3) a producao individual de borracha em kg/planta/ano e maior nos clones RRIM 600, PB 217 e PR 255, enquanto a producao total em kg/ha/ano e superior nos clones RRIM 600 e PB 235; 4)os clones PB 217 e PR 255 sao menos adaptados a regiao, apresentando menores valores de estande final, circunferencia do caule, porcentagem de plantas em sangria e de producao total de borracha por hectare.
Resumo:
RESUMO: O presente trabalho teve como objetivo avaliar o teor e a produtividade de genótipos de girassol semeados em segunda safra no ano de 2014 em Campo Novo do Parecis ? MT, no campo experimental do Instituto Federal de Educação Ciência e Tecnologia de Mato Grosso. O delineamento experimental utilizado foi o de blocos casualizados, com 16 tratamentos (16 genótipos) e quatro repetições. As parcelas experimentais foram constituídas de 4 linhas com 6,5 m de comprimento, com espaçamento entrelinhas de 0,45 m, contendo área de 11,7 m², totalizando uma área de 748 m². Foi utilizada a população de 45000 plantas por hectare. Os dados foram submetidos à análise de variância e ao teste Scott-Knott, a 5% de probabilidade. Os genótipos que se destacaram em relação à produtividade de aquênios foram o MG 360, AGUARÁ 06, MG 305, AGUARÁ 04, CF 101, SYN 045, GNZ NEON, HELIO 251 e SYN 3950HO. Para o teor de óleo nos aquênios e produtividade de óleo, o genótipo MG 360 apresentou o maior valor e se destacou em relação aos demais genótipos analisados. ABSTRACT: This study aimed to evaluate genotypes of sunflower seeded second harvest in the year 2014 in Campus Campo Novo do Parecis, in the experimental field of the Instituto Federal de Educação Ciência e Tecnologia de Mato Grosso. The experimental design was a randomized block design with treatments 16 (16 genotypes) and four replications. The experimental plots consisted of four rows 6.5 m long with row spacing of 0.45 m, containing area of 11.7 m², totaling an area of 748 m². The population of 45000 plants per hectare is used. Data were subjected to analysis of variance and the Scott-Knott test at 5 % probability. The genotypes that stood out in relation to achenes productivity were MG 360, AGUARÁ 06, MG 305, AGUARÁ 04, CF 101, SYN 045, GNZ NEON, HELIO 251 and SYN 3950HO. For oil content and oil productivity, MG 360 genotype showed the highest value and stood out in relation to other genotypes.
Resumo:
RESUMO: O presente trabalho teve como objetivo avaliar genótipos de girassol semeados em segunda safra no ano de 2014 em Campo Novo do Parecis ? MT, no campo experimental do Instituto Federal de Educação Ciência e Tecnologia de Mato Grosso. O delineamento experimental utilizado foi o de blocos casualizados, com 16 tratamentos (16 genótipos) e quatro repetições. As parcelas experimentais foram constituídas de 4 linhas com 6,5 m de comprimento, com espaçamento entrelinhas de 0,45 m, contendo área de 11,7 m², totalizando uma área de 748m². Foi utilizada a população de 45000 plantas por hectare. Os dados foram submetidos à análise de variância e ao teste Scott-Knott, a 5% de probabilidade. Para a massa de mil aquênios, os genótipos que se destacaram foram BRS 323, MG 360 e M734 enquanto que as os mais produtivos foram os genótipos MG 360, AGUARÁ 06, MG 305, AGUARÁ 04, CF 101, SYN 045, GNZ NEON, HELIO 251 e SYN 3950HO. ABSTRACT: This study aimed to evaluate genotypes of sunflower seeded second harvest in the year 2014 in Campus Campo Novo do Parecis, in the experimental field of the Instituto Federal de Educação Ciência e Tecnologia de Mato Grosso. The experimental design was a randomized block design with treatments 16 (16 genotypes) and four replications. The experimental plots consisted of four rows 6.5 m long with row spacing of 0.45 m, containing area of 11.7 m², totaling an area of 748 m². The population of 45000 plants per hectare is used. Data were subjected to analysis of variance and the Scott - Knott test at 5 % probability. For the mass of thousand achenes, genotypes that stood out were BRS 323, MG 360 and M734 while the most productive genotypes were the MG 360, AGUARÁ 06, MG 305, AGUARÁ 04, CF 101, SYN 045, GNZ NEON, HELIO 251 and SYN 3950HO.
Resumo:
This paper presents experimental results that aimed to investigate the effects of soil liquefaction on the modal parameters (i.e. frequency and damping ratio) of pile-supported structures. The tests were carried out using the shaking table facility of the Bristol Laboratory for Advanced Dynamics Engineering (BLADE) at the University of Bristol (UK) whereby four pile-supported structures (two single piles and two pile groups) with and without superstructure mass were tested. The experimental investigation aimed to monitor the variation in natural frequency and damping of the four physical models at different degrees of excess pore water pressure generation and in full-liquefaction condition. The experimental results showed that the natural frequency of pile-supported structures may decrease considerably owing to the loss of lateral support offered by the soil to the pile. On the other hand, the damping ratio of structure may increase to values in excess of 20%. These findings have important design consequences: (a) for low-period structures, substantial reduction of spectral acceleration is expected; (b) during and after liquefaction, the response of the system may be dictated by the interactions of multiple loadings, that is, horizontal, axial and overturning moment, which were negligible prior to liquefaction; and (c) with the onset of liquefaction due to increased flexibility of pile-supported structure, larger spectral displacement may be expected, which in turn may enhance Pdelta effects and consequently amplification of overturning moment. Practical implications for pile design are discussed.