838 resultados para Mesh generation from image data
Resumo:
Sea surface temperature (SST) datasets have been generated from satellite observations for the period 1991–2010, intended for use in climate science applications. Attributes of the datasets specifically relevant to climate applications are: first, independence from in situ observations; second, effort to ensure homogeneity and stability through the time-series; third, context-specific uncertainty estimates attached to each SST value; and, fourth, provision of estimates of both skin SST (the fundamental measure- ment, relevant to air-sea fluxes) and SST at standard depth and local time (partly model mediated, enabling comparison with his- torical in situ datasets). These attributes in part reflect requirements solicited from climate data users prior to and during the project. Datasets consisting of SSTs on satellite swaths are derived from the Along-Track Scanning Radiometers (ATSRs) and Advanced Very High Resolution Radiometers (AVHRRs). These are then used as sole SST inputs to a daily, spatially complete, analysis SST product, with a latitude-longitude resolution of 0.05°C and good discrimination of ocean surface thermal features. A product user guide is available, linking to reports describing the datasets’ algorithmic basis, validation results, format, uncer- tainty information and experimental use in trial climate applications. Future versions of the datasets will span at least 1982–2015, better addressing the need in many climate applications for stable records of global SST that are at least 30 years in length.
Resumo:
he perspective European Supergrid would consist of an integrated power system network, where electricity demands from one country could be met by generation from another country. This paper makes use of a bi-linear fixed-effects model to analyse the determinants for trading electricity across borders among 34 countries connected by the European Supergrid. The key question that this paper aims to address is the extent to which the privatisation of European electricity markets has brought about higher cross-border trade of electricity. The analysis makes use of distance, price ratios, gate closure times, size of peaks and aggregate demand as standard determinants. Controlling for other standard determinants, it is concluded that privatisation in most cases led to higher power exchange and that the benefits are more significant where privatisation measures have been in place for a longer period.
Resumo:
The REgents PARk and Tower Environmental Experiment (REPARTEE) comprised two campaigns in London in October 2006 and October/November 2007. The experiment design involved measurements at a heavily trafficked roadside site, two urban background sites and an elevated site at 160–190 m above ground on the BT Tower, supplemented in the second campaign by Doppler lidar measurements of atmospheric vertical structure. A wide range of measurements of airborne particle physical metrics and chemical composition were made as well as measurements of a considerable range of gas phase species and the fluxes of both particulate and gas phase substances. Significant findings include (a) demonstration of the evaporation of traffic-generated nanoparticles during both horizontal and vertical atmospheric transport; (b) generation of a large base of information on the fluxes of nanoparticles, accumulation mode particles and specific chemical components of the aerosol and a range of gas phase species, as well as the elucidation of key processes and comparison with emissions inventories; (c) quantification of vertical gradients in selected aerosol and trace gas species which has demonstrated the important role of regional transport in influencing concentrations of sulphate, nitrate and secondary organic compounds within the atmosphere of London; (d) generation of new data on the atmospheric structure and turbulence above London, including the estimation of mixed layer depths; (e) provision of new data on trace gas dispersion in the urban atmosphere through the release of purposeful tracers; (f) the determination of spatial differences in aerosol particle size distributions and their interpretation in terms of sources and physico-chemical transformations; (g) studies of the nocturnal oxidation of nitrogen oxides and of the diurnal behaviour of nitrate aerosol in the urban atmosphere, and (h) new information on the chemical composition and source apportionment of particulate matter size fractions in the atmosphere of London derived both from bulk chemical analysis and aerosol mass spectrometry with two instrument types.
Resumo:
This paper details a strategy for modifying the source code of a complex model so that the model may be used in a data assimilation context, {and gives the standards for implementing a data assimilation code to use such a model}. The strategy relies on keeping the model separate from any data assimilation code, and coupling the two through the use of Message Passing Interface (MPI) {functionality}. This strategy limits the changes necessary to the model and as such is rapid to program, at the expense of ultimate performance. The implementation technique is applied in different models with state dimension up to $2.7 \times 10^8$. The overheads added by using this implementation strategy in a coupled ocean-atmosphere climate model are shown to be an order of magnitude smaller than the addition of correlated stochastic random errors necessary for some nonlinear data assimilation techniques.
Resumo:
Human brain imaging techniques, such as Magnetic Resonance Imaging (MRI) or Diffusion Tensor Imaging (DTI), have been established as scientific and diagnostic tools and their adoption is growing in popularity. Statistical methods, machine learning and data mining algorithms have successfully been adopted to extract predictive and descriptive models from neuroimage data. However, the knowledge discovery process typically requires also the adoption of pre-processing, post-processing and visualisation techniques in complex data workflows. Currently, a main problem for the integrated preprocessing and mining of MRI data is the lack of comprehensive platforms able to avoid the manual invocation of preprocessing and mining tools, that yields to an error-prone and inefficient process. In this work we present K-Surfer, a novel plug-in of the Konstanz Information Miner (KNIME) workbench, that automatizes the preprocessing of brain images and leverages the mining capabilities of KNIME in an integrated way. K-Surfer supports the importing, filtering, merging and pre-processing of neuroimage data from FreeSurfer, a tool for human brain MRI feature extraction and interpretation. K-Surfer automatizes the steps for importing FreeSurfer data, reducing time costs, eliminating human errors and enabling the design of complex analytics workflow for neuroimage data by leveraging the rich functionalities available in the KNIME workbench.
Resumo:
Operational forecasting centres are currently developing data assimilation systems for coupled atmosphere-ocean models. Strongly coupled assimilation, in which a single assimilation system is applied to a coupled model, presents significant technical and scientific challenges. Hence weakly coupled assimilation systems are being developed as a first step, in which the coupled model is used to compare the current state estimate with observations, but corrections to the atmosphere and ocean initial conditions are then calculated independently. In this paper we provide a comprehensive description of the different coupled assimilation methodologies in the context of four dimensional variational assimilation (4D-Var) and use an idealised framework to assess the expected benefits of moving towards coupled data assimilation. We implement an incremental 4D-Var system within an idealised single column atmosphere-ocean model. The system has the capability to run both strongly and weakly coupled assimilations as well as uncoupled atmosphere or ocean only assimilations, thus allowing a systematic comparison of the different strategies for treating the coupled data assimilation problem. We present results from a series of identical twin experiments devised to investigate the behaviour and sensitivities of the different approaches. Overall, our study demonstrates the potential benefits that may be expected from coupled data assimilation. When compared to uncoupled initialisation, coupled assimilation is able to produce more balanced initial analysis fields, thus reducing initialisation shock and its impact on the subsequent forecast. Single observation experiments demonstrate how coupled assimilation systems are able to pass information between the atmosphere and ocean and therefore use near-surface data to greater effect. We show that much of this benefit may also be gained from a weakly coupled assimilation system, but that this can be sensitive to the parameters used in the assimilation.
Resumo:
We utilized an ecosystem process model (SIPNET, simplified photosynthesis and evapotranspiration model) to estimate carbon fluxes of gross primary productivity and total ecosystem respiration of a high-elevation coniferous forest. The data assimilation routine incorporated aggregated twice-daily measurements of the net ecosystem exchange of CO2 (NEE) and satellite-based reflectance measurements of the fraction of absorbed photosynthetically active radiation (fAPAR) on an eight-day timescale. From these data we conducted a data assimilation experiment with fifteen different combinations of available data using twice-daily NEE, aggregated annual NEE, eight-day f AP AR, and average annual fAPAR. Model parameters were conditioned on three years of NEE and fAPAR data and results were evaluated to determine the information content from the different combinations of data streams. Across the data assimilation experiments conducted, model selection metrics such as the Bayesian Information Criterion and Deviance Information Criterion obtained minimum values when assimilating average annual fAPAR and twice-daily NEE data. Application of wavelet coherence analyses showed higher correlations between measured and modeled fAPAR on longer timescales ranging from 9 to 12 months. There were strong correlations between measured and modeled NEE (R2, coefficient of determination, 0.86), but correlations between measured and modeled eight-day fAPAR were quite poor (R2 = −0.94). We conclude that this inability to determine fAPAR on eight-day timescale would improve with the considerations of the radiative transfer through the plant canopy. Modeled fluxes when assimilating average annual fAPAR and annual NEE were comparable to corresponding results when assimilating twice-daily NEE, albeit at a greater uncertainty. Our results support the conclusion that for this coniferous forest twice-daily NEE data are a critical measurement stream for the data assimilation. The results from this modeling exercise indicate that for this coniferous forest, average annuals for satellite-based fAPAR measurements paired with annual NEE estimates may provide spatial detail to components of ecosystem carbon fluxes in proximity of eddy covariance towers. Inclusion of other independent data streams in the assimilation will also reduce uncertainty on modeled values.
Resumo:
In this invited article the authors present an evaluative report on the development of the MESHGuides project (http://www.meshguides.org/). MESHGuides’ objective is to provide education with an international knowledge management system. MESHGuides were conceived as research summaries for supporting teachers’ in developing evidence-based practice. Their aim is to enhance teachers’ capacity to engage actively with research in their own classrooms. The original thinking for MESH arose from the work of UK-based academics Professor Marilyn Leask and Dr Sarah Younie in response to a desire, which has recently gathered momentum in the UK, for the development of a more research-informed teaching profession and for the establishment of an on-line platform to support evidence-based practice (DfE, 2015; Leask and Younie 2001; OECD 2009). The focus of this article is on how the MESHGuides project was conceived and structured, the technical systems supporting it and the practical reality for academics and teachers of composing and using MESHGuides. The project and the guides are in the early stages of development, and discussion indicates future possibilities for more global engagement with this knowledge management system.
Resumo:
Periocular recognition has recently become an active topic in biometrics. Typically it uses 2D image data of the periocular region. This paper is the first description of combining 3D shape structure with 2D texture. A simple and effective technique using iterative closest point (ICP) was applied for 3D periocular region matching. It proved its strength for relatively unconstrained eye region capture, and does not require any training. Local binary patterns (LBP) were applied for 2D image based periocular matching. The two modalities were combined at the score-level. This approach was evaluated using the Bosphorus 3D face database, which contains large variations in facial expressions, head poses and occlusions. The rank-1 accuracy achieved from the 3D data (80%) was better than that for 2D (58%), and the best accuracy (83%) was achieved by fusing the two types of data. This suggests that significant improvements to periocular recognition systems could be achieved using the 3D structure information that is now available from small and inexpensive sensors.
Resumo:
Lake surface water temperatures (LSWTs) of 246 globally distributed large lakes were derived from Along-Track Scanning Radiometers (ATSR) for the period 1991–2011. The climatological cycles of mean LSWT derived from these data quantify on a global scale the responses of large lakes' surface temperatures to the annual cycle of forcing by solar radiation and the ambient meteorological conditions. LSWT cycles reflect the twice annual peak in net solar radiation for lakes between 1°S to 12°N. For lakes without a lake-mean seasonal ice cover, LSWT extremes exceed air temperatures by 0.5–1.7 °C for maximum and 0.7–1.9 °C for minimum temperature. The summer maximum LSWTs of lakes from 25°S to 35°N show a linear decrease with increasing altitude; −3.76 ± 0.17 °C km−1 (inline image = 0.95), marginally lower than the corresponding air temperature decrease with altitude −4.15 ± 0.24 °C km−1 (inline image = 0.95). Lake altitude of tropical lakes account for 0.78–0.83 (inline image) of the variation in the March to June LSWT–air temperature differences, with differences decreasing by 1.9 °C as the altitude increases from 500 to 1800 m above sea level (a.s.l.) We define an ‘open water phase’ as the length of time the lake-mean LSWT remains above 4 °C. There is a strong global correlation between the start and end of the lake-mean open water phase and the spring and fall 0 °C air temperature transition days, (inline image = 0.74 and 0.80, respectively), allowing for a good estimation of timing and length of the open water phase of lakes without LSWT observations. Lake depth, lake altitude and distance from coast further explain some of the inter-lake variation in the start and end of the open water phase.
Resumo:
A set of four eddy-permitting global ocean reanalyses produced in the framework of the MyOcean project have been compared over the altimetry period 1993–2011. The main differences among the reanalyses used here come from the data assimilation scheme implemented to control the ocean state by inserting reprocessed observations of sea surface temperature (SST), in situ temperature and salinity profiles, sea level anomaly and sea-ice concentration. A first objective of this work includes assessing the interannual variability and trends for a series of parameters, usually considered in the community as essential ocean variables: SST, sea surface salinity, temperature and salinity averaged over meaningful layers of the water column, sea level, transports across pre-defined sections, and sea ice parameters. The eddy-permitting nature of the global reanalyses allows also to estimate eddy kinetic energy. The results show that in general there is a good consistency between the different reanalyses. An intercomparison against experiments without data assimilation was done during the MyOcean project and we conclude that data assimilation is crucial for correctly simulating some quantities such as regional trends of sea level as well as the eddy kinetic energy. A second objective is to show that the ensemble mean of reanalyses can be evaluated as one single system regarding its reliability in reproducing the climate signals, where both variability and uncertainties are assessed through the ensemble spread and signal-to-noise ratio. The main advantage of having access to several reanalyses differing in the way data assimilation is performed is that it becomes possible to assess part of the total uncertainty. Given the fact that we use very similar ocean models and atmospheric forcing, we can conclude that the spread of the ensemble of reanalyses is mainly representative of our ability to gauge uncertainty in the assimilation methods. This uncertainty changes a lot from one ocean parameter to another, especially in global indices. However, despite several caveats in the design of the multi-system ensemble, the main conclusion from this study is that an eddy-permitting multi-system ensemble approach has become mature and our results provide a first step towards a systematic comparison of eddy-permitting global ocean reanalyses aimed at providing robust conclusions on the recent evolution of the oceanic state.
Resumo:
This paper describes new advances in the exploitation of oxygen A-band measurements from POLDER3 sensor onboard PARASOL, satellite platform within the A-Train. These developments result from not only an account of the dependence of POLDER oxygen parameters to cloud optical thickness τ and to the scene's geometrical conditions but also, and more importantly, from the finer understanding of the sensitivity of these parameters to cloud vertical extent. This sensitivity is made possible thanks to the multidirectional character of POLDER measurements. In the case of monolayer clouds that represent most of cloudy conditions, new oxygen parameters are obtained and calibrated from POLDER3 data colocalized with the measurements of the two active sensors of the A-Train: CALIOP/CALIPSO and CPR/CloudSat. From a parameterization that is (μs, τ) dependent, with μs the cosine of the solar zenith angle, a cloud top oxygen pressure (CTOP) and a cloud middle oxygen pressure (CMOP) are obtained, which are estimates of actual cloud top and middle pressures (CTP and CMP). Performances of CTOP and CMOP are presented by class of clouds following the ISCCP classification. In 2008, the coefficient of the correlation between CMOP and CMP is 0.81 for cirrostratus, 0.79 for stratocumulus, 0.75 for deep convective clouds. The coefficient of the correlation between CTOP and CTP is 0.75, 0.73, and 0.79 for the same cloud types. The score obtained by CTOP, defined as the confidence in the retrieval for a particular range of inferred value and for a given error, is higher than the one of MODIS CTP estimate. Scores of CTOP are the highest for bin value of CTP superior in numbers. For liquid (ice) clouds and an error of 30 hPa (50 hPa), the score of CTOP reaches 50% (70%). From the difference between CTOP and CMOP, a first estimate of the cloud vertical extent h is possible. A second estimate of h comes from the correlation between the angular standard deviation of POLDER oxygen pressure σPO2 and the cloud vertical extent. This correlation is studied in detail in the case of liquid clouds. It is shown to be spatially and temporally robust, except for clouds above land during winter months. The analysis of the correlation's dependence on the scene's characteristics leads to a parameterization providing h from σPO2. For liquid water clouds above ocean in 2008, the mean difference between the actual cloud vertical extent and the one retrieved from σPO2 (from the pressure difference) is 5 m (−12 m). The standard deviation of the mean difference is close to 1000 m for the two methods. POLDER estimates of the cloud geometrical thickness obtain a global score of 50% confidence for a relative error of 20% (40%) of the estimate for ice (liquid) clouds over ocean. These results need to be validated outside of the CALIPSO/CloudSat track.
Resumo:
The classification of galaxies as star forming or active is generally done in the ([O III]/H beta, [N II]/H alpha) plane. The Sloan Digital Sky Survey (SDSS) has revealed that, in this plane, the distribution of galaxies looks like the two wings of a seagull. Galaxies in the right wing are referred to as Seyfert/LINERs, leading to the idea that non-stellar activity in galaxies is a very common phenomenon. Here, we argue that a large fraction of the systems in the right wing could actually be galaxies which stopped forming stars. The ionization in these `retired` galaxies would be produced by hot post-asymptotic giant branch stars and white dwarfs. Our argumentation is based on a stellar population analysis of the galaxies via our STARLIGHT code and on photoionization models using the Lyman continuum radiation predicted for this population. The proportion of LINER galaxies that can be explained in such a way is, however, uncertain. We further show how observational selection effects account for the shape of the right wing. Our study suggests that nuclear activity may not be as common as thought. If retired galaxies do explain a large part of the seagull`s right wing, some of the work concerning nuclear activity in galaxies, as inferred from SDSS data, will have to be revised.
Resumo:
This paper is concerned with the computational efficiency of fuzzy clustering algorithms when the data set to be clustered is described by a proximity matrix only (relational data) and the number of clusters must be automatically estimated from such data. A fuzzy variant of an evolutionary algorithm for relational clustering is derived and compared against two systematic (pseudo-exhaustive) approaches that can also be used to automatically estimate the number of fuzzy clusters in relational data. An extensive collection of experiments involving 18 artificial and two real data sets is reported and analyzed. (C) 2011 Elsevier B.V. All rights reserved.
Resumo:
A large amount of biological data has been produced in the last years. Important knowledge can be extracted from these data by the use of data analysis techniques. Clustering plays an important role in data analysis, by organizing similar objects from a dataset into meaningful groups. Several clustering algorithms have been proposed in the literature. However, each algorithm has its bias, being more adequate for particular datasets. This paper presents a mathematical formulation to support the creation of consistent clusters for biological data. Moreover. it shows a clustering algorithm to solve this formulation that uses GRASP (Greedy Randomized Adaptive Search Procedure). We compared the proposed algorithm with three known other algorithms. The proposed algorithm presented the best clustering results confirmed statistically. (C) 2009 Elsevier Ltd. All rights reserved.