935 resultados para complex data


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Communication signal processing applications often involve complex-valued (CV) functional representations for signals and systems. CV artificial neural networks have been studied theoretically and applied widely in nonlinear signal and data processing [1–11]. Note that most artificial neural networks cannot be automatically extended from the real-valued (RV) domain to the CV domain because the resulting model would in general violate Cauchy-Riemann conditions, and this means that the training algorithms become unusable. A number of analytic functions were introduced for the fully CV multilayer perceptrons (MLP) [4]. A fully CV radial basis function (RBF) nework was introduced in [8] for regression and classification applications. Alternatively, the problem can be avoided by using two RV artificial neural networks, one processing the real part and the other processing the imaginary part of the CV signal/system. A even more challenging problem is the inverse of a CV

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work presents a model study for the formation of a dimeric dioxomolybdenum(VI) complex [MoO2L]2, generated by simultaneous satisfaction of acceptor and donor character existing in the corresponding monomeric Mo(VI) complex MoO2L. This mononuclear complex is specially designed to contain a coordinatively unsaturated Mo(VI) acceptor centre and a free donor group, (e.g. –NH2 group) strategically placed in the ligand skeleton [H2L = 2-hydroxyacetophenonehydrazone of 2-aminobenzoylhydrazine]. Apart from the dimer [MoO2L]2, complexes of the type MoO2L·B (where B = CH3OH, γ-picoline and imidazole) are also reported. All the complexes are characterized by elemental analysis, spectroscopic (UV–Vis, IR, 1H NMR) techniques and cyclic voltammetry. Single crystal X-ray structures of [MoO2L]2 (1), MoO2L·CH3OH (2), and MoO2L.(γ-pic) (3) have been determined and discussed. DFT calculation on these complexes corroborates experimental data and provides clue for the facile formation of this type of dimer not reported previously. The process of dimer formation may also be viewed as an interaction between two molecules of a specially designed complex acting as a monodentate ligand. This work is expected to open up a new field of design and synthesis of dimeric complexes through the process of symbiotic donor–acceptor (acid–base) interaction between two molecules of a specially designed monomer.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Smart healthcare is a complex domain for systems integration due to human and technical factors and heterogeneous data sources involved. As a part of smart city, it is such a complex area where clinical functions require smartness of multi-systems collaborations for effective communications among departments, and radiology is one of the areas highly relies on intelligent information integration and communication. Therefore, it faces many challenges regarding integration and its interoperability such as information collision, heterogeneous data sources, policy obstacles, and procedure mismanagement. The purpose of this study is to conduct an analysis of data, semantic, and pragmatic interoperability of systems integration in radiology department, and to develop a pragmatic interoperability framework for guiding the integration. We select an on-going project at a local hospital for undertaking our case study. The project is to achieve data sharing and interoperability among Radiology Information Systems (RIS), Electronic Patient Record (EPR), and Picture Archiving and Communication Systems (PACS). Qualitative data collection and analysis methods are used. The data sources consisted of documentation including publications and internal working papers, one year of non-participant observations and 37 interviews with radiologists, clinicians, directors of IT services, referring clinicians, radiographers, receptionists and secretary. We identified four primary phases of data analysis process for the case study: requirements and barriers identification, integration approach, interoperability measurements, and knowledge foundations. Each phase is discussed and supported by qualitative data. Through the analysis we also develop a pragmatic interoperability framework that summaries the empirical findings and proposes recommendations for guiding the integration in the radiology context.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Single-carrier (SC) block transmission with frequency-domain equalisation (FDE) offers a viable transmission technology for combating the adverse effects of long dispersive channels encountered in high-rate broadband wireless communication systems. However, for high bandwidthefficiency and high power-efficiency systems, the channel can generally be modelled by the Hammerstein system that includes the nonlinear distortion effects of the high power amplifier (HPA) at transmitter. For such nonlinear Hammerstein channels, the standard SC-FDE scheme no longer works. This paper advocates a complex-valued (CV) B-spline neural network based nonlinear SC-FDE scheme for Hammerstein channels. Specifically, We model the nonlinear HPA, which represents the CV static nonlinearity of the Hammerstein channel, by a CV B-spline neural network, and we develop two efficient alternating least squares schemes for estimating the parameters of the Hammerstein channel, including both the channel impulse response coefficients and the parameters of the CV B-spline model. We also use another CV B-spline neural network to model the inversion of the nonlinear HPA, and the parameters of this inverting B-spline model can easily be estimated using the standard least squares algorithm based on the pseudo training data obtained as a natural byproduct of the Hammerstein channel identification. Equalisation of the SC Hammerstein channel can then be accomplished by the usual one-tap linear equalisation in frequency domain as well as the inverse B-spline neural network model obtained in time domain. Extensive simulation results are included to demonstrate the effectiveness of our nonlinear SC-FDE scheme for Hammerstein channels.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ras of complex proteins (ROC) domains were identified in 2003 as GTP binding modules in large multidomain proteins from Dictyostelium discoideum. Research into the function of these domains exploded with their identification in a number of proteins linked to human disease, including leucine-rich repeat kinase 2 (LRRK2) and death-associated protein kinase 1 (DAPK1) in Parkinson’s disease and cancer, respectively. This surge in research has resulted in a growing body of data revealing the role that ROC domains play in regulating protein function and signaling pathways. In this review, recent advances in the structural informa- tion available for proteins containing ROC domains, along with insights into enzymatic function and the integration of ROC domains as molecular switches in a cellular and organismal context, are explored.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The human ROCO proteins are a family of multi-domain proteins sharing a conserved ROC-COR supra-domain. The family has four members: leu- cine-rich repeat kinase 1 (LRRK1), leucine-rich repeat kinase 2 (LRRK2), death-associated protein kinase 1 (DAPK1) and malignant fibrous histiocy- toma amplified sequences with leucine-rich tandem repeats 1 (MASL1). Previous studies of LRRK1/2 and DAPK1 have shown that the ROC (Ras of complex proteins) domain can bind and hydrolyse GTP, but the cellular consequences of this activity are still unclear. Here, the first biochemical characterization of MASL1 and the impact of GTP binding on MASL1 complex formation are reported. The results demonstrate that MASL1, similar to other ROCO proteins, can bind guanosine nucleotides via its ROC domain. Furthermore, MASL1 exists in two distinct cellular com- plexes associated with heat shock protein 60, and the formation of a low molecular weight pool of MASL1 is modulated by GTP binding. Finally, loss of GTP enhances MASL1 toxicity in cells. Taken together, these data point to a central role for the ROC/GTPase domain of MASL1 in the reg- ulation of its cellular function.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Diatom, geochemical and isotopic data provide a record of environmental change in Laguna La Gaiba, lowland Bolivia (17°450S, 57°350W), over the last ca. 25 000 years. High-resolution diatom analysis around the Last Glacial–Interglacial Transition provides new insights into this period of change. The full and late glacial lake was generally quite shallow, but with evidence of periodic flooding. At about 13 100 cal a BP, just before the start of the Younger Dryas chronozone, the diatoms indicate shallower water conditions, but there is a marked change at about 12 200 cal a BP indicating the onset of a period of high variability, with rising water levels punctuated by periodic drying. From ca. 11 800 to 10 000 cal a BP, stable, deeper water conditions persisted. There is evidence for drying in the early to middle Holocene, but not as pronounced as that reported from elsewhere in the southern hemisphere tropics of South America. This was followed by the onset of wetter conditions in the late Holocene consistent with insolation forcing. Conditions very similar to present were established about 2100 cal a BP. A complex response to both insolation forcing and millennial-scale events originating in the North Atlantic is noted.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Population modelling is increasingly recognised as a useful tool for pesticide risk assessment. For vertebrates that may ingest pesticides with their food, such as woodpigeon (Columba palumbus), population models that simulate foraging behaviour explicitly can help predicting both exposure and population-level impact. Optimal foraging theory is often assumed to explain the individual-level decisions driving distributions of individuals in the field, but it may not adequately predict spatial and temporal characteristics of woodpigeon foraging because of the woodpigeons’ excellent memory, ability to fly long distances, and distinctive flocking behaviour. Here we present an individual-based model (IBM) of the woodpigeon. We used the model to predict distributions of foraging woodpigeons that use one of six alternative foraging strategies: optimal foraging, memory-based foraging and random foraging, each with or without flocking mechanisms. We used pattern-oriented modelling to determine which of the foraging strategies is best able to reproduce observed data patterns. Data used for model evaluation were gathered during a long-term woodpigeon study conducted between 1961 and 2004 and a radiotracking study conducted in 2003 and 2004, both in the UK, and are summarised here as three complex patterns: the distributions of foraging birds between vegetation types during the year, the number of fields visited daily by individuals, and the proportion of fields revisited by them on subsequent days. The model with a memory-based foraging strategy and a flocking mechanism was the only one to reproduce these three data patterns, and the optimal foraging model produced poor matches to all of them. The random foraging strategy reproduced two of the three patterns but was not able to guarantee population persistence. We conclude that with the memory-based foraging strategy including a flocking mechanism our model is realistic enough to estimate the potential exposure of woodpigeons to pesticides. We discuss how exposure can be linked to our model, and how the model could be used for risk assessment of pesticides, for example predicting exposure and effects in heterogeneous landscapes planted seasonally with a variety of crops, while accounting for differences in land use between landscapes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a novel approach to the automatic classification of very large data sets composed of terahertz pulse transient signals, highlighting their potential use in biochemical, biomedical, pharmaceutical and security applications. Two different types of THz spectra are considered in the classification process. Firstly a binary classification study of poly-A and poly-C ribonucleic acid samples is performed. This is then contrasted with a difficult multi-class classification problem of spectra from six different powder samples that although have fairly indistinguishable features in the optical spectrum, they also possess a few discernable spectral features in the terahertz part of the spectrum. Classification is performed using a complex-valued extreme learning machine algorithm that takes into account features in both the amplitude as well as the phase of the recorded spectra. Classification speed and accuracy are contrasted with that achieved using a support vector machine classifier. The study systematically compares the classifier performance achieved after adopting different Gaussian kernels when separating amplitude and phase signatures. The two signatures are presented as feature vectors for both training and testing purposes. The study confirms the utility of complex-valued extreme learning machine algorithms for classification of the very large data sets generated with current terahertz imaging spectrometers. The classifier can take into consideration heterogeneous layers within an object as would be required within a tomographic setting and is sufficiently robust to detect patterns hidden inside noisy terahertz data sets. The proposed study opens up the opportunity for the establishment of complex-valued extreme learning machine algorithms as new chemometric tools that will assist the wider proliferation of terahertz sensing technology for chemical sensing, quality control, security screening and clinic diagnosis. Furthermore, the proposed algorithm should also be very useful in other applications requiring the classification of very large datasets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Early in 1996, the latest of the European incoherent-scatter (EISCAT) radars came into operation on the Svalbard islands. The EISCAT Svalbard Radar (ESR) has been built in order to study the ionosphere in the northern polar cap and in particular, the dayside cusp. Conditions in the upper atmosphere in the cusp region are complex, with magnetosheath plasma cascading freely into the atmosphere along open magnetic field lines as a result of magnetic reconnection at the dayside magnetopause. A model has been developed to predict the effects of pulsed reconnection and the subsequent cusp precipitation in the ionosphere. Using this model we have successfully recreated some of the major features seen in photometer and satellite data within the cusp. In this paper, the work is extended to predict the signatures of pulsed reconnection in ESR data when the radar is pointed along the magnetic field. It is expected that enhancements in both electron concentration and electron temperature will be observed. Whether these enhancements are continuous in time or occur as a series of separate events is shown to depend critically on where the open/closed field-line boundary is with respect to the radar. This is shown to be particularly true when reconnection pulses are superposed on a steady background rate.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In recent years, there has been an increasing interest in the adoption of emerging ubiquitous sensor network (USN) technologies for instrumentation within a variety of sustainability systems. USN is emerging as a sensing paradigm that is being newly considered by the sustainability management field as an alternative to traditional tethered monitoring systems. Researchers have been discovering that USN is an exciting technology that should not be viewed simply as a substitute for traditional tethered monitoring systems. In this study, we investigate how a movement monitoring measurement system of a complex building is developed as a research environment for USN and related decision-supportive technologies. To address the apparent danger of building movement, agent-mediated communication concepts have been designed to autonomously manage large volumes of exchanged information. In this study, we additionally detail the design of the proposed system, including its principles, data processing algorithms, system architecture, and user interface specifics. Results of the test and case study demonstrate the effectiveness of the USN-based data acquisition system for real-time monitoring of movement operations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In common with other positive-strand RNA viruses, replication of feline calicivirus (FCV) results in rearrangement of intracellular membranes and production of numerous membrane-bound vesicular structures on which viral genome replication is thought to occur. In this study, bioinformatics approaches have identified three of the FCV non-structural proteins, namely p32, p39 and p30, as potential transmembrane proteins. These proteins were able to target enhanced cyan fluorescent protein to membrane fractions where they behaved as integral membrane proteins. Immunofluorescence microscopy of these proteins expressed in cells showed co-localization with endoplasmic reticulum (ER) markers. Further electron microscopy analysis of cells co-expressing FCV p39 or p30 with a horseradish peroxidase protein containing the KDEL ER retention motif demonstrated gross morphological changes to the ER. Similar reorganization patterns, especially for those produced by p30, were observed in naturally infected Crandel-Rees feline kidney cells. Together, the data demonstrate that the p32, p39 and p30 proteins of FCV locate to the ER and lead to reorganization of ER membranes. This suggests that they may play a role in the generation of FCV replication complexes and that the endoplasmic reticulum may represent the potential source of the membrane vesicles induced during FCV infection.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: In many experimental pipelines, clustering of multidimensional biological datasets is used to detect hidden structures in unlabelled input data. Taverna is a popular workflow management system that is used to design and execute scientific workflows and aid in silico experimentation. The availability of fast unsupervised methods for clustering and visualization in the Taverna platform is important to support a data-driven scientific discovery in complex and explorative bioinformatics applications. Results: This work presents a Taverna plugin, the Biological Data Interactive Clustering Explorer (BioDICE), that performs clustering of high-dimensional biological data and provides a nonlinear, topology preserving projection for the visualization of the input data and their similarities. The core algorithm in the BioDICE plugin is Fast Learning Self Organizing Map (FLSOM), which is an improved variant of the Self Organizing Map (SOM) algorithm. The plugin generates an interactive 2D map that allows the visual exploration of multidimensional data and the identification of groups of similar objects. The effectiveness of the plugin is demonstrated on a case study related to chemical compounds. Conclusions: The number and variety of available tools and its extensibility have made Taverna a popular choice for the development of scientific data workflows. This work presents a novel plugin, BioDICE, which adds a data-driven knowledge discovery component to Taverna. BioDICE provides an effective and powerful clustering tool, which can be adopted for the explorative analysis of biological datasets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fire activity has varied globally and continuously since the last glacial maximum (LGM) in response to long-term changes in global climate and shorter-term regional changes in climate, vegetation, and human land use. We have synthesized sedimentary charcoal records of biomass burning since the LGM and present global maps showing changes in fire activity for time slices during the past 21,000 years (as differences in charcoal accumulation values compared to pre-industrial). There is strong broad-scale coherence in fire activity after the LGM, but spatial heterogeneity in the signals increases thereafter. In North America, Europe and southern South America, charcoal records indicate less-than-present fire activity during the deglacial period, from 21,000 to ∼11,000 cal yr BP. In contrast, the tropical latitudes of South America and Africa show greater-than-present fire activity from ∼19,000 to ∼17,000 cal yr BP and most sites from Indochina and Australia show greater-than-present fire activity from 16,000 to ∼13,000 cal yr BP. Many sites indicate greater-than-present or near-present activity during the Holocene with the exception of eastern North America and eastern Asia from 8,000 to ∼3,000 cal yr BP, Indonesia and Australia from 11,000 to 4,000 cal yr BP, and southern South America from 6,000 to 3,000 cal yr BP where fire activity was less than present. Regional coherence in the patterns of change in fire activity was evident throughout the post-glacial period. These complex patterns can largely be explained in terms of large-scale climate controls modulated by local changes in vegetation and fuel load

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Social network has gained remarkable attention in the last decade. Accessing social network sites such as Twitter, Facebook LinkedIn and Google+ through the internet and the web 2.0 technologies has become more affordable. People are becoming more interested in and relying on social network for information, news and opinion of other users on diverse subject matters. The heavy reliance on social network sites causes them to generate massive data characterised by three computational issues namely; size, noise and dynamism. These issues often make social network data very complex to analyse manually, resulting in the pertinent use of computational means of analysing them. Data mining provides a wide range of techniques for detecting useful knowledge from massive datasets like trends, patterns and rules [44]. Data mining techniques are used for information retrieval, statistical modelling and machine learning. These techniques employ data pre-processing, data analysis, and data interpretation processes in the course of data analysis. This survey discusses different data mining techniques used in mining diverse aspects of the social network over decades going from the historical techniques to the up-to-date models, including our novel technique named TRCM. All the techniques covered in this survey are listed in the Table.1 including the tools employed as well as names of their authors.