924 resultados para Iterative probing
Resumo:
A study is made of the recognition and transformation of figures by iterative arrays of finite state automata. A figure is a finite rectangular two-dimensional array of symbols. The iterative arrays considered are also finite, rectangular, and two-dimensional. The automata comprising any given array are called cells and are assumed to be isomorphic and to operate synchronously with the state of a cell at time t+1 being a function of the states of it and its four nearest neighbors at time t. At time t=0 each cell is placed in one of a fixed number of initial states. The pattern of initial states thus introduced represents the figure to be processed. The resulting sequence of array states represents a computation based on the input figure. If one waits for a specially designated cell to indicate acceptance or rejection of the figure, the array is said to be working on a recognition problem. If one waits for the array to come to a stable configuration representing an output figure, the array is said to be working on a transformation problem.
Resumo:
M. Galea and Q. Shen. Iterative vs Simultaneous Fuzzy Rule Induction. Proceedings of the 14th International Conference on Fuzzy Systems, pages 767-772.
Resumo:
M. Galea, Q. Shen and V. Singh. Encouraging Complementary Fuzzy Rules within Iterative Rule Learning. Proceedings of the 2005 UK Workshop on Computational Intelligence, pages 15-22.
Resumo:
We present results of calculations [1] that employ a new mixed quantum classical iterative density matrix propagation approach (ILDM , or so called Is‐Landmap) [2] to explore the survival of coherence in different photo synthetic models. Our model studies confirm the long lived quantum coherence , while conventional theoretical tools (such as Redfield equation) fail to describe these phenomenon [3,4]. Our ILDM method is a numerical exactly propagation scheme and can be served as a bench mark calculation tools[2]. Result get from ILDM and from other recent methods have been compared and show agreement with each other[4,5]. Long lived coherence plateau has been attribute to the shift of harmonic potential due to the system bath interaction, and the harvesting efficiency is a balance between the coherence and dissipation[1]. We use this approach to investigate the excitation energy transfer dynamics in various light harvesting complex include Fenna‐Matthews‐Olsen light harvesting complex[1] and Cryptophyte Phycocyanin 645 [6]. [1] P.Huo and D.F.Coker ,J. Chem. Phys. 133, 184108 (2010) . [2] E.R. Dunkel, S. Bonella, and D.F. Coker, J. Chem. Phys. 129, 114106 (2008). [3] A. Ishizaki and G.R. Fleming, J. Chem. Phys. 130, 234111 (2009). [4] A. Ishizaki and G.R. Fleming, Proc. Natl. Acad. Sci. 106, 17255 (2009). [5] G. Tao and W.H. Miller, J. Phys. Chem. Lett. 1, 891 (2010). [6] P.Huo and D.F.Coker in preparation
Resumo:
For communication-intensive parallel applications, the maximum degree of concurrency achievable is limited by the communication throughput made available by the network. In previous work [HPS94], we showed experimentally that the performance of certain parallel applications running on a workstation network can be improved significantly if a congestion control protocol is used to enhance network performance. In this paper, we characterize and analyze the communication requirements of a large class of supercomputing applications that fall under the category of fixed-point problems, amenable to solution by parallel iterative methods. This results in a set of interface and architectural features sufficient for the efficient implementation of the applications over a large-scale distributed system. In particular, we propose a direct link between the application and network layer, supporting congestion control actions at both ends. This in turn enhances the system's responsiveness to network congestion, improving performance. Measurements are given showing the efficacy of our scheme to support large-scale parallel computations.
Resumo:
Replication is a commonly proposed solution to problems of scale associated with distributed services. However, when a service is replicated, each client must be assigned a server. Prior work has generally assumed that assignment to be static. In contrast, we propose dynamic server selection, and show that it enables application-level congestion avoidance. To make dynamic server selection practical, we demonstrate the use of three tools. In addition to direct measurements of round-trip latency, we introduce and validate two new tools: bprobe, which estimates the maximum possible bandwidth along a given path; and cprobe, which estimates the current congestion along a path. Using these tools we demonstrate dynamic server selection and compare it to previous static approaches. We show that dynamic server selection consistently outperforms static policies by as much as 50%. Furthermore, we demonstrate the importance of each of our tools in performing dynamic server selection.
Resumo:
Growing interest in inference and prediction of network characteristics is justified by its importance for a variety of network-aware applications. One widely adopted strategy to characterize network conditions relies on active, end-to-end probing of the network. Active end-to-end probing techniques differ in (1) the structural composition of the probes they use (e.g., number and size of packets, the destination of various packets, the protocols used, etc.), (2) the entity making the measurements (e.g. sender vs. receiver), and (3) the techniques used to combine measurements in order to infer specific metrics of interest. In this paper, we present Periscope: a Linux API that enables the definition of new probing structures and inference techniques from user space through a flexible interface. PeriScope requires no support from clients beyond the ability to respond to ICMP ECHO REQUESTs and is designed to minimize user/kernel crossings and to ensure various constraints (e.g., back-to-back packet transmissions, fine-grained timing measurements) We show how to use Periscope for two different probing purposes, namely the measurement of shared packet losses between pairs of endpoints and for the measurement of subpath bandwidth. Results from Internet experiments for both of these goals are also presented.
Resumo:
info:eu-repo/semantics/published
Resumo:
The idealized system of an atomically flat metallic surface [highly oriented pyrolytic graphite (HOPG)] and an organic monolayer (porphyrin) was used to determine whether the dielectric function and associated properties of thin films can be accessed with scanning-near-field scanning optical microscopy (s-NSOM). Here, we demonstrate the use of harmonics up to fourth order and the polarization dependence of incident light to probe dielectric properties on idealized samples of monolayers of organic molecules on atomically smooth substrates. An analytical treatment of light/sample interaction using the s-NSOM tip was developed in order to quantify the dielectric properties. The theoretical analysis and numerical modeling, as well as experimental data, demonstrate that higher order harmonic scattering can be used to extract the dielectric properties of materials with tens of nanometer spatial resolution. To date, the third harmonic provides the best lateral resolution (∼50 nm) and dielectric constant contrast for a porphyrin film on HOPG. © 2009 American Institute of Physics.
Resumo:
Ultraviolet-visible spectroscopy readily discerns the two types of melanin pigments (eumelanin and pheomelanin), although fundamental details regarding the optical properties and pigment heterogeneity are more difficult to disentangle via analysis of the broad featureless absorption spectrum alone. We employed nonlinear transient absorption spectroscopy to study different melanin pigments at near-infrared wavelengths. Excited-state absorption, ground-state depletion, and stimulated emission signal contributions were distinguished for natural and synthetic eumelanins and pheomelanins. A starker contrast among the pigments is observed in the nonlinear excitation regime because they all exhibit distinct transient absorptive amplitudes, phase shifts, and nonexponential population dynamics spanning the femtosecond-nanosecond range. In this manner, different pigments within the pheomelanin subclass were distinguished in synthetic and human hair samples. These results highlight the potential of nonlinear spectroscopies to deliver an in situ analysis of natural melanins in tissue that are otherwise difficult to extract and purify.
Resumo:
Metals support surface plasmons at optical wavelengths and have the ability to localize light to subwavelength regions. The field enhancements that occur in these regions set the ultimate limitations on a wide range of nonlinear and quantum optical phenomena. We found that the dominant limiting factor is not the resistive loss of the metal, but rather the intrinsic nonlocality of its dielectric response. A semiclassical model of the electronic response of a metal places strict bounds on the ultimate field enhancement. To demonstrate the accuracy of this model, we studied optical scattering from gold nanoparticles spaced a few angstroms from a gold film. The bounds derived from the models and experiments impose limitations on all nanophotonic systems.
Resumo:
Although many feature selection methods for classification have been developed, there is a need to identify genes in high-dimensional data with censored survival outcomes. Traditional methods for gene selection in classification problems have several drawbacks. First, the majority of the gene selection approaches for classification are single-gene based. Second, many of the gene selection procedures are not embedded within the algorithm itself. The technique of random forests has been found to perform well in high-dimensional data settings with survival outcomes. It also has an embedded feature to identify variables of importance. Therefore, it is an ideal candidate for gene selection in high-dimensional data with survival outcomes. In this paper, we develop a novel method based on the random forests to identify a set of prognostic genes. We compare our method with several machine learning methods and various node split criteria using several real data sets. Our method performed well in both simulations and real data analysis.Additionally, we have shown the advantages of our approach over single-gene-based approaches. Our method incorporates multivariate correlations in microarray data for survival outcomes. The described method allows us to better utilize the information available from microarray data with survival outcomes.
Resumo:
A regularized algorithm for the recovery of band-limited signals from noisy data is described. The regularization is characterized by a single parameter. Iterative and non-iterative implementations of the algorithm are shown to have useful properties, the former offering the advantage of flexibility and the latter a potential for rapid data processing. Comparative results, using experimental data obtained in laser anemometry studies with a photon correlator, are presented both with and without regularization. © 1983 Taylor & Francis Ltd.
Resumo:
The tomography problem is investigated when the available projections are restricted to a limited angular domain. It is shown that a previous algorithm proposed for extrapolating the data to the missing cone in Fourier space is unstable in the presence of noise because of the ill-posedness of the problem. A regularized algorithm is proposed, which converges to stable solutions. The efficiency of both algorithms is tested by means of numerical simulations. © 1983 Taylor and Francis Group, LLC.