471 resultados para Component reuse


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hydrogels are promising materials for cartilage repair, but the properties required for optimal functional outcomes are not yet known. In this study, we functionalized four materials that are commonly used in cartilage tissue engineering and evaluated them using in vitro cultures. Gelatin, hyaluronic acid, polyethylene glycol, and alginate were functionalized with methacrylic anhydride to make them photocrosslinkable. We found that the responses of encapsulated human chondrocytes were highly dependent on hydrogel type. Gelatin hydrogels supported cell proliferation and the deposition of a glycosaminoglycan rich matrix with significant mechanical functionality. However, cells had a dedifferentiated phenotype, with high expression of collagen type I. Chondrocytes showed the best redifferentiation in hyaluronic acid hydrogels, but the newly formed matrix was highly localized to the pericellular regions, and these gels degraded rapidly. Polyethylene glycol hydrogels, as a bioinert control, did not promote any strong responses. Alginate hydrogels did not support the deposition of new matrix, and the stiffness decreased during culture. The markedly different response of chondrocytes to these four photocrosslinkable hydrogels demonstrates the importance of material properties for chondrogenesis and extracellular matrix production, which are critical for effective cartilage repair.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The mean shift tracker has achieved great success in visual object tracking due to its efficiency being nonparametric. However, it is still difficult for the tracker to handle scale changes of the object. In this paper, we associate a scale adaptive approach with the mean shift tracker. Firstly, the target in the current frame is located by the mean shift tracker. Then, a feature point matching procedure is employed to get the matched pairs of the feature point between target regions in the current frame and the previous frame. We employ FAST-9 corner detector and HOG descriptor for the feature matching. Finally, with the acquired matched pairs of the feature point, the affine transformation between target regions in the two frames is solved to obtain the current scale of the target. Experimental results show that the proposed tracker gives satisfying results when the scale of the target changes, with a good performance of efficiency.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Bearing faults are the most common cause of wind turbine failures. Unavailability and maintenance cost of wind turbines are becoming critically important, with their fast growing in electric networks. Early fault detection can reduce outage time and costs. This paper proposes Anomaly Detection (AD) machine learning algorithms for fault diagnosis of wind turbine bearings. The application of this method on a real data set was conducted and is presented in this paper. For validation and comparison purposes, a set of baseline results are produced using the popular one-class SVM methods to examine the ability of the proposed technique in detecting incipient faults.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Overvoltage and overloading due to high utilization of PVs are the main power quality concerns for future distribution power systems. This paper proposes a distributed control coordination strategy to manage multiple PVs within a network to overcome these issues. PVs reactive power is used to deal with over-voltages and PVs active power curtailment are regulated to avoid overloading. The proposed control structure is used to share the required contribution fairly among PVs, in proportion to their ratings. This approach is examined on a practical distribution network with multiple PVs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Understanding the dynamics of disease spread is of crucial importance, in contexts such as estimating load on medical services to risk assessment and intervention policies against large-scale epidemic outbreaks. However, most of the information is available after the spread itself, and preemptive assessment is far from trivial. Here, we investigate the use of agent-based simulations to model such outbreaks in a stylised urban environment. For most diseases, infection of a new individual may occur from casual contact in crowds as well as from repeated interactions with social partners such as work colleagues or family members. Our model therefore accounts for these two phenomena.Presented in this paper is the initial framework for such a model, detailing implementation of geographical features and generation of social structures. Preliminary results are a promising step towards large-scale simulations and evaluation of potential intervention policies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Several algorithms and techniques widely used in Computer Science have been adapted from, or inspired by, known biological phenomena. This is a consequence of the multidisciplinary background of most early computer scientists. The field has now matured, and permits development of tools and collaborative frameworks which play a vital role in advancing current biomedical research. In this paper, we briefly present examples of the former, and elaborate upon two of the latter, applied to immunological modelling and as a new paradigm in gene expression.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we investigate the effect of mobility constraints on epidemic broadcast mechanisms in DTNs (Delay-Tolerant Networks). Major factors affecting epidemic broadcast performances are its forwarding algorithm and node mobility. The impact of forwarding algorithm and node mobility on epidemic broadcast mechanisms has been actively studied in the literature, but those studies generally use unconstrained mobility models. The objective of this paper is therefore to quantitatively investigate the effect of mobility constraints on epidemic broadcast mechanisms. We evaluate the performances of three classes of epidemic broadcast mechanisms - P-BCAST (PUSH-based BroadCast), SA-BCAST (Self-Adaptive BroadCast), and HP-BCAST (History-based P-BCAST) - with a random waypoint mobility model with mobility constraints. Our finding includes that the existence of mobility constraints significantly improves the reach ability and dissemination speed of epidemic broadcast mechanisms while degrading their efficiency.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In some delay-tolerant communication systems such as vehicular ad-hoc networks, information flow can be represented as an infectious process, where each entity having already received the information will try to share it with its neighbours. The random walk and random waypoint models are popular analysis tools for these epidemic broadcasts, and represent two types of random mobility. In this paper, we introduce a simulation framework investigating the impact of a gradual increase of bias in path selection (i.e. reduction of randomness), when moving from the former to the latter. Randomness in path selection can significantly alter the system performances, in both regular and irregular network structures. The implications of these results for real systems are discussed in details.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

One of the main challenges in data analytics is that discovering structures and patterns in complex datasets is a computer-intensive task. Recent advances in high-performance computing provide part of the solution. Multicore systems are now more affordable and more accessible. In this paper, we investigate how this can be used to develop more advanced methods for data analytics. We focus on two specific areas: model-driven analysis and data mining using optimisation techniques.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

As computational models in fields such as medicine and engineering get more refined, resource requirements are increased. In a first instance, these needs have been satisfied using parallel computing and HPC clusters. However, such systems are often costly and lack flexibility. HPC users are therefore tempted to move to elastic HPC using cloud services. One difficulty in making this transition is that HPC and cloud systems are different, and performance may vary. The purpose of this study is to evaluate cloud services as a means to minimise both cost and computation time for large-scale simulations, and to identify which system properties have the most significant impact on performance. Our simulation results show that, while the performance of Virtual CPU (VCPU) is satisfactory, network throughput may lead to difficulties.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Most real-life data analysis problems are difficult to solve using exact methods, due to the size of the datasets and the nature of the underlying mechanisms of the system under investigation. As datasets grow even larger, finding the balance between the quality of the approximation and the computing time of the heuristic becomes non-trivial. One solution is to consider parallel methods, and to use the increased computational power to perform a deeper exploration of the solution space in a similar time. It is, however, difficult to estimate a priori whether parallelisation will provide the expected improvement. In this paper we consider a well-known method, genetic algorithms, and evaluate on two distinct problem types the behaviour of the classic and parallel implementations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In providing simultaneous information on expression profiles for thousands of genes, microarray technologies have, in recent years, been largely used to investigate mechanisms of gene expression. Clustering and classification of such data can, indeed, highlight patterns and provide insight on biological processes. A common approach is to consider genes and samples of microarray datasets as nodes in a bipartite graphs, where edges are weighted e.g. based on the expression levels. In this paper, using a previously-evaluated weighting scheme, we focus on search algorithms and evaluate, in the context of biclustering, several variations of Genetic Algorithms. We also introduce a new heuristic “Propagate”, which consists in recursively evaluating neighbour solutions with one more or one less active conditions. The results obtained on three well-known datasets show that, for a given weighting scheme,optimal or near-optimal solutions can be identified.