8 resultados para Risk and performance

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The present work summarizes research related to the definition of nutrient recommendations for feeds used in the intensive production of rabbit's meat. Fibre is the main chemical constituent of rabbit diets that typically contain 320 to 360 and 50 to 90 g/kg of insoluble and soluble fibre, respectively. Instead, the dietary contents of cereal grains (∼120 to 160 g/kg), fat (15 to 25 g/kg) and protein concentrates (150 to 180 g/kg) are usually low with respect to other intensively reared monogastric animals. Cell wall constituents are not well digested in rabbits, but this effect is compensated by its stimulus of gut motility, which leads to an increasing rate of passage of digesta, and allows achieving an elevated dry matter intake. A high feed consumption and an adequate balance in essential nutrients are required to sustain the elevated needs of high-productive rabbits measured either as reproductive yield, milk production or growth rate in the fattening period. Around weaning, pathologies occur in a context of incomplete development of the digestive physiology of young rabbits. The supply of balanced diets has also been related to the prevention of disorders by means of three mechanisms: (i) promoting a lower retention time of the digesta in the digestive tract through feeding fibre sources with optimal chemical and physical characteristics, (ii) restricting feed intake after weaning or (iii) causing a lower flow of easily available substrates into the fermentative area by modifying feed composition (e.g. by lowering protein and starch contents, increasing its digestibility or partially substituting insoluble with soluble fibre), or by delaying age at weaning. The alteration in the gut microbiota composition has been postulated as the possible primary cause of these pathologies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A trial was conducted to examine the effects of increasing levels of wheat in the diet and xylanase (ES) supplementation on nitrogen and ether extract retention, pH of the GIT, productive performance from 25 to 47 wks of age, and enzyme activity at the small intestine level. The basal diets (from 25 to 33 wks and from 33 to 47 wks) consisted of soybean meal and corn, and the wheat was introduced in the experimental diets at expenses of corn, primarily.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Although several profiling techniques for identifying performance bottlenecks in logic programs have been developed, they are generally not automatic and in most cases they do not provide enough information for identifying the root causes of such bottlenecks. This complicates using their results for guiding performance improvement. We present a profiling method and tool that provides such explanations. Our profiler associates cost centers to certain program elements and can measure different types of resource-related properties that affect performance, preserving the precedence of cost centers in the cali graph. It includes an automatic method for detecting procedures that are performance bottlenecks. The profiling tool has been integrated in a previously developed run-time checking framework to allow verification of certain properties when they cannot be verified statically. The approach allows checking global computational properties which require complex instrumentation tracking information about previous execution states, such as, e.g., that the execution time accumulated by a given procedure is not greater than a given bound. We have built a prototype implementation, integrated it in the Ciao/CiaoPP system and successfully applied it to performance improvement, automatic optimization (e.g., resource-aware specialization of programs), run-time checking, and debugging of global computational properties (e.g., resource usage) in Prolog programs.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Although several profiling techniques for identifying performance bottlenecks in logic programs have been developed, they are generally not automatic and in most cases they do not provide enough information for identifying the root causes of such bottlenecks. This complicates using their results for guiding performance improvement. We present a profiling method and tool that provides such explanations. Our profiler associates cost centers to certain program elements and can measure different types of resource-related properties that affect performance, preserving the precedence of cost centers in the call graph. It includes an automatic method for detecting procedures that are performance bottlenecks. The profiling tool has been integrated in a previously developed run-time checking framework to allow verification of certain properties when they cannot be verified statically. The approach allows checking global computational properties which require complex instrumentation tracking information about previous execution states, such as, e.g., that the execution time accumulated by a given procedure is not greater than a given bound. We have built a prototype implementation, integrated it in the Ciao/CiaoPP system and successfully applied it to performance improvement, automatic optimization (e.g., resource-aware specialization of programs), run-time checking, and debugging of global computational properties (e.g., resource usage) in Prolog programs.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Most CPV systems are based on Fresnel lenses. Among these, LPI-patented Fresnel-Köhler (FK) concentrator outstands owing to performance and practical reasons. The VentanaTM power train is the first off-the-shelf commercial product based on the FK and comprises both the primary (POE) lenses (a 36-units 1×1 m2 acrylic panel manufactured by EVONIK and 10×) and glass (or Savosil) secondary optics (SOE). This high concentration optical train (Cg=1,024×, ~250mm optical depth) fits with 5×5 mm2 (at least) solar cells. The optical train is the fruit of a 1-year development that has included design, modeling, prototyping and characterization, and through the process LPI had the opportunity to find out how well the actual performance correlates with models, but also learned practical aspects of a CPV system of this kind, some of which have very positive impact on system performance and reliability.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The first step in order to comply with the European Union goals of Near to Zero Energy Buildings is to reduce the energy consumption in buildings. Most of the building consumption is related to the use of active systems to maintain the interior comfort. Passive design strategies contribute to improve the interior comfort conditions, increasing the energy efficiency in buildings and reducing their energy consumption. In this work, an analysis of the passive strategies used in Net Energy Plus Houses has been made. The participating houses of the Solar Decathlon Europe 2012 competition were used as case studies. The passive design strategies of these houses were compared with the annual simulations, and the competition monitored data, especially during the Passive Monitored Period. The analysis included the thermal properties of the building envelope, geometric parameters, ratios and others passive solutions such as Thermal Energy Storage systems, evaporative cooling, night ventilation, solar gains and night sky radiation cooling. The results reflect the impact of passive design strategies on the houses' comfort and efficiency, as well as their influence in helping to achieve the Zero Energy Buildings category.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La importancia de los sistemas de recomendación ha experimentado un crecimiento exponencial como consecuencia del auge de las redes sociales. En esta tesis doctoral presentaré una amplia visión sobre el estado del arte de los sistemas de recomendación. Incialmente, estos estaba basados en fitrado demográfico, basado en contendio o colaborativo. En la actualidad, estos sistemas incorporan alguna información social al proceso de recomendación. En el futuro utilizarán información implicita, local y personal proveniente del Internet de las cosas. Los sistemas de recomendación basados en filtrado colaborativo se pueden modificar con el fin de realizar recomendaciones a grupos de usuarios. Existen trabajos previos que han incluido estas modificaciones en diferentes etapas del algoritmo de filtrado colaborativo: búsqueda de los vecinos, predicción de las votaciones y elección de las recomendaciones. En esta tesis doctoral proporcionaré un nuevo método que realizar el proceso de unficación (pasar de varios usuarios a un grupo) en el primer paso del algoritmo de filtrado colaborativo: cálculo de la métrica de similaridad. Proporcionaré una formalización completa del método propuesto. Explicaré cómo obtener el conjunto de k vecinos del grupo de usuarios y mostraré cómo obtener recomendaciones usando dichos vecinos. Asimismo, incluiré un ejemplo detallando cada paso del método propuesto en un sistema de recomendación compuesto por 8 usuarios y 10 items. Las principales características del método propuesto son: (a) es más rápido (más eficiente) que las alternativas proporcionadas por otros autores, y (b) es al menos tan exacto y preciso como otras soluciones estudiadas. Para contrastar esta hipótesis realizaré varios experimentos que miden la precisión, la exactitud y el rendimiento del método. Los resultados obtenidos se compararán con los resultados de otras alternativas utilizadas en la recomendación de grupos. Los experimentos se realizarán con las bases de datos de MovieLens y Netflix. ABSTRACT The importance of recommender systems has grown exponentially with the advent of social networks. In this PhD thesis I will provide a wide vision about the state of the art of recommender systems. They were initially based on demographic, contentbased and collaborative filtering. Currently, these systems incorporate some social information to the recommendation process. In the future, they will use implicit, local and personal information from the Internet of Things. As we will see here, recommender systems based on collaborative filtering can be used to perform recommendations to group of users. Previous works have made this modification in different stages of the collaborative filtering algorithm: establishing the neighborhood, prediction phase and determination of recommended items. In this PhD thesis I will provide a new method that carry out the unification process (many users to one group) in the first stage of the collaborative filtering algorithm: similarity metric computation. I will provide a full formalization of the proposed method. I will explain how to obtain the k nearest neighbors of the group of users and I will show how to get recommendations using those users. I will also include a running example of a recommender system with 8 users and 10 items detailing all the steps of the method I will present. The main highlights of the proposed method are: (a) it will be faster (more efficient) that the alternatives provided by other authors, and (b) it will be at least as precise and accurate as other studied solutions. To check this hypothesis I will conduct several experiments measuring the accuracy, the precision and the performance of my method. I will compare these results with the results generated by other methods of group recommendation. The experiments will be carried out using MovieLens and Netflix datasets.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

One of the most demanding needs in cloud computing and big data is that of having scalable and highly available databases. One of the ways to attend these needs is to leverage the scalable replication techniques developed in the last decade. These techniques allow increasing both the availability and scalability of databases. Many replication protocols have been proposed during the last decade. The main research challenge was how to scale under the eager replication model, the one that provides consistency across replicas. This thesis provides an in depth study of three eager database replication systems based on relational systems: Middle-R, C-JDBC and MySQL Cluster and three systems based on In-Memory Data Grids: JBoss Data Grid, Oracle Coherence and Terracotta Ehcache. Thesis explore these systems based on their architecture, replication protocols, fault tolerance and various other functionalities. It also provides experimental analysis of these systems using state-of-the art benchmarks: TPC-C and TPC-W (for relational systems) and Yahoo! Cloud Serving Benchmark (In- Memory Data Grids). Thesis also discusses three Graph Databases, Neo4j, Titan and Sparksee based on their architecture and transactional capabilities and highlights the weaker transactional consistencies provided by these systems. It discusses an implementation of snapshot isolation in Neo4j graph database to provide stronger isolation guarantees for transactions.