52 resultados para Structured mealtimes


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Grounded on Raymond Williams‘s definition of knowable community as a cultural tool to analyse literary texts, the essay reads the texts D.H.Lawrence wrote while travelling in the Mediterranean (Twilight in Italy, Sea and Sardinia and Etruscan Places) as knowable communities, bringing to the discussion the wide importance of literature not only as an object for aesthetic or textual readings, but also as a signifying practice which tells stories of culture. Departing from some considerations regarding the historical development of the relationship between literature and culture, the essay analyses the ways D. H. Lawrence constructed maps of meaning, where the readers, in a dynamic relation with the texts, apprehend experiences, structures and feelings; putting into perspective Williams‘s theory of culture as a whole way of life, it also analyses the ways the author communicates and organizes these experiences, creating a space of communication and operating at different levels of reality: on the one hand, the reality of the whole way of Italian life, and, on the other hand, the reality of the reader who aspires to make sense and to create an interpretative context where all the information is put, and, also, the reality of the writer in the poetic act of writing. To read these travel writings as knowable communities is to understand them as a form that invents a community with no other existence but that of the literary text. The cultural construction we find in these texts is the result of the selection, and interpretation done by D.H.Lawrence, as well as the product of the author‘s enunciative positions, and of his epistemological and ontological filigrees of existence, structured by the conditions of possibility. In the rearticulation of the text, of the writer and of the reader, in a dynamic and shared process of discursive alliances, we understand that Lawrence tells stories of the Mediterranean through his literary art.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Os direitos de transmissão (Transmission Rigths – TRs) correspondem, na sua essência, a contratos que conferem aos seus proprietários o direito de transmitir energia eléctrica, por um determinado caminho, a um preço fixo. Na actualidade vive-se uma era de liberalização dos mercados de energia eléctrica nos quais, no caso concreto dos mercados de TRs, os diversos agentes podem ceder os TRs que possuem a outros agentes desde que cumpram certos requisitos impostos pelo sistema na figura do seu operador de sistema (Independent System Operator – ISO). Neste sentido, o ISO oferece, aos diversos agentes do mercado, algumas ferramentas que lhes permite transaccionar, sob sua orientação, os seus respectivos TRs fazendo-se cumprir todos os requisitos indispensáveis para o efeito. A mais popular dessas ferramentas, nos principais mercados energético da actualidade, é o leilão. Com o presente trabalho de dissertação é pretendido apresentar-se um modelo para a resolução do problema inerente a um leilão de TRs em ambiente de mercado, neste caso concreto de direitos de transmissão financeiros de energia eléctrica (Financial Transmission Rigths – FTRs). Neste sentido foi desenvolvido um simulador informático (SIM_AuctFTR) que implementa um modelo para este tipo de problemas. Este trabalho foi estruturado essencialmente em três etapas com objectivos inerentes. Assim, numa primeira fase da realização deste trabalho, foi realizado um estudo de diversos conceitos e metodologias inerentes ao problema de leilão de FTRs em ambiente de mercado, suportado por uma sólida base bibliográfica. A segunda fase tratou-se do desenvolvimento do algoritmo da aplicação computacional que solucione um problema de um leilão de FTRs. O SIM_AuctFTR foi desenvolvido sobre o pressuposto da maximização dos rendimentos financeiros provenientes da atribuição dos FTRs propostos a leilão, tendo estes últimos de coabitar no sistema respeitando as restrições técnicas a que este são inerentes, mesmo na eventualidade da ocorrência de contingências de nível ‘n-1’. Por último, numa terceira fase, partindo da aplicação desenvolvida, foram realizados 4 casos de estudos com outras tantas redes eléctricas de forma a testar a robustez da ferramenta desenvolvida.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este estudo tem como objetivo compreender a fase de preparação da expatriação de empresas portuguesas para Moçambique. Mais concretamente, procura explorar as práticas de Gestão de Recursos Humanos desenvolvidas durante a fase de preparação de missões internacionais para Moçambique. Para a realização do estudo foram selecionadas 5 empresas portuguesas que têm experiência de expatriação de colaboradores para Moçambique. Foi utilizada a metodologia qualitativa, optando-se pelo estudo caso múltiplo e exploratório. O instrumento de recolha de informação foi a entrevista semiestruturada. Ao todo, foram realizadas 12 entrevistas, das quais 5 destinaram-se a representantes organizacionais e 7 a expatriados/repatriados das 5 empresas em estudo. Os resultados sugerem que as empresas portuguesas preparam a expatriação para Moçambique. Porém, essa preparação é feita de forma informal, não existindo ainda um programa formalizado de preparação de expatriações, nomeadamente para Moçambique. Na parte final da dissertação são discutidos os principais resultados obtidos e apresentadas as conclusões do estudo aqui levado a cabo.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we discuss the mathematical aspects of the Heisenberg uncertainty principle within local fractional Fourier analysis. The Schrödinger equation and Heisenberg uncertainty principles are structured within local fractional operators.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper proposes a new architecture targeting real-time and reliable Distributed Computer-Controlled Systems (DCCS). This architecture provides a structured approach for the integration of soft and/or hard real-time applications with Commercial O -The-Shelf (COTS) components. The Timely Computing Base model is used as the reference model to deal with the heterogeneity of system components with respect to guaranteeing the timeliness of applications. The reliability and availability requirements of hard real-time applications are guaranteed by a software-based fault-tolerance approach.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper describes how MPEG-4 object based video (obv) can be used to allow selected objects to be inserted into the play-out stream to a specific user based on a profile derived for that user. The application scenario described here is for personalized product placement, and considers the value of this application in the current and evolving commercial media distribution market given the huge emphasis media distributors are currently placing on targeted advertising. This level of application of video content requires a sophisticated content description and metadata system (e.g., MPEG-7). The scenario considers the requirement for global libraries to provide the objects to be inserted into the streams. The paper then considers the commercial trading of objects between the libraries, video service providers, advertising agencies and other parties involved in the service. Consequently a brokerage of video objects is proposed based on negotiation and trading using intelligent agents representing the various parties. The proposed Media Brokerage Platform is a multi-agent system structured in two layers. In the top layer, there is a collection of coarse grain agents representing the real world players – the providers and deliverers of media contents and the market regulator profiler – and, in the bottom layer, there is a set of finer grain agents constituting the marketplace – the delegate agents and the market agent. For knowledge representation (domain, strategic and negotiation protocols) we propose a Semantic Web approach based on ontologies. The media components contents should be represented in MPEG-7 and the metadata describing the objects to be traded should follow a specific ontology. The top layer content providers and deliverers are modelled by intelligent autonomous agents that express their will to transact – buy or sell – media components by registering at a service registry. The market regulator profiler creates, according to the selected profile, a market agent, which, in turn, checks the service registry for potential trading partners for a given component and invites them for the marketplace. The subsequent negotiation and actual transaction is performed by delegate agents in accordance with their profiles and the predefined rules of the market.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Embedded real-time applications increasingly present high computation requirements, which need to be completed within specific deadlines, but that present highly variable patterns, depending on the set of data available in a determined instant. The current trend to provide parallel processing in the embedded domain allows providing higher processing power; however, it does not address the variability in the processing pattern. Dimensioning each device for its worst-case scenario implies lower average utilization, and increased available, but unusable, processing in the overall system. A solution for this problem is to extend the parallel execution of the applications, allowing networked nodes to distribute the workload, on peak situations, to neighbour nodes. In this context, this report proposes a framework to develop parallel and distributed real-time embedded applications, transparently using OpenMP and Message Passing Interface (MPI), within a programming model based on OpenMP. The technical report also devises an integrated timing model, which enables the structured reasoning on the timing behaviour of these hybrid architectures.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Mestrado em Ensino Precoce do Inglês

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Not just with the emergence but also with the growing of the electronic market, that is, the growth of online suppliers of services and products and Internet users (potential consumers), the necessary conditions to the affirmation of the agile/virtual enterprises (A/VE) as a present and future enterprise organizational model are created. In this context, it is our understanding that the broker may have an important role in its development, namely, if the broker performs functions for the A/VE with better efficacy and efficiency. In this article, we will present first a revision of the broker’s models in a structured form. We present a taxonomy of possible broker’s functions for the broker’s actuation near the A/VE and then the classification of the literature broker’s models. This classification will permit an analysis of a broker’s model and establish a mainframe for our broker’s model according to the BM_Virtual Enterprise Architecture Reference Model (BM_VEARM).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Most of the traditional software and database development approaches tend to be serial, not evolutionary and certainly not agile, especially on data-oriented aspects. Most of the more commonly used methodologies are strict, meaning they’re composed by several stages each with very specific associated tasks. A clear example is the Rational Unified Process (RUP), divided into Business Modeling, Requirements, Analysis & Design, Implementation, Testing and Deployment. But what happens when the needs of a well design and structured plan, meet the reality of a small starting company that aims to build an entire user experience solution. Here resource control and time productivity is vital, requirements are in constant change, and so is the product itself. In order to succeed in this environment a highly collaborative and evolutionary development approach is mandatory. The implications of constant changing requirements imply an iterative development process. Project focus is on Data Warehouse development and business modeling. This area is usually a tricky one. Business knowledge is part of the enterprise, how they work, their goals, what is relevant for analyses are internal business processes. Throughout this document it will be explained why Agile Modeling development was chosen. How an iterative and evolutionary methodology, allowed for reasonable planning and documentation while permitting development flexibility, from idea to product. More importantly how it was applied on the development of a Retail Focused Data Warehouse. A productized Data Warehouse built on the knowledge of not one but several client needs. One that aims not just to store usual business areas but create an innovative sets of business metrics by joining them with store environment analysis, converting Business Intelligence into Actionable Business Intelligence.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Dissertação apresentada ao Instituto Superior de Contabilidade e Administração do Porto para a obtenção do Grau de Mestre em Empreendedorismo e Internacionalização Orientador: Professor Doutor José de Freitas Santos

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper shows that a hierarchical architecture, distributing several control actions in growing levels of complexity and using resources of reconfigurable computing, enables one to take into account the ease of future modifications, updates and improvements in robotic applications. An experimental example of a Stewart—Gough platform control (a platform applied as the solution to countless practical problems) is presented using reconfigurable computing. The software and hardware developed are structured in independent blocks. This open architecture implementation allows easy expansion of the system and better adaptation of the platform to its related tasks.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Dissertação apresentada ao Instituto Politécnico do Porto para obtenção do Grau de Mestre em Gestão das Organizações, Ramo de Gestão de Empresas Orientada pelo Professor Doutor José Freitas Santos

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The dynamic of the international business and its multidimensional nature requires the understanding of the complexities of different contexts dictated by cultural differences between countries. The purpose of this paper is to study, in depth howsmall and medium-sized companies develop their international marketing mix strategy in their overseas subsidiaries. We use the theoretical construct of Hofstede (1980, 1991) in the dimensions of Power Distance (PD), Uncertainty Avoidance (UA), Individualism (IND), Masculinity (MASC) and Long-Term Orientation (LTO) to assess the cross cultural differences between countries and the business practices to analyze the adaptation or standardization of the international marketing mix strategy of foreign Portuguese subsidiaries. Ourstudy uses an exploratoryand qualitative methodology. We conducted semi-structured interviews in order to achieve a good understanding ofinternational marketing mix strategy of four companies from different sectors. Our results show that the national cultural differences have great influence in the marketing strategy of the subsidiary. The business practices adjustments in the subsidiary company that proved to be necessary conditions for their performance are conducted by the products augmented offerings concerning the characteristics of the product, design and brand name in order to meet the requirements and specificities of the host country of the subsidiary.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The main purpose of this work was the development of procedures for the simulation of atmospheric ows over complex terrain, using OpenFOAM. For this aim, tools and procedures were developed apart from this code for the preprocessing and data extraction, which were thereafter applied in the simulation of a real case. For the generation of the computational domain, a systematic method able to translate the terrain elevation model to a native OpenFOAM format (blockMeshDict) was developed. The outcome was a structured mesh, in which the user has the ability to de ne the number of control volumes and its dimensions. With this procedure, the di culties of case set up and the high computation computational e ort reported in literature associated to the use of snappyHexMesh, the OpenFOAM resource explored until then for the accomplishment of this task, were considered to be overwhelmed. Developed procedures for the generation of boundary conditions allowed for the automatic creation of idealized inlet vertical pro les, de nition of wall functions boundary conditions and the calculation of internal eld rst guesses for the iterative solution process, having as input experimental data supplied by the user. The applicability of the generated boundary conditions was limited to the simulation of turbulent, steady-state, incompressible and neutrally strati ed atmospheric ows, always recurring to RaNS (Reynolds-averaged Navier-Stokes) models. For the modelling of terrain roughness, the developed procedure allowed to the user the de nition of idealized conditions, like an uniform aerodynamic roughness length or making its value variable as a function of topography characteristic values, or the using of real site data, and it was complemented by the development of techniques for the visual inspection of generated roughness maps. The absence and the non inclusion of a forest canopy model limited the applicability of this procedure to low aerodynamic roughness lengths. The developed tools and procedures were then applied in the simulation of a neutrally strati ed atmospheric ow over the Askervein hill. In the performed simulations was evaluated the solution sensibility to di erent convection schemes, mesh dimensions, ground roughness and formulations of the k - ε and k - ω models. When compared to experimental data, calculated values showed a good agreement of speed-up in hill top and lee side, with a relative error of less than 10% at a height of 10 m above ground level. Turbulent kinetic energy was considered to be well simulated in the hill windward and hill top, and grossly predicted in the lee side, where a zone of ow separation was also identi ed. Despite the need of more work to evaluate the importance of the downstream recirculation zone in the quality of gathered results, the agreement between the calculated and experimental values and the OpenFOAM sensibility to the tested parameters were considered to be generally in line with the simulations presented in the reviewed bibliographic sources.