19 resultados para Continuous Attention
em Instituto Politécnico do Porto, Portugal
Resumo:
Lifelong learning (LLL) has received increasing attention in recent years. It implies that learning should take place at all stages of the “life cycle and it should be life-wide, that is embedded in all life contexts from the school to the work place, the home and the community” (Green, 2002, p.613). The ‘learning society’, is the vision of a society where there are recognized opportunities for learning for every person, wherever they are and however old they happen to be. Globalization and the rise of new information technologies are some of the driving forces that cause depreciation of specialised competences. This happens very quickly in terms of economic value; consequently, workers of all skills levels, during their working life, must have the opportunity to update “their technical skills and enhance general skills to keep pace with continuous technological change and new job requirements” (Fahr, 2005, p. 75). It is in this context that LLL tops the policy agenda of international bodies, national governments and non-governmental organizations, in the field of education and training, to justify the need for LLL opportunities for the population as they face contemporary employability challenges. It is in this context that the requirement and interest to analyse the behaviour patterns of adult learners has developed over the last few years
Resumo:
IRMA International Conference under the theme Managing Worldwide Operations and Communications with Information Technology, May 19-23, Vancouver, British Columbia, Canada
Resumo:
Trabalho de Projeto apresentado ao Instituto Superior de Contabilidade e Administração do Porto para obtenção do grau de Mestre em Auditoria Orientado por: Doutora Alcina Augusta de Sena Portugal Dias
Resumo:
Nos dias de hoje é necessário criar hábitos de vida mais saudáveis que contribuam para o bem-estar da população. Adoptar medidas e práticas de modo regular e disciplinado, pode diminuir o risco do aparecimento de determinadas doenças, como a obesidade, as doenças cardiovasculares, a hipertensão, a diabetes, alguns tipos de cancro e tantas outras. É também importante salientar que, uma alimentação cuidada dá saúde e aumenta a esperança média de vida. Em Portugal, nos últimos anos, os costumes alimentares da população têm vindo a alterar-se significativamente. As refeições caseiras confeccionadas com produtos frescos dão lugar à designada “cultura do fast food”. Em contrapartida, os consumidores são cada vez mais exigentes, estando em permanente alerta no que se refere ao estado dos alimentos. A rotulagem de um produto, para além da função publicitária, tem vindo a ser objecto de legislação específica de forma a fornecer informação simples e clara, correspondente à composição, qualidade, quantidade, validade ou outras características do produto. Estas informações devem ser acessíveis a qualquer tipo de público, com mais ou menos formação e de qualquer estrato social. A qualidade e segurança dos produtos deve basear-se na garantia de que todos os ingredientes, materiais de embalagem e processos produtivos são adequados à produção de produtos seguros, saudáveis e saborosos. A Silliker Portugal, S.A. é uma empresa independente de prestação de serviços para o sector agro-alimentar, líder mundial na prestação de serviços para a melhoria da qualidade e segurança alimentar. A Silliker dedica-se a ajudar as empresas a encontrar soluções para os desafios actuais do sector, oferecendo uma ampla gama de serviços, onde se inclui o serviço de análises microbiológicas, químicas e sensorial; consultadoria em segurança alimentar e desenvolvimento; auditorias; rotulagem e legislação. A actualização permanente de procedimentos na procura de uma melhoria contínua é um dos objectivos da empresa. Para responder a um dos desafios colocados à Silliker, surgiu este trabalho, que consistiu no desenvolvimento de um novo método para determinação de ácidos gordos e da gordura total em diferentes tipos de alimentos e comparação dos resultados, com os obtidos com o método analítico até então adoptado. Se a gordura é um elemento de grande importância na alimentação, devido às suas propriedades nutricionais e organoléticas, recentemente, os investigadores têm focado a sua atenção nos mais diversos ácidos gordos (saturados, monoinsaturados e polinsaturados), em particular nos ácidos gordos essenciais e nos isómeros do ácido linoleico conjugado (CLA), uma mistura de isómeros posicionais e geométricos do ácido linoleico com actividade biológica importante. A técnica usada nas determinações foi a cromatografia gasosa com ionização de chama, GC-FID, tendo as amostras sido previamente tratadas e extraídas de acordo com o tipo de matriz. A metodologia analítica desenvolvida permitiu a correcta avaliação do perfil em ácidos gordos, tendo-se para isso usado uma mistura de 37 ésteres metílicos, em que o ácido gordo C13:0 foi usado como padrão interno. A identificação baseou-se nos tempos de retenção de cada ácido gordo da mistura e para a quantificação usaram-se os factores de resposta. A validação do método implementado foi baseada nos resultados obtidos no estudo de três matrizes relativas a materiais certificados pela BIPEA (Bureau Interprofessionnel des Etudes Analytiques), para o que foram efectuadas doze réplicas de cada matriz. Para cada réplica efectuada foi calculado o teor de matéria gorda, sendo posteriormente o resultado comparado com o emitido pela entidade certificada. Após análise de cada constituinte foi também possível calcular o teor em ácidos gordos saturados, monoinsaturados e polinsaturados. A determinação do perfil em ácidos gordos dos materiais certificados foi aceitável atendendo aos valores obtidos, os quais se encontravam no intervalo de valores admissíveis indicados nos relatórios. A quantificação da matéria gorda no que se refere à matriz de “Paté à Tartinier” apresentou um z-score de 4,3, o que de acordo com as exigências internas da Silliker, não é válido. Para as outras duas matrizes (“Mélange Nutritif” e “Plat cuisiné à base de viande”) os valores de z-score foram, respectivamente, 0,7 e -1,0, o que permite concluir a validade do método. Para que o método possa vir a ser adoptado como método alternativo é necessário um estudo mais alargado relativamente a amostras com diferentes composições. O método foi aplicado na análise de amostras de fiambre, leite gordo, queijo, ovo com ómega 3, amendoim e óleo de girassol, e os resultados foram comparados com os obtidos pelo método até então adoptado.
Resumo:
In this work three natural waste materials containing chitin were used as adsorbents for textile dyestuffs, namely the Anodonta (Anodonta cygnea) shell, the Sepia (Sepia officinalis) and the Squid (Loligo vulgaris) pens. The selected dyestuffs were the Cibacron green T3G-E (CI reactive green 12), and the Solophenyl green BLE 155% (CI direct green 26), both from CIBA, commonly used in cellulosic fibres dyeing, the most used fibres in the textile industry. Batch equilibrium studies showed that the materials’ adsorption capacities increase after a simple and inexpensive chemical treatment, which increases their porosity and chitin relative content. Kinetic studies suggested the existence of a high internal resistance in both systems. Fixed bed column experiments performed showed an improvement in adsorbents’ behaviour after chemical treatment. However, in the column experiments, the biodegradation was the main mechanism of dyestuff removal, allowing the materials’ bioregeneration. The adsorption was strongly reduced by the pore clogging effect of the biomass. The deproteinised Squid pen (grain size 0.500–1.41 mm) is the adsorbent with highest adsorption capacity (0.27 and 0.037 g/g, respectively, for the reactive and direct dyestuffs, at 20ºC), followed by the demineralised Sepia pen and Anodonta shell, behaving like pure chitin in all experiments, but showing inferior performances than the granular activated carbon tested in the column experiments.
Resumo:
We introduce a new wavelet transform within the framework of the local fractional calculus. An illustrative example of local fractional wavelet transform is also presented.
Resumo:
The higher education system in Europe is currently under stress and the debates over its reform and future are gaining momentum. Now that, for most countries, we are in a time for change, in the overall society and the whole education system, the legal and political dimensions have gained prominence, which has not been followed by a more integrative approach of the problem of order, its reform and the issue of regulation, beyond the typical static and classical cost-benefit analyses. The two classical approaches for studying (and for designing the policy measures of) the problem of the reform of the higher education system - the cost-benefit analysis and the legal scholarship description - have to be integrated. This is the argument of our paper that the very integration of economic and legal approaches, what Warren Samuels called the legal-economic nexus, is meaningful and necessary, especially if we want to address the problem of order (as formulated by Joseph Spengler) and the overall regulation of the system. On the one hand, and without neglecting the interest and insights gained from the cost-benefit analysis, or other approaches of value for money assessment, we will focus our study on the legal, social and political aspects of the regulation of the higher education system and its reform in Portugal. On the other hand, the economic and financial problems have to be taken into account, but in a more inclusive way with regard to the indirect and other socio-economic costs not contemplated in traditional or standard assessments of policies for the tertiary education sector. In the first section of the paper, we will discuss the theoretical and conceptual underpinning of our analysis, focusing on the evolutionary approach, the role of critical institutions, the legal-economic nexus and the problem of order. All these elements are related to the institutional tradition, from Veblen and Commons to Spengler and Samuels. The second section states the problem of regulation in the higher education system and the issue of policy formulation for tackling the problem. The current situation is clearly one of crisis with the expansion of the cohorts of young students coming to an end and the recurrent scandals in private institutions. In the last decade, after a protracted period of extension or expansion of the system, i. e., the continuous growth of students, universities and other institutions are competing harder to gain students and have seen their financial situation at risk. It seems that we are entering a period of radical uncertainty, higher competition and a new configuration that is slowly building up is the growth in intensity, which means upgrading the quality of the higher learning and getting more involvement in vocational training and life-long learning. With this change, and along with other deep ones in the Portuguese society and economy, the current regulation has shown signs of maladjustment. The third section consists of our conclusions on the current issue of regulation and policy challenge. First, we underline the importance of an evolutionary approach to a process of change that is essentially dynamic. A special attention will be given to the issues related to an evolutionary construe of policy analysis and formulation. Second, the integration of law and economics, through the notion of legal economic nexus, allows us to better define the issues of regulation and the concrete problems that the universities are facing. One aspect is the instability of the political measures regarding the public administration and on which the higher education system depends financially, legally and institutionally, to say the least. A corollary is the lack of clear strategy in the policy reforms. Third, our research criticizes several studies, such as the one made by the OECD in late 2006 for the Ministry of Science, Technology and Higher Education, for being too static and neglecting fundamental aspects of regulation such as the logic of actors, groups and organizations who are major players in the system. Finally, simply changing the legal rules will not necessary per se change the behaviors that the authorities want to change. By this, we mean that it is not only remiss of the policy maker to ignore some of the critical issues of regulation, namely the continuous non-respect by academic management and administrative bodies of universities of the legal rules that were once promulgated. Changing the rules does not change the problem, especially without the necessary debates form the different relevant quarters that make up the higher education system. The issues of social interaction remain as intact. Our treatment of the matter will be organized in the following way. In the first section, the theoretical principles are developed in order to be able to study more adequately the higher education transformation with a modest evolutionary theory and a legal and economic nexus of the interactions of the system and the policy challenges. After describing, in the second section, the recent evolution and current working of the higher education in Portugal, we will analyze the legal framework and the current regulatory practices and problems in light of the theoretical framework adopted. We will end with some conclusions on the current problems of regulation and the policy measures that are discusses in recent years.
Resumo:
In this paper, the fractional Fourier transform (FrFT) is applied to the spectral bands of two component mixture containing oxfendazole and oxyclozanide to provide the multicomponent quantitative prediction of the related substances. With this aim in mind, the modulus of FrFT spectral bands are processed by the continuous Mexican Hat family of wavelets, being denoted by MEXH-CWT-MOFrFT. Four modulus sets are obtained for the parameter a of the FrFT going from 0.6 up to 0.9 in order to compare their effects upon the spectral and quantitative resolutions. Four linear regression plots for each substance were obtained by measuring the MEXH-CWT-MOFrFT amplitudes in the application of the MEXH family to the modulus of the FrFT. This new combined powerful tool is validated by analyzing the artificial samples of the related drugs, and it is applied to the quality control of the commercial veterinary samples.
Resumo:
Towpregs based on different fibres and thermoplastic matrices were processed for highly demanding and more commercial applications by different composite processing technologies. In the technologies used, compression moulding and pultrusion, the final composite pr ocessing parameters were studied in order to obtain composites with adequate properties at industrial compatible production rates. The produced towpregs were tested to verify its polymer content and degree of impregnation. The obtained results have shown t hat the coating line enabled to produce, with efficiency and industrial scale speed rates, thermoplastic matrix towpregs that may be used to manufacture composites for advanced and larger volume commercial markets.
Resumo:
Manufacturing processes need permanently to innovate and optimize because any can be susceptible to continuous improvement. Innovation and commitment to the development of these new solutions resulting from existing expertise and the continuing need to increase productivity, flexibility and ensuring the necessary quality of the manufactured products. To increase flexibility, it is necessary to significantly reduce set-up times and lead time in order to ensure the delivery of products ever faster. This objective can be achieved through a normalization of the pultrusion line elements. Implicitly, there is an increase of productivity by this way. This work is intended to optimize the pultrusion process of structural profiles. We consider all elements of the system from the storehouse of the fibers (rack) to the pultrusion die. Particular attention was devoted to (a) the guidance system of the fibers and webs, (b) the resin container where the fibers are impregnated, (c) standard plates positioning of the fibers towards the entrance to the spinneret and also (d) reviewed the whole process of assembling and fixing the die as well as its the heating system. With the implementation of these new systems was achieved a significant saving of time set-up and were clearly reduced the unit costs of production. Quality assurance was also increased.
Resumo:
Num passado relativamente recente, o processo de demolição tem experimentado várias mudanças, que incidem principalmente na forma como o edifício ou estrutura vai ser demolido. Mais recentemente, o surgimento das preocupações ambientais levam hoje a falar de "desconstrução", em função do prejuízo das demolições. O termo desconstrução é usado para descrever o processo de desmantelamento e remoção seletiva de materiais de construção em vez da demolição tradicional. O sucesso da desconstrução vem em função do planeamento antecipado, da aplicação contínua das regras de segurança e da formação e informação para todos os trabalhadores. A triagem de materiais de demolição é de certa forma imposta pela preocupação das empresas que contribuem para a economia de recursos naturais. Esta metodologia é uma melhoria contínua, porém muito mais diversificados os materiais obtidos e a sua heterogeneidade. A correta gestão do fluxo de resíduos de construção e demolição é importante e o seu interesse na reciclagem de resíduos tem aumentado constantemente. Esse interesse é dirigido pelo grande volume desses materiais, os custos associados e uma maior consciência das oportunidades de reciclar. Auxiliando também no desenvolvimento dos mercados locais para os resíduos, tanto para os materiais que são diretamente reutilizados, como os que são reciclados, sendo novamente aplicados na construção civil. Esta dissertação aborda os métodos comummente utilizados na demolição de edifícios que dão mais atenção ao planeamento e execução dos trabalhos de demolição, garantindo a segurança no local de trabalho.
Resumo:
To meet the increasing demands of the complex inter-organizational processes and the demand for continuous innovation and internationalization, it is evident that new forms of organisation are being adopted, fostering more intensive collaboration processes and sharing of resources, in what can be called collaborative networks (Camarinha-Matos, 2006:03). Information and knowledge are crucial resources in collaborative networks, being their management fundamental processes to optimize. Knowledge organisation and collaboration systems are thus important instruments for the success of collaborative networks of organisations having been researched in the last decade in the areas of computer science, information science, management sciences, terminology and linguistics. Nevertheless, research in this area didn’t give much attention to multilingual contexts of collaboration, which pose specific and challenging problems. It is then clear that access to and representation of knowledge will happen more and more on a multilingual setting which implies the overcoming of difficulties inherent to the presence of multiple languages, through the use of processes like localization of ontologies. Although localization, like other processes that involve multilingualism, is a rather well-developed practice and its methodologies and tools fruitfully employed by the language industry in the development and adaptation of multilingual content, it has not yet been sufficiently explored as an element of support to the development of knowledge representations - in particular ontologies - expressed in more than one language. Multilingual knowledge representation is then an open research area calling for cross-contributions from knowledge engineering, terminology, ontology engineering, cognitive sciences, computational linguistics, natural language processing, and management sciences. This workshop joined researchers interested in multilingual knowledge representation, in a multidisciplinary environment to debate the possibilities of cross-fertilization between knowledge engineering, terminology, ontology engineering, cognitive sciences, computational linguistics, natural language processing, and management sciences applied to contexts where multilingualism continuously creates new and demanding challenges to current knowledge representation methods and techniques. In this workshop six papers dealing with different approaches to multilingual knowledge representation are presented, most of them describing tools, approaches and results obtained in the development of ongoing projects. In the first case, Andrés Domínguez Burgos, Koen Kerremansa and Rita Temmerman present a software module that is part of a workbench for terminological and ontological mining, Termontospider, a wiki crawler that aims at optimally traverse Wikipedia in search of domainspecific texts for extracting terminological and ontological information. The crawler is part of a tool suite for automatically developing multilingual termontological databases, i.e. ontologicallyunderpinned multilingual terminological databases. In this paper the authors describe the basic principles behind the crawler and summarized the research setting in which the tool is currently tested. In the second paper, Fumiko Kano presents a work comparing four feature-based similarity measures derived from cognitive sciences. The purpose of the comparative analysis presented by the author is to verify the potentially most effective model that can be applied for mapping independent ontologies in a culturally influenced domain. For that, datasets based on standardized pre-defined feature dimensions and values, which are obtainable from the UNESCO Institute for Statistics (UIS) have been used for the comparative analysis of the similarity measures. The purpose of the comparison is to verify the similarity measures based on the objectively developed datasets. According to the author the results demonstrate that the Bayesian Model of Generalization provides for the most effective cognitive model for identifying the most similar corresponding concepts existing for a targeted socio-cultural community. In another presentation, Thierry Declerck, Hans-Ulrich Krieger and Dagmar Gromann present an ongoing work and propose an approach to automatic extraction of information from multilingual financial Web resources, to provide candidate terms for building ontology elements or instances of ontology concepts. The authors present a complementary approach to the direct localization/translation of ontology labels, by acquiring terminologies through the access and harvesting of multilingual Web presences of structured information providers in the field of finance, leading to both the detection of candidate terms in various multilingual sources in the financial domain that can be used not only as labels of ontology classes and properties but also for the possible generation of (multilingual) domain ontologies themselves. In the next paper, Manuel Silva, António Lucas Soares and Rute Costa claim that despite the availability of tools, resources and techniques aimed at the construction of ontological artifacts, developing a shared conceptualization of a given reality still raises questions about the principles and methods that support the initial phases of conceptualization. These questions become, according to the authors, more complex when the conceptualization occurs in a multilingual setting. To tackle these issues the authors present a collaborative platform – conceptME - where terminological and knowledge representation processes support domain experts throughout a conceptualization framework, allowing the inclusion of multilingual data as a way to promote knowledge sharing and enhance conceptualization and support a multilingual ontology specification. In another presentation Frieda Steurs and Hendrik J. Kockaert present us TermWise, a large project dealing with legal terminology and phraseology for the Belgian public services, i.e. the translation office of the ministry of justice, a project which aims at developing an advanced tool including expert knowledge in the algorithms that extract specialized language from textual data (legal documents) and whose outcome is a knowledge database including Dutch/French equivalents for legal concepts, enriched with the phraseology related to the terms under discussion. Finally, Deborah Grbac, Luca Losito, Andrea Sada and Paolo Sirito report on the preliminary results of a pilot project currently ongoing at UCSC Central Library, where they propose to adapt to subject librarians, employed in large and multilingual Academic Institutions, the model used by translators working within European Union Institutions. The authors are using User Experience (UX) Analysis in order to provide subject librarians with a visual support, by means of “ontology tables” depicting conceptual linking and connections of words with concepts presented according to their semantic and linguistic meaning. The organizers hope that the selection of papers presented here will be of interest to a broad audience, and will be a starting point for further discussion and cooperation.
Resumo:
Coal contains trace elements and naturally occurring radionuclides such as 40K, 232Th, 238U. When coal is burned, minerals, including most of the radionuclides, do not burn and concentrate in the ash several times in comparison with their content in coal. Usually, a small fraction of the fly ash produced (2-5%) is released into the atmosphere. The activities released depend on many factors (concentration in coal, ash content and inorganic matter of the coal, combustion temperature, ratio between bottom and fly ash, filtering system). Therefore, marked differences should be expected between the by-products produced and the amount of activity discharged (per unit of energy produced) from different coal-fired power plants. In fact, the effects of these releases on the environment due to ground deposition have been received some attention but the results from these studies are not unanimous and cannot be understood as a generic conclusion for all coal-fired power plants. In this study, the dispersion modelling of natural radionuclides was carried out to assess the impact of continuous atmospheric releases from a selected coal plant. The natural radioactivity of the coal and the fly ash were measured and the dispersion was modelled by a Gaussian plume estimating the activity concentration at different heights up to a distance of 20 km in several wind directions. External and internal doses (inhalation and ingestion) and the resulting risk were calculated for the population living within 20 km from the coal plant. In average, the effective dose is lower than the ICRP’s limit and the risk is lower than the U.S. EPA’s limit. Therefore, in this situation, the considered exposure does not pose any risk. However, when considering the dispersion in the prevailing wind direction, these values are significant due to an increase of 232Th and 226Ra concentrations in 75% and 44%, respectively.
Resumo:
Trabalho de Projeto
Resumo:
The recent changes concerning the consumers’ active participation in the efficient management of load devices for one’s own interest and for the interest of the network operator, namely in the context of demand response, leads to the need for improved algorithms and tools. A continuous consumption optimization algorithm has been improved in order to better manage the shifted demand. It has been done in a simulation and user-interaction tool capable of being integrated in a multi-agent smart grid simulator already developed, and also capable of integrating several optimization algorithms to manage real and simulated loads. The case study of this paper enhances the advantages of the proposed algorithm and the benefits of using the developed simulation and user interaction tool.