954 resultados para Adaptação à Universidade


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nowadays, the importance of using software processes is already consolidated and is considered fundamental to the success of software development projects. Large and medium software projects demand the definition and continuous improvement of software processes in order to promote the productive development of high-quality software. Customizing and evolving existing software processes to address the variety of scenarios, technologies, culture and scale is a recurrent challenge required by the software industry. It involves the adaptation of software process models for the reality of their projects. Besides, it must also promote the reuse of past experiences in the definition and development of software processes for the new projects. The adequate management and execution of software processes can bring a better quality and productivity to the produced software systems. This work aimed to explore the use and adaptation of consolidated software product lines techniques to promote the management of the variabilities of software process families. In order to achieve this aim: (i) a systematic literature review is conducted to identify and characterize variability management approaches for software processes; (ii) an annotative approach for the variability management of software process lines is proposed and developed; and finally (iii) empirical studies and a controlled experiment assess and compare the proposed annotative approach against a compositional one. One study a comparative qualitative study analyzed the annotative and compositional approaches from different perspectives, such as: modularity, traceability, error detection, granularity, uniformity, adoption, and systematic variability management. Another study a comparative quantitative study has considered internal attributes of the specification of software process lines, such as modularity, size and complexity. Finally, the last study a controlled experiment evaluated the effort to use and the understandability of the investigated approaches when modeling and evolving specifications of software process lines. The studies bring evidences of several benefits of the annotative approach, and the potential of integration with the compositional approach, to assist the variability management of software process lines

Relevância:

20.00% 20.00%

Publicador:

Resumo:

OBJETIVO: caracterizar o desempenho de escolares da 1ª série na bateria de identificação de erros de reversão e inversão na escrita. MÉTODOS: participaram deste estudo 30 escolares de 1ª série de ensino público do município da cidade de Marília-SP, de ambos os gêneros, na faixa etária de 7 anos a 7 anos e 11 meses de idade. Como procedimento foi realizada a adaptação brasileira da bateria de identificação dos erros de reversão e inversão na escrita. Esta bateria é composta por 3 testes: teste de desempenho contínuo, teste de memória de curto prazo e teste de controle de escrita. Cada teste é composto por subtestes que visam identificar a capacidade dos escolares em identificar as inversões e reversões de letras e números isolados e letras e números em sequências. RESULTADOS: os resultados revelaram que os escolares da 1ª série apresentaram desempenho superior para identificação de reversão e inversão de letras dentro de palavras do que isoladamente, identificação da reversão para letras e números com desempenho superior se comparado a inversão de letras na palavra. Além disso, os escolares deste estudo apresentaram menor tempo para identificação de reversão e inversão de letras e números isolados do que em sequência. CONCLUSÃO: a Bateria adaptada para este estudo permitiu conhecer o perfil dos escolares de 1ª série quanto a identificação dos erros de inversão e reversão, entretanto, é necessário novas aplicações para verificar a ocorrência de identificação destes erros em populações de diferentes fase de alfabetização.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

OBJETIVO: caracterizar a inserção de egressos do Curso de Fonoaudiologia da Universidade Estadual Paulista (UNESP) - Marília, em Programas de Pós-Graduação (PPG) Stricto Sensu brasileiros. MÉTODO: foram utilizadas listas de graduados e Curriculum Vitae do egresso e do orientador. RESULTADOS: dos 537 formados, 16,57% cursaram/estavam cursando PPG e destes, 98,88% em mestrado e 37,08% também em doutorado. Na grande área de conhecimento, 50% dos egressos de mestrado vincularam-se predominantemente a programas em Ciências da Saúde, 31,80% em Ciências Humanas e 13,64% em Linguística, Letras e Artes. No doutorado, 33, 33% em Ciências Humanas, 30,30% em Ciências da Saúde e em Linguística, Letras e Artes. Quanto à área de conhecimento, predominou a vinculação, no mestrado, de 30,68% em Fonoaudiologia, 28,41% em Educação, 13,64% em Linguística e 9,09% em Medicina I; e, no doutorado, de 33,33% em Educação, 30,30% em Linguística e 9,09% em Fonoaudiologia; 55,68% dissertações e 51,52% teses focalizaram a linguagem. A UNESP predominou com 39,77% no mestrado e 48,48% no doutorado. Predominou a vinculação a Programas com conceito 4 para 52,27% dos egressos do mestrado e 45,45% do doutorado. Quando constou a informação (55,68%), todos receberam fomento. O Teste de Razão de Verossimilhança não indicou diferenças significativas dos percentuais obtidos entre o mestrado e o doutorado. CONCLUSÃO: os resultados superaram os apresentados para o mesmo Estado, mostraram a característica interdisciplinar da Ciência Fonoaudiológica e o predomínio de temática em linguagem.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Conselho Nacional de Desenvolvimento Científico e Tecnológico

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ubiquitous computing systems operate in environments where the available resources significantly change during the system operation, thus requiring adaptive and context aware mechanisms to sense changes in the environment and adapt to new execution contexts. Motivated by this requirement, a framework for developing and executing adaptive context aware applications is proposed. The PACCA framework employs aspect-oriented techniques to modularize the adaptive behavior and to keep apart the application logic from this behavior. PACCA uses abstract aspect concept to provide flexibility by addition of new adaptive concerns that extend the abstract aspect. Furthermore, PACCA has a default aspect model that considers habitual adaptive concerns in ubiquitous applications. It exploits the synergy between aspect-orientation and dynamic composition to achieve context-aware adaptation, guided by predefined policies and aim to allow software modules on demand load making possible better use of mobile devices and yours limited resources. A Development Process for the ubiquitous applications conception is also proposed and presents a set of activities that guide adaptive context-aware developer. Finally, a quantitative study evaluates the approach based on aspects and dynamic composition for the construction of ubiquitous applications based in metrics

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The use of middleware technology in various types of systems, in order to abstract low-level details related to the distribution of application logic, is increasingly common. Among several systems that can be benefited from using these components, we highlight the distributed systems, where it is necessary to allow communications between software components located on different physical machines. An important issue related to the communication between distributed components is the provision of mechanisms for managing the quality of service. This work presents a metamodel for modeling middlewares based on components in order to provide to an application the abstraction of a communication between components involved in a data stream, regardless their location. Another feature of the metamodel is the possibility of self-adaptation related to the communication mechanism, either by updating the values of its configuration parameters, or by its replacement by another mechanism, in case of the restrictions of quality of service specified are not being guaranteed. In this respect, it is planned the monitoring of the communication state (application of techniques like feedback control loop), analyzing performance metrics related. The paradigm of Model Driven Development was used to generate the implementation of a middleware that will serve as proof of concept of the metamodel, and the configuration and reconfiguration policies related to the dynamic adaptation processes. In this sense was defined the metamodel associated to the process of a communication configuration. The MDD application also corresponds to the definition of the following transformations: the architectural model of the middleware in Java code, and the configuration model to XML

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Distributed multimedia systems have highly variable characteristics, resulting in new requirements while new technologies become available or in the need for adequacy in accordance with the amount of available resources. So, these systems should provide support for dynamic adaptations in order to adjust their structures and behaviors at runtime. This paper presents an approach to adaptation model-based and proposes a reflective and component-based framework for construction and support of self-adaptive distributed multimedia systems, providing many facilities for the development and evolution of such systems, such as dynamic adaptation. The propose is to keep one or more models to represent the system at runtime, so some external entity can perform an analysis of these models by identifying problems and trying to solve them. These models integrate the reflective meta-level, acting as a system self-representation. The framework defines a meta-model for description of self-adaptive distributed multimedia applications, which can represent components and their relationships, policies for QoS specification and adaptation actions. Additionally, this paper proposes an ADL and architecture for model-based adaptation. As a case study, this paper presents some scenarios to demonstrate the application of the framework in practice, with and without the use of ADL, as well as check some characteristics related to dynamic adaptation

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The process for choosing the best components to build systems has become increasingly complex. It becomes more critical if it was need to consider many combinations of components in the context of an architectural configuration. These circumstances occur, mainly, when we have to deal with systems involving critical requirements, such as the timing constraints in distributed multimedia systems, the network bandwidth in mobile applications or even the reliability in real-time systems. This work proposes a process of dynamic selection of architectural configurations based on non-functional requirements criteria of the system, which can be used during a dynamic adaptation. This proposal uses the MAUT theory (Multi-Attribute Utility Theory) for decision making from a finite set of possibilities, which involve multiple criteria to be analyzed. Additionally, it was proposed a metamodel which can be used to describe the application s requirements in terms of the non-functional requirements criteria and their expected values, to express them in order to make the selection of the desired configuration. As a proof of concept, it was implemented a module that performs the dynamic choice of configurations, the MoSAC. This module was implemented using a component-based development approach (CBD), performing a selection of architectural configurations based on the proposed selection process involving multiple criteria. This work also presents a case study where an application was developed in the context of Digital TV to evaluate the time spent on the module to return a valid configuration to be used in a middleware with autoadaptative features, the middleware AdaptTV

Relevância:

20.00% 20.00%

Publicador:

Resumo:

On the last years, several middleware platforms for Wireless Sensor Networks (WSN) were proposed. Most of these platforms does not consider issues of how integrate components from generic middleware architectures. Many requirements need to be considered in a middleware design for WSN and the design, in this case, it is possibility to modify the source code of the middleware without changing the external behavior of the middleware. Thus, it is desired that there is a middleware generic architecture that is able to offer an optimal configuration according to the requirements of the application. The adoption of middleware based in component model consists of a promising approach because it allows a better abstraction, low coupling, modularization and management features built-in middleware. Another problem present in current middleware consists of treatment of interoperability with external networks to sensor networks, such as Web. Most current middleware lacks the functionality to access the data provided by the WSN via the World Wide Web in order to treat these data as Web resources, and they can be accessed through protocols already adopted the World Wide Web. Thus, this work presents the Midgard, a component-based middleware specifically designed for WSNs, which adopts the architectural patterns microkernel and REST. The microkernel architectural complements the component model, since microkernel can be understood as a component that encapsulates the core system and it is responsible for initializing the core services only when needed, as well as remove them when are no more needed. Already REST defines a standardized way of communication between different applications based on standards adopted by the Web and enables him to treat WSN data as web resources, allowing them to be accessed through protocol already adopted in the World Wide Web. The main goals of Midgard are: (i) to provide easy Web access to data generated by WSN, exposing such data as Web resources, following the principles of Web of Things paradigm and (ii) to provide WSN application developer with capabilities to instantiate only specific services required by the application, thus generating a customized middleware and saving node resources. The Midgard allows use the WSN as Web resources and still provide a cohesive and weakly coupled software architecture, addressing interoperability and customization. In addition, Midgard provides two services needed for most WSN applications: (i) configuration and (ii) inspection and adaptation services. New services can be implemented by others and easily incorporated into the middleware, because of its flexible and extensible architecture. According to the assessment, the Midgard provides interoperability between the WSN and external networks, such as web, as well as between different applications within a single WSN. In addition, we assessed the memory consumption, the application image size, the size of messages exchanged in the network, and response time, overhead and scalability on Midgard. During the evaluation, the Midgard proved satisfies their goals and shown to be scalable without consuming resources prohibitively

Relevância:

20.00% 20.00%

Publicador:

Resumo:

With the advance of the Cloud Computing paradigm, a single service offered by a cloud platform may not be enough to meet all the application requirements. To fulfill such requirements, it may be necessary, instead of a single service, a composition of services that aggregates services provided by different cloud platforms. In order to generate aggregated value for the user, this composition of services provided by several Cloud Computing platforms requires a solution in terms of platforms integration, which encompasses the manipulation of a wide number of noninteroperable APIs and protocols from different platform vendors. In this scenario, this work presents Cloud Integrator, a middleware platform for composing services provided by different Cloud Computing platforms. Besides providing an environment that facilitates the development and execution of applications that use such services, Cloud Integrator works as a mediator by providing mechanisms for building applications through composition and selection of semantic Web services that take into account metadata about the services, such as QoS (Quality of Service), prices, etc. Moreover, the proposed middleware platform provides an adaptation mechanism that can be triggered in case of failure or quality degradation of one or more services used by the running application in order to ensure its quality and availability. In this work, through a case study that consists of an application that use services provided by different cloud platforms, Cloud Integrator is evaluated in terms of the efficiency of the performed service composition, selection and adaptation processes, as well as the potential of using this middleware in heterogeneous computational clouds scenarios

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Self-adaptive software system is able to change its structure and/or behavior at runtime due to changes in their requirements, environment or components. One way to archieve self-adaptation is the use a sequence of actions (known as adaptation plans) which are typically defined at design time. This is the approach adopted by Cosmos - a Framework to support the configuration and management of resources in distributed environments. In order to deal with the variability inherent of self-adaptive systems, such as, the appearance of new components that allow the establishment of configurations that were not envisioned at development time, this dissertation aims to give Cosmos the capability of generating adaptation plans of runtime. In this way, it was necessary to perform a reengineering of the Cosmos Framework in order to allow its integration with a mechanism for the dynamic generation of adaptation plans. In this context, our work has been focused on conducting a reengineering of Cosmos. Among the changes made to in the Cosmos, we can highlight: changes in the metamodel used to represent components and applications, which has been redefined based on an architectural description language. These changes were propagated to the implementation of a new Cosmos prototype, which was then used for developing a case study application for purpose of proof of concept. Another effort undertaken was to make Cosmos more attractive by integrating it with another platform, in the case of this dissertation, the OSGi platform, which is well-known and accepted by the industry

Relevância:

20.00% 20.00%

Publicador:

Resumo:

One way to deal with the high complexity of current software systems is through selfadaptive systems. Self-adaptive system must be able to monitor themselves and their environment, analyzing the monitored data to determine the need for adaptation, decide how the adaptation will be performed, and finally, make the necessary adjustments. One way to perform the adaptation of a system is generating, at runtime, the process that will perform the adaptation. One advantage of this approach is the possibility to take into account features that can only be evaluated at runtime, such as the emergence of new components that allow new architectural arrangements which were not foreseen at design time. In this work we have as main objective the use of a framework for dynamic generation of processes to generate architectural adaptation plans on OSGi environment. Our main interest is evaluate how this framework for dynamic generation of processes behave in new environments

Relevância:

20.00% 20.00%

Publicador:

Resumo:

OBJETIVO: Avaliar o efeito da desnutrição protéica na parede intestinal do rato através da medida de força de ruptura e dosagem do colágeno tecidual no íleo e cólon distal. MÉTODOS: Foram utilizados 120 ratos, pesando em média 100g, que receberam durante 07 dias uma dieta padrão, contendo 20% de caseína para adaptação dos animais as condições do biotério. Após esse período os animais foram divididos em dois grupos de 60, o controle denominado grupo um que recebeu a dieta padrão, e o grupo teste denominado grupo dois, que recebeu dieta hipoprotéica contendo 2% de caseína. Os dois grupos receberam suas respectivas dietas por um período de 21 dias. Após esse período iniciou-se o sacrifício seqüencial dos animais em ambos os grupos, em número de 12 animais em cada momento, correspondendo ao dia Zero (MO), 4º dia (M1), 7º dia (M2), 14º dia (M3), e 21º dia (M4) sendo mantida a mesma dieta até o final do sacrifício. em cada momento foram avaliados o peso corpóreo, albumina sanguínea, hidroxiprolina tecidual, relação hidroxiprolina/proteína tecidual e a força de ruptura no segmento ileal e cólico dos animais. RESULTADOS: Observou-se que a força de ruptura do segmento ileal e do cólon distal foi menor nos animais desnutridos (Grupo 2). A perda da resistência mecânica foi maior no segmento do cólon distal do que no segmento ileal, provavelmente pela menor concentração do colágeno tecidual no cólon distal. CONCLUSÃO: A desnutrição protéica induz a diminuição da resistência mecânica no íleo e no cólon distal associado a diminuição do colágeno tecidual na parede intestinal.