861 resultados para Future Technology
Resumo:
In recent years, learning analytics (LA) has attracted a great deal of attention in technology-enhanced learning (TEL) research as practitioners, institutions, and researchers are increasingly seeing the potential that LA has to shape the future TEL landscape. Generally, LA deals with the development of methods that harness educational data sets to support the learning process. This paper provides a foundation for future research in LA. It provides a systematic overview on this emerging field and its key concepts through a reference model for LA based on four dimensions, namely data, environments, context (what?), stakeholders (who?), objectives (why?), and methods (how?). It further identifies various challenges and research opportunities in the area of LA in relation to each dimension.
Resumo:
Transcatheter aortic valve replacement (TAVR) constitutes a relatively new treatment option for the patients with severe symptomatic aortic stenosis. Evidence from registries and randomized control trials has underscored the value of this treatment in inoperable and high risk populations, while new developments in valve technology and TAVR enabling devices have reduced the risk of complications, simplified the procedure, and broadened the applications of this therapy. The initial promising clinical results and the potential of an effective less invasive treatment of aortic stenosis has not only created high expectations but also the need to address the pitfalls of TAVR technology. The evolving knowledge concerning the groups of patients who would benefit from this treatment, the limited long term follow-up data, the concerns about devices' long term durability, and the severity of complications remain important caveats which restrict the widespread clinical adoption of TAVR. The aim of this review article is to present the recent advances, highlight the limitations of TAVR technology, and discuss the future perspectives in this rapidly evolving field.
Resumo:
Swiss lake-side settlements dating between 4300 and 800 BC were first recognized in the early 19th century and between 1854 and 1880 early research boomed due to the first scientific studies and the artificial lowering of lakes in Western Switzerland. In the 20th century underwater and wetland archaeology experienced an enormous surge not only because of large rescue excavations, due to extensive motorway construction projects but also due to the evolution of modern IT technology in the 1970s. For the first time huge quantities of ancient wooden structures could be dated by dendrochronology. This produced a quantum leap in the 150 years of pile-dwelling research. In 2011, the UNESCO World Heritage Committee recognized the outstanding universal value of these sites. This article presents an overview about Swiss pile-dwellings of the Neolithic and the Bronze Age and the results of two recent diploma works (case study Sutz-Lattrigen Haupstation innen and case study Seedorf Lobsigensee) as examples of research and cooperation between universities and government agencies for cultural heritage management.
Resumo:
How do developers and designers of a new technology make sense of intended users? The critical groundwork for user-centred technology development begins not by involving actual users' exposure to the technological artefact but much earlier, with designers' and developers' vision of future users. Thus, anticipating intended users is critical to technology uptake. We conceptualise the anticipation of intended users as a form of prospective sensemaking in technology development. Employing a narrative analytical approach and drawing on four key communities in the development of Grid computing, we reconstruct how each community anticipated the intended Grid user. Based on our findings, we conceptualise user anticipation in Terms of two key dimensions, namely the intended possibility to inscribe user needs into the technological artefact as well as the intended scope of the application domain. In turn, these dimensions allow us to develop an initial typology of intended user concepts that in turn might provide a key building block towards a generic typology of intended users.
Resumo:
Maternal thromboembolism and a spectrum of placenta-mediated complications including the pre-eclampsia syndromes, fetal growth restriction, fetal loss, and abruption manifest a shared etiopathogenesis and predisposing risk factors. Furthermore, these maternal and fetal complications are often linked to subsequent maternal health consequences that comprise the metabolic syndrome, namely, thromboembolism, chronic hypertension, and type II diabetes. Traditionally, several lines of evidence have linked vasoconstriction, excessive thrombosis and inflammation, and impaired trophoblast invasion at the uteroplacental interface as hallmark features of the placental complications. "Omic" technologies and biomarker development have been largely based upon advances in vascular biology, improved understanding of the molecular basis and biochemical pathways responsible for the clinically relevant diseases, and increasingly robust large cohort and/or registry based studies. Advances in understanding of innate and adaptive immunity appear to play an important role in several pregnancy complications. Strategies aimed at improving prediction of these pregnancy complications are often incorporating hemodynamic blood flow data using non-invasive imaging technologies of the utero-placental and maternal circulations early in pregnancy. Some evidence suggests that a multiple marker approach will yield the best performing prediction tools, which may then in turn offer the possibility of early intervention to prevent or ameliorate these pregnancy complications. Prediction of maternal cardiovascular and non-cardiovascular consequences following pregnancy represents an important area of future research, which may have significant public health consequences not only for cardiovascular disease, but also for a variety of other disorders, such as autoimmune and neurodegenerative diseases.
Resumo:
SOLUTIONS (2013 to 2018) is a European Union Seventh Framework Programme Project (EU-FP7). The project aims to deliver a conceptual framework to support the evidence-based development of environmental policies with regard to water quality. SOLUTIONS will develop the tools for the identification, prioritisation and assessment of those water contaminants that may pose a risk to ecosystems and human health. To this end, a new generation of chemical and effect-based monitoring tools is developed and integrated with a full set of exposure, effect and risk assessment models. SOLUTIONS attempts to address legacy, present and future contamination by integrating monitoring and modelling based approaches with scenarios on future developments in society, economy and technology and thus in contamination. The project follows a solutions-oriented approach by addressing major problems of water and chemicals management and by assessing abatement options. SOLUTIONS takes advantage of the access to the infrastructure necessary to investigate the large basins of the Danube and Rhine as well as relevant Mediterranean basins as case studies, and puts major efforts on stakeholder dialogue and support. Particularly, the EU Water Framework Directive (WFD) Common Implementation Strategy (CIS) working groups, International River Commissions, and water works associations are directly supported with consistent guidance for the early detection, identification, prioritisation, and abatement of chemicals in the water cycle. SOLUTIONS will give a specific emphasis on concepts and tools for the impact and risk assessment of complex mixtures of emerging pollutants, their metabolites and transformation products. Analytical and effect-based screening tools will be applied together with ecological assessment tools for the identification of toxicants and their impacts. The SOLUTIONS approach is expected to provide transparent and evidence-based candidates or River Basin Specific Pollutants in the case study basins and to assist future review of priority pollutants under the WFD as well as potential abatement options.
Resumo:
The medical education community is working-across disciplines and across the continuum-to address the current challenges facing the medical education system and to implement strategies to improve educational outcomes. Educational technology offers the promise of addressing these important challenges in ways not previously possible. The authors propose a role for virtual patients (VPs), which they define as multimedia, screen-based interactive patient scenarios. They believe VPs offer capabilities and benefits particularly well suited to addressing the challenges facing medical education. Well-designed, interactive VP-based learning activities can promote the deep learning that is needed to handle the rapid growth in medical knowledge. Clinically oriented learning from VPs can capture intrinsic motivation and promote mastery learning. VPs can also enhance trainees' application of foundational knowledge to promote the development of clinical reasoning, the foundation of medical practice. Although not the entire solution, VPs can support competency-based education. The data created by the use of VPs can serve as the basis for multi-institutional research that will enable the medical education community both to better understand the effectiveness of educational interventions and to measure progress toward an improved system of medical education.
Resumo:
Purpose. The purpose of this study was to investigate the impact of a motivational weight management DVD on knowledge of obesity related diseases, readiness, motivation, and self-efficacy to lose weight, connectedness to their care provider, and patients return to clinic. Design. A randomized control trial was conducted in which 40 overweight/obese adolescents and their parents/caregivers were randomly assigned to standard care alone or standard care plus DVD. Subjects completed a set of pre- and post-questionnaire measures. A group of 22 patients was also formed as a historical control group in order to account for the potential effect of extra attention given to subjects prospectively enrolled. Methods. The adolescents and their parent/caregiver were placed into a patient room. Consent was obtained and a set of written pre-questionnaires were given to both the parent and the adolescent. Standard care was provided to all patients by the Registered Dietitian and physician; the DVD was shown in addition to standard care among the intervention group. A set of post-questionnaires were given and compensation was provided. Analysis. Groups were compared to determine equivalence at baseline. Analysis of covariance was used to evaluate changes over time, while controlling for pre-test scores and race/ethnicity. Results. Parents who viewed the DVD experienced greater changes in correct knowledge as compared to parents who did not view the DVD. Conclusion. Our study found only one substantial benefit of the DVD beyond standard clinical practices. This is an important area for change as it increased awareness of obesity as a serious disease and has future clinical implications.^
Resumo:
Information technology (IT) in the hospital organization is fast becoming a key asset, particularly in light of recent reform legislation in the United States calling for expanding the role of IT in our health care system. Future payment reductions to hospitals included in current health reform are based on expected improvements in hospital operating efficiency. Since over half of hospital expenses are for labor, improved efficiency in use of labor resources can be critical in meeting this challenge. Policy makers have touted the value of IT investments to improve efficiency in response to payment reductions. ^ This study was the first to directly examine the relationship between electronic health record (EHR) technology and staffing efficiency in hospitals. As the hospital has a myriad of outputs for inpatient and outpatient care, efficiency was measured using an industry standard performance metric – full time equivalent employees per adjusted occupied bed (FTE/AOB). Three hypotheses were tested in this study.^ To operationalize EHR technology adoption, we developed three constructs to model adoption, each of which was tested by separate hypotheses. The first hypothesis that a larger number of EHR applications used by a hospital would be associated with greater staffing efficiency (or lower values of FTE/AOB) was not accepted. Association between staffing efficiency and specific EHR applications was the second hypothesis tested and accepted with some applications showing significant impacts on observed values for FTE/AOB. Finally, the hypothesis that the longer an EHR application was used in a hospital would be associated with greater labor efficiency was not accepted as the model showed few statistically significant relationships to FTE/AOB performance. Generally, there does not appear a strong relationship between EHR usage and improved labor efficiency in hospitals.^ While returns on investment from EHR usage may not come from labor efficiencies, they may be better sought using measures of quality, contribution to an efficient and effective local health care system, and improved customer satisfaction through greater patient throughput.^
Resumo:
In the IT industry, there has been a remarkable increase in the demand for system LSI. A system LSI must be produced, tailor-designed for each electrical appliance. It is said that this production method has made the IC cycle ambiguous in recent years. It can be sought that the choice of whether the economy pursues a development path centering on technology which is tradable or technology which is embodied in labor, depends on the historical background. The relationship between these two types of technologies is changing rapidly every one or two years. In this background, the analysis is focused on the new trend of technology. In the section 2, the newest trend of technology in the field of system LSI is explained. Then, which kind of technology will be developed and how it will have an affect in the near future, is considered.
Resumo:
Developing countries are experiencing unprecedented levels of economic growth. As a result, they will be responsible for most of the future growth in energy demand and greenhouse gas (GHG) emissions. Curbing GHG emissions in developing countries has become one of the cornerstones of a future international agreement under the United Nations Framework Convention for Climate Change (UNFCCC). However, setting caps for developing countries’ GHG emissions has encountered strong resistance in the current round of negotiations. Continued economic growth that allows poverty eradication is still the main priority for most developing countries, and caps are perceived as a constraint to future growth prospects. The development, transfer and use of low-carbon technologies have more positive connotations, and are seen as the potential path towards low-carbon development. So far, the success of the UNFCCC process in improving the levels of technology transfer (TT) to developing countries has been limited. This thesis analyses the causes for such limited success and seeks to improve on the understanding about what constitutes TT in the field of climate change, establish the factors that enable them in developing countries and determine which policies could be implemented to reinforce these factors. Despite the wide recognition of the importance of technology and knowledge transfer to developing countries in the climate change mitigation policy agenda, this issue has not received sufficient attention in academic research. Current definitions of climate change TT barely take into account the perspective of actors involved in actual climate change TT activities, while respective measurements do not bear in mind the diversity of channels through which these happen and the outputs and effects that they convey. Furthermore, the enabling factors for TT in non-BRIC (Brazil, Russia, India, China) developing countries have been seldom investigated, and policy recommendations to improve the level and quality of TTs to developing countries have not been adapted to the specific needs of highly heterogeneous countries, commonly denominated as “developing countries”. This thesis contributes to enriching the climate change TT debate from the perspective of a smaller emerging economy (Chile) and by undertaking a quantitative analysis of enabling factors for TT in a large sample of developing countries. Two methodological approaches are used to study climate change TT: comparative case study analysis and quantitative analysis. Comparative case studies analyse TT processes in ten cases based in Chile, all of which share the same economic, technological and policy frameworks, thus enabling us to draw conclusions on the enabling factors and obstacles operating in TT processes. The quantitative analysis uses three methodologies – principal component analysis, multiple regression analysis and cluster analysis – to assess the performance of developing countries in a number of enabling factors and the relationship between these factors and indicators of TT, as well as to create groups of developing countries with similar performances. The findings of this thesis are structured to provide responses to four main research questions: What constitutes technology transfer and how does it happen? Is it possible to measure technology transfer, and what are the main challenges in doing so? Which factors enable climate change technology transfer to developing countries? And how do different developing countries perform in these enabling factors, and how can differentiated policy priorities be defined accordingly? vi Resumen Los paises en desarrollo estan experimentando niveles de crecimiento economico sin precedentes. Como consecuencia, se espera que sean responsables de la mayor parte del futuro crecimiento global en demanda energetica y emisiones de Gases de Efecto de Invernadero (GEI). Reducir las emisiones de GEI en los paises en desarrollo es por tanto uno de los pilares de un futuro acuerdo internacional en el marco de la Convencion Marco de las Naciones Unidas para el Cambio Climatico (UNFCCC). La posibilidad de compromisos vinculantes de reduccion de emisiones de GEI ha sido rechazada por los paises en desarrollo, que perciben estos limites como frenos a su desarrollo economico y a su prioridad principal de erradicacion de la pobreza. El desarrollo, transferencia y uso de tecnologias bajas en carbono tiene connotaciones mas positivas y se percibe como la via hacia un crecimiento bajo en carbono. Hasta el momento, la UNFCCC ha tenido un exito limitado en la promocion de transferencias de tecnologia (TT) a paises en desarrollo. Esta tesis analiza las causas de este resultado y busca mejorar la comprension sobre que constituye transferencia de tecnologia en el area de cambio climatico, cuales son los factores que la facilitan en paises en desarrollo y que politicas podrian implementarse para reforzar dichos factores. A pesar del extendido reconocimiento sobre la importancia de la transferencia de tecnologia a paises en desarrollo en la agenda politica de cambio climatico, esta cuestion no ha sido suficientemente atendida por la investigacion existente. Las definiciones actuales de transferencia de tecnologia relacionada con la mitigacion del cambio climatico no tienen en cuenta la diversidad de canales por las que se manifiestan o los efectos que consiguen. Los factores facilitadores de TT en paises en desarrollo no BRIC (Brasil, Rusia, India y China) apenas han sido investigados, y las recomendaciones politicas para aumentar el nivel y la calidad de la TT no se han adaptado a las necesidades especificas de paises muy heterogeneos aglutinados bajo el denominado grupo de "paises en desarrollo". Esta tesis contribuye a enriquecer el debate sobre la TT de cambio climatico con la perspectiva de una economia emergente de pequeno tamano (Chile) y el analisis cuantitativo de factores que facilitan la TT en una amplia muestra de paises en desarrollo. Se utilizan dos metodologias para el estudio de la TT a paises en desarrollo: analisis comparativo de casos de estudio y analisis cuantitativo basado en metodos multivariantes. Los casos de estudio analizan procesos de TT en diez casos basados en Chile, para derivar conclusiones sobre los factores que facilitan u obstaculizan el proceso de transferencia. El analisis cuantitativo multivariante utiliza tres metodologias: regresion multiple, analisis de componentes principales y analisis cluster. Con dichas metodologias se busca analizar el posicionamiento de diversos paises en cuanto a factores que facilitan la TT; las relaciones entre dichos factores e indicadores de transferencia tecnologica; y crear grupos de paises con caracteristicas similares que podrian beneficiarse de politicas similares para la promocion de la transferencia de tecnologia. Los resultados de la tesis se estructuran en torno a cuatro preguntas de investigacion: .Que es la transferencia de tecnologia y como ocurre?; .Es posible medir la transferencia de tecnologias de bajo carbono?; .Que factores facilitan la transferencia de tecnologias de bajo carbono a paises en desarrollo? y .Como se puede agrupar a los paises en desarrollo en funcion de sus necesidades politicas para la promocion de la transferencia de tecnologias de bajo carbono?
Resumo:
The main objective of this course, conducted by Jóvenes Nucleares (Spanish Young Generation in Nuclear, JJNN), a non-profit organization that depends on the Spanish Nuclear Society (SNE) is to pass on basic knowledge about Science and Nuclear Technology to the general public, mostly students and introduce them to its most relevant points. The purposes of this course are to provide general information, to answer the most common questions about Nuclear Energy and to motivate the young students to start a career in nuclear. Therefore, it is directed mainly to high school and university students, but also to general people that wants to learn about the key issues of such an important matter in our society. Anybody could attend the course, as no specific scientific education is required. The course is done at least once a year, during the Annual Meeting of the Spanish Nuclear Society, which takes place in a different Spanish city each time. The course is done also to whichever university or institution that asks for it to JJNN, with the only limit of the presenter´s availability. The course is divided into the following chapters: Physical nuclear and radiation principles, Nuclear power plants, Nuclear safety, Nuclear fuel, Radioactive waste, Decommission of nuclear facilities, Future nuclear power plants, Other uses of nuclear technology, Nuclear energy, climate change and sustainable development. The course is divided into 15 minutes lessons on the above topics, imparted by young professionals, experts in the field that belongs either to the Spanish Young Generation in Nuclear, either to companies and institutions related with nuclear energy. At the end of the course, a 200 pages book with the contents of the course is handed to every member of the audience. This book is also distributed in other course editions at high schools and universities in order to promote the scientific dissemination of the Nuclear Technology. As an extra motivation, JJNN delivers a course certificate to the assistants. At the end of the last edition course, in Santiago de Compostela, the assistants were asked to provide a feedback about it. Some really interesting lessons were learned, that will be very useful to improve next editions of the course. As a general conclusion of the courses it can be said that many of the students that have assisted to the course have increased their motivation in the nuclear field, and hopefully it will help the young talents to choose the nuclear field to develop their careers
Resumo:
Wireless communication is the transfer of information from one place to another without using wires. From the earliest times, humans have felt the need to develop techniques of remote communication. From this need arose the smoke signals, communication by sun reflection in mirrors and so on. But today the telecommunications electronic devices such as telephone, television, radio or computer. Radio and television are used for one-way communication. Telephone and computer are used for two-way communication. In wireless networks there is almost unlimited mobility, we can access the network almost anywhere or anytime. In wired networks we have the restriction of using the services in fixed area services. The demand of the wireless is increasing very fast; everybody wants broadband services anywhere anytime. WiMAX (Worldwide Interoperability for Microwave Access) is a broadband wireless technology based on IEEE 802.16-2004 and IEEE 802.16e-2005 that appears to solve this demand. WIMAX is a system that allows wireless data transmission in areas of up to 48 km of radius. It is designed as a wireless alternative to ADSL and a way to connect nodes in wireless metropolitan areas network. Unlike wireless systems that are limited in most cases, about 100 meter, providing greater coverage and more bandwidth. WIMAX promises to achieve high data transmission rates over large areas with a great amount of users. This alternative to the networks of broadband access common as DSL o Wi-Fi, can give broadband access to places quickly to rural areas and developing areas around the world. This paper is a study of WIMAX technology and market situation. First, the paper is responsible for explaining the technical aspects of WIMAX. For this gives an overview of WIMAX standards, physical layer, MAC layer and WiMAX, Technology and Market Beijing University of Post and Telecommunications 2 WIMAX network architecture. Second, the paper address the issue of market in which provides an overview of development and deployment of WIMAX to end the future development trend of WIMAX is addressed. RESUMEN: Por comunicaciones inalámbricas se entiende la transferencia de información desde un lugar a otro sin la necesidad de un soporte físico como es por ejemplo el cable. Por lo que remontándose a los principios de la existencia del ser humano, nos damos cuenta de que el ser humano siempre ha sentido la necesidad de desarrollar técnicas para lograr comunicarse a distancia con sus semejantes. De dicha necesidad, surgieron técnicas tan ancestrales como puede ser la comunicación mediante señales de humo o por reflexión de los rayos solares en espejos entre otras. La curiosidad del ser humano y la necesidad de comunicarse a distancia fue la que llevó a Alexander Graham Bell a inventar el teléfono en 1876. La aparición de un dispositivo que permitía comunicarse a distancia permitiendo escuchar la voz de aquella persona con la que se quería hablar, supuso una revolución no solo en el panorama tecnológico, si no también en el panorama social. Pues a parte de permitir comunicaciones a larga distancia, solventó el problema de la comunicación en “tiempo real”. A raíz de este invento, la tecnología en materia de comunicación ha ido avanzando significativamente, más concretamente en lo referido a las comunicaciones inalámbricas. En 1973 se realizó la primera llamada desde un terminal móvil aunque no fue hasta 1983 cuando se empezó a comercializar dicho terminal, lo que supuso un cambio de hábitos y costumbres para la sociedad. Desde la aparición del primer móvil el crecimiento del mercado ha sido exponencial, lo que ha repercutido en una demanda impensable de nuevas aplicaciones integradas en dichos dispositivos móviles que satisfagan las necesidades que día a día autogenera la sociedad. Tras conseguir realizar llamadas a larga distancia de forma inalámbrica, el siguiente paso fue la creación de los SMS (Short Message System) lo que supuso una nueva revolución además de abaratar costes al usuario a la hora de comunicarse. Pero el gran reto para la industria de las comunicaciones móviles surgió con la aparición de internet. Todo el mundo sentía la necesidad de poder conectarse a esa gran base de datos que es internet en cualquier parte y en cualquier momento. Las primeras conexiones a internet desde dispositivos móviles se realizaron a través de la tecnología WAP (Wireless Application Protocol) hasta la aparición de la tecnología GPRS que permitía la conexión mediante protocolo TCP/IP. A partir de estas conexiones han surgido otras tecnologías, como EDGE, HSDPA, etc., que permitían y permiten la conexión a internet desde dispositivos móviles. Hoy en día la demanda de servicios de red inalámbrica crece de forma rápida y exponencial, todo el mundo quiere servicios de banda ancha en cualquier lugar y en cualquier momento. En este documento se analiza la tecnología WiMAX ( Worldwide Interoperability for Microwave Access) que es una tecnología de banda ancha basada en el estándar IEEE 802.16 creada para brindar servicios a la demanda emergente en la banda ancha desde un punto de vista tecnológico, donde se da una visión de la parte técnica de la tecnología; y desde el punto de vista del mercado, donde se analiza el despliegue y desarrollo de la tecnología desde el punto de vista de negocio. WiMAX es una tecnología que permite la transmisión inalámbrica de datos en áreas de hasta 48Km de radio y que está diseñada como alternativa inalámbrica para ADSL y para conectar nodos de red inalámbrica en áreas metropolitanas. A diferencia de los sistemas inalámbricos existentes que están limitados en su mayoría a unos cientos de metros, WiMAX ofrece una mayor cobertura y un mayor ancho de banda que permita dar soporte a nuevas aplicaciones, además de alcanzar altas tasas de transmisión de datos en grandes áreas con una gran cantidad de usuarios. Se trata de una alternativa a las redes de acceso de banda ancha como DSL o Wi-Fi, que puede dar acceso de banda ancha a lugares tales como zonas rurales o zonas en vías de desarrollo por todo el mundo con rapidez. Existen dos tecnologías de WiMAX, WiMAX fijo (basado en el estándar IEEE 802.16d-2004) y WiMAX móvil (basado en el estándar IEEE 802.16e-2005). La tecnología fija está diseñada para comunicaciones punto a multipunto, mientras que la fija lo está para comunicaciones multipunto a multipunto. WiMAX móvil se basa en la tecnología OFDM que ofrece ventajas en términos de latencia, eficiencia en el uso del espectro y soporte avanzado para antenas. La modulación OFDM es muy robusta frente al multitrayecto, que es muy habitual en los canales de radiodifusión, frente al desvanecimiento debido a las condiciones meteorológicas y frente a las interferencias de RF. Una vez creada la tecnología WiMAX, poseedora de las características idóneas para solventar la demanda del mercado, ha de darse el siguiente paso, hay que convencer a la industria de las telecomunicaciones de que dicha tecnología realmente es la solución para que apoyen su implantación en el mercado de la banda ancha para las redes inalámbricas. Es aquí donde entra en juego el estudio del mercado que se realiza en este documento. WiMAX se enfrenta a un mercado exigente en el que a parte de tener que dar soporte a la demanda técnica, ha de ofrecer una rentabilidad económica a la industria de las comunicaciones móviles y más concretamente a las operadoras móviles que son quienes dentro del sector de las telecomunicaciones finalmente han de confiar en la tecnología para dar soporte a sus usuarios ya que estos al fin y al cabo lo único que quieren es que su dispositivo móvil satisfaga sus necesidades independientemente de la tecnología que utilicen para tener acceso a la red inalámbrica de banda ancha. Quizás el mayor problema al que se ha enfrentado WiMAX haya sido la situación económica en la que se encuentra el mundo. WiMAX a comenzado su andadura en uno de los peores momentos, pero aun así se presenta como una tecnología capaz de ayudar al mundo a salir hacia delante en estos tiempos tan duros. Finalmente se analiza uno de los debates existentes hoy en día en el sector de las comunicaciones móviles, WiMAX vs. LTE. Como se puede observar en el documento realmente una tecnología no saldrá victoriosa frente a la otra, si no que ambas tecnologías podrán coexistir y trabajar de forma conjunta.
Resumo:
The future Internet is expected to be composed of a mesh of interoperable web services accessible from all over the web. This approach has not yet caught on since global user?service interaction is still an open issue. This paper states one vision with regard to next-generation front-end Web 2.0 technology that will enable integrated access to services, contents and things in the future Internet. In this paper, we illustrate how front-ends that wrap traditional services and resources can be tailored to the needs of end users, converting end users into prosumers (creators and consumers of service-based applications). To do this, we propose an architecture that end users without programming skills can use to create front-ends, consult catalogues of resources tailored to their needs, easily integrate and coordinate front-ends and create composite applications to orchestrate services in their back-end. The paper includes a case study illustrating that current user-centred web development tools are at a very early stage of evolution. We provide statistical data on how the proposed architecture improves these tools. This paper is based on research conducted by the Service Front End (SFE) Open Alliance initiative.
Resumo:
Area, launched in 1999 with the Bologna Declaration, has bestowed such a magnitude and unprecedented agility to the transformation process undertaken by European universities. However, the change has been more profound and drastic with regards to the use of new technologies both inside and outside the classroom. This article focuses on the study and analysis of the technology’s history within the university education and its impact on teachers, students and teaching methods. All the elements that have been significant and innovative throughout the history inside the teaching process have been analyzed, from the use of blackboard and chalk during lectures, the use of slide projectors and transparent slides, to the use of electronic whiteboards and Internet nowadays. The study is complemented with two types of surveys that have been performed among teachers and students during the school years 1999 - 2011 in the School of Civil Engineering at the Polytechnic University of Madrid. The pros and cons of each of the techniques and methodologies used in the learning process over the last decades are described, unfolding how they have affected the teacher, who has evolved from writing on a whiteboard to project onto a screen, the student, who has evolved from taking handwritten notes to download information or search the Internet, and the educational process, that has evolved from the lecture to acollaborative learning and project-based learning. It is unknown how the process of learning will evolve in the future, but we do know the consequences that some of the multimedia technologies are having on teachers, students and the learning process. It is our goal as teachers to keep ourselves up to date, in order to offer the student adequate technical content, while providing proper motivation through the use of new technologies. The study provides a forecast in the evolution of multimedia within the classroom and the renewal of the education process, which in our view, will set the basis for future learning process within the context of this new interactive era.