900 resultados para Local computer network
Resumo:
BACKGROUND: To asses the clinical profile, treatment outcome and prognostic factors in primary breast lymphoma (PBL). METHODS: Between 1970 and 2000, 84 consecutive patients with PBL were treated in 20 institutions of the Rare Cancer Network. Forty-six patients had Ann Arbor stage IE, 33 stage IIE, 1 stage IIIE, 2 stage IVE and 2 an unknown stage. Twenty-one underwent a mastectomy, 39 conservative surgery and 23 biopsy; 51 received radiotherapy (RT) with (n = 37) or without (n = 14) chemotherapy. Median RT dose was 40 Gy (range 12-55 Gy). RESULTS: Ten (12%) patients progressed locally and 43 (55%) had a systemic relapse. Central nervous system (CNS) was the site of relapse in 12 (14%) cases. The 5-yr overall survival, lymphoma-specific survival, disease-free survival and local control rates were 53%, 59%, 41% and 87% respectively. In the univariate analyses, favorable prognostic factors were early stage, conservative surgery, RT administration and combined modality treatment. Multivariate analysis showed that early stage and the use of RT were favorable prognostic factors. CONCLUSION: The outcome of PBL is fair. Local control is excellent with RT or combined modality treatment but systemic relapses, including that in the CNS, occurs frequently.
Resumo:
The final year project came to us as an opportunity to get involved in a topic which has appeared to be attractive during the learning process of majoring in economics: statistics and its application to the analysis of economic data, i.e. econometrics.Moreover, the combination of econometrics and computer science is a very hot topic nowadays, given the Information Technologies boom in the last decades and the consequent exponential increase in the amount of data collected and stored day by day. Data analysts able to deal with Big Data and to find useful results from it are verydemanded in these days and, according to our understanding, the work they do, although sometimes controversial in terms of ethics, is a clear source of value added both for private corporations and the public sector. For these reasons, the essence of this project is the study of a statistical instrument valid for the analysis of large datasets which is directly related to computer science: Partial Correlation Networks.The structure of the project has been determined by our objectives through the development of it. At first, the characteristics of the studied instrument are explained, from the basic ideas up to the features of the model behind it, with the final goal of presenting SPACE model as a tool for estimating interconnections in between elements in large data sets. Afterwards, an illustrated simulation is performed in order to show the power and efficiency of the model presented. And at last, the model is put into practice by analyzing a relatively large data set of real world data, with the objective of assessing whether the proposed statistical instrument is valid and useful when applied to a real multivariate time series. In short, our main goals are to present the model and evaluate if Partial Correlation Network Analysis is an effective, useful instrument and allows finding valuable results from Big Data.As a result, the findings all along this project suggest the Partial Correlation Estimation by Joint Sparse Regression Models approach presented by Peng et al. (2009) to work well under the assumption of sparsity of data. Moreover, partial correlation networks are shown to be a very valid tool to represent cross-sectional interconnections in between elements in large data sets.The scope of this project is however limited, as there are some sections in which deeper analysis would have been appropriate. Considering intertemporal connections in between elements, the choice of the tuning parameter lambda, or a deeper analysis of the results in the real data application are examples of aspects in which this project could be completed.To sum up, the analyzed statistical tool has been proved to be a very useful instrument to find relationships that connect the elements present in a large data set. And after all, partial correlation networks allow the owner of this set to observe and analyze the existing linkages that could have been omitted otherwise.
Resumo:
Combinatorial optimization involves finding an optimal solution in a finite set of options; many everyday life problems are of this kind. However, the number of options grows exponentially with the size of the problem, such that an exhaustive search for the best solution is practically infeasible beyond a certain problem size. When efficient algorithms are not available, a practical approach to obtain an approximate solution to the problem at hand, is to start with an educated guess and gradually refine it until we have a good-enough solution. Roughly speaking, this is how local search heuristics work. These stochastic algorithms navigate the problem search space by iteratively turning the current solution into new candidate solutions, guiding the search towards better solutions. The search performance, therefore, depends on structural aspects of the search space, which in turn depend on the move operator being used to modify solutions. A common way to characterize the search space of a problem is through the study of its fitness landscape, a mathematical object comprising the space of all possible solutions, their value with respect to the optimization objective, and a relationship of neighborhood defined by the move operator. The landscape metaphor is used to explain the search dynamics as a sort of potential function. The concept is indeed similar to that of potential energy surfaces in physical chemistry. Borrowing ideas from that field, we propose to extend to combinatorial landscapes the notion of the inherent network formed by energy minima in energy landscapes. In our case, energy minima are the local optima of the combinatorial problem, and we explore several definitions for the network edges. At first, we perform an exhaustive sampling of local optima basins of attraction, and define weighted transitions between basins by accounting for all the possible ways of crossing the basins frontier via one random move. Then, we reduce the computational burden by only counting the chances of escaping a given basin via random kick moves that start at the local optimum. Finally, we approximate network edges from the search trajectory of simple search heuristics, mining the frequency and inter-arrival time with which the heuristic visits local optima. Through these methodologies, we build a weighted directed graph that provides a synthetic view of the whole landscape, and that we can characterize using the tools of complex networks science. We argue that the network characterization can advance our understanding of the structural and dynamical properties of hard combinatorial landscapes. We apply our approach to prototypical problems such as the Quadratic Assignment Problem, the NK model of rugged landscapes, and the Permutation Flow-shop Scheduling Problem. We show that some network metrics can differentiate problem classes, correlate with problem non-linearity, and predict problem hardness as measured from the performances of trajectory-based local search heuristics.
Resumo:
A mathematical model of the voltage drop which arises in on-chip power distribution networks is used to compare the maximum voltage drop in the case of different geometric arrangements of the pads supplying power to the chip. These include the square or Manhattan power pad arrangement, which currently predominates, as well as equilateral triangular and hexagonal arrangements. In agreement with the findings in the literature and with physical and SPICE models, the equilateral triangular power pad arrangement is found to minimize the maximum voltage drop. This headline finding is a consequence of relatively simple formulas for the voltage drop, with explicit error bounds, which are established using complex analysis techniques, and elliptic functions in particular.
Resumo:
El principal objectiu d'aquest treball és implementar i exposar una descripció teòrica per a diferents esquemes de Physical Layer Network Coding. Utilitzant un esquema bàsic com a punt de partida, el projecte presenta la construcció i l'anàlisis de diferents esquemes de comunicació on la complexitat va augmentant a mesura que anem avançant en el projecte. El treball està estructurat en diferents parts: primer, es presenta una introducció a Physical Layer Network Coding i a Lattice Network Codes. A continuació, s'introdueixen les eines matemàtiques necessàries per entendre el CF System. Després, s'analitza i implementa el primer esquema bàsic. A partir del qual, implementem una versió vectorial del CF System i una versió codificada amb un Hamming q-ari. Finalment, s'estudien i implementen diferents estratègies per millorar la matriu de coeficients A.
Resumo:
Integration of biological data of various types and the development of adapted bioinformatics tools represent critical objectives to enable research at the systems level. The European Network of Excellence ENFIN is engaged in developing an adapted infrastructure to connect databases, and platforms to enable both the generation of new bioinformatics tools and the experimental validation of computational predictions. With the aim of bridging the gap existing between standard wet laboratories and bioinformatics, the ENFIN Network runs integrative research projects to bring the latest computational techniques to bear directly on questions dedicated to systems biology in the wet laboratory environment. The Network maintains internally close collaboration between experimental and computational research, enabling a permanent cycling of experimental validation and improvement of computational prediction methods. The computational work includes the development of a database infrastructure (EnCORE), bioinformatics analysis methods and a novel platform for protein function analysis FuncNet.
Resumo:
VVALOSADE is a research project of professor Anita Lukka's VALORE research team in the Lappeenranta University of Technology. The VALOSADE includes the ELO technology program of Tekes. SMILE is one of four subprojects of the VALOSADE. The SMILE study focuses on the case of the company network that is composed of small and micro-sized mechanical maintenance service providers and forest industry as large-scale customers. The basic principle of the SMILE study is the communication and ebusiness in supply and demand networks. The aim of the study is to develop ebusiness strategy, ebusiness model and e-processes among the SME local service providers, and onthe other hand, between the local service provider network and the forest industry customers in a maintenance and operations service business. A literature review, interviews and benchmarking are used as research methods in this qualitative case study. The first SMILE report, 'Ebusiness between Global Company and Its Local SME Supplier Network', concentrated on creating background for the SMILE study by studying general trends of ebusiness in supply chains and networks of different industries. This second phase of the study concentrates on case network background, such as business relationships, information systems and business objectives; core processes in maintenance and operations service network; development needs in communication among the network participants; and ICT solutions to respond needs in changing environment. In the theory part of the report, different ebusiness models and frameworks are introduced. Those models and frameworks are compared to empirical case data. From that analysis of the empirical data, therecommendations for the development of the network information system are derived. In process industry such as the forest industry, it is crucial to achieve a high level of operational efficiency and reliability, which sets up great requirements for maintenance and operations. Therefore, partnerships or strategic alliances are needed between the network participants. In partnerships and alliances, deep communication is important, and therefore the information systems in the network also are critical. Communication, coordination and collaboration will increase in the case network in the future, because network resources must be optimised to improve competitive capability of the forest industry customers and theefficiency of their service providers. At present, ebusiness systems are not usual in this maintenance network. A network information system among the forest industry customers and their local service providers actually is the only genuinenetwork information system in this total network. However, the utilisation of that system has been quite insignificant. The current system does not add value enough either to the customers or to the local service providers. At present, thenetwork information system is the infomediary that share static information forthe network partners. The network information system should be the transaction intermediary, which integrates internal processes of the network companies; the network information system, which provides common standardised processes for thelocal service providers; and the infomediary, which share static and dynamic information on right time, on right partner, on right costs, on right format and on right quality. This study provides recommendations how to develop this system in the future to add value to the network companies. Ebusiness scenarios, vision, objectives, strategies, application architecture, ebusiness model, core processes and development strategy must be considered when the network information system will be developed in the next development step. The core processes in the case network are demand/capacity management, customer/supplier relationship management, service delivery management, knowledge management and cash flow management. Most benefits from ebusiness solutions come from the electrifying of operational level processes, such as service delivery management and cash flow management.
Resumo:
Perceiving the world visually is a basic act for humans, but for computers it is still an unsolved problem. The variability present innatural environments is an obstacle for effective computer vision. The goal of invariant object recognition is to recognise objects in a digital image despite variations in, for example, pose, lighting or occlusion. In this study, invariant object recognition is considered from the viewpoint of feature extraction. Thedifferences between local and global features are studied with emphasis on Hough transform and Gabor filtering based feature extraction. The methods are examined with respect to four capabilities: generality, invariance, stability, and efficiency. Invariant features are presented using both Hough transform and Gabor filtering. A modified Hough transform technique is also presented where the distortion tolerance is increased by incorporating local information. In addition, methods for decreasing the computational costs of the Hough transform employing parallel processing and local information are introduced.
Resumo:
En este trabajo intentamos comprobar, entre otros factores, la configuración de medios innovadores a escala local, donde se forma una red de empresas, se aplica un saber tradicional, una cultura y un capital social que se adapta perfectamente a los mercados internacionales más competitivos. Concretamente, el municipio objeto de estudio, A Estrada (Pontevedra), presenta un desarrollo económico apoyado en la fabricación de muebles.
Resumo:
It has been convincingly argued that computer simulation modeling differs from traditional science. If we understand simulation modeling as a new way of doing science, the manner in which scientists learn about the world through models must also be considered differently. This article examines how researchers learn about environmental processes through computer simulation modeling. Suggesting a conceptual framework anchored in a performative philosophical approach, we examine two modeling projects undertaken by research teams in England, both aiming to inform flood risk management. One of the modeling teams operated in the research wing of a consultancy firm, the other were university scientists taking part in an interdisciplinary project experimenting with public engagement. We found that in the first context the use of standardized software was critical to the process of improvisation, the obstacles emerging in the process concerned data and were resolved through exploiting affordances for generating, organizing, and combining scientific information in new ways. In the second context, an environmental competency group, obstacles were related to the computer program and affordances emerged in the combination of experience-based knowledge with the scientists' skill enabling a reconfiguration of the mathematical structure of the model, allowing the group to learn about local flooding.
Resumo:
This article explores how to enrich scaffolding processes among university students using specific Computer Supported Collaborative Learning –CSCL- software. A longitudinal case study was designed, in which eighteen students participated in a twelve-month learning project. During this period the students followed an instructional process, using the CSCL software to support and improve the students’ interaction processes, in particular the processes of giving and receiving assistance. Our research analyzed the evolution of the quality of the students’ interaction processes and the students’ learning results. The effects of the students’ participation in the CSCL environment have been described in terms of their development of affective, cognitive and metacognitive learning processes. Our results showed that the specific activities that students performed while working with the CSCL system triggered specific learning processes, which had a positive incidence on their learning results.
Resumo:
Tässä työssä on esitetty sen ohjelmiston kehittämisen prosessi, joka on tarkoitettu annettavien palveluiden valvottavaksi käyttäen prototyyppimallia. Raportti sisältää vaatimusten, kohteisiin suunnatun analyysin ja suunnittelun, realisointiprosessien kuvauksen ja prototyypin testauksen. Ohjelmiston käyttöala – antavien palveluiden valvonta. Vaatimukset sovellukselle analysoitiin ohjelmistomarkkinoiden perusteella sekä ohjelmiston engineeringin periaatteiden mukaisesti. Ohjelmiston prototyyppi on realisoitu käyttäen asiakas-/palvelinhybridimallia sekä ralaatiokantaa. Kehitetty ohjelmisto on tarkoitettu venäläisille tietokonekerhoille, jotka erikoistuvat pelipalvelinten antamiseen.
Resumo:
The networking and digitalization of audio equipment has created a need for control protocols. These protocols offer new services to customers and ensure that the equipment operates correctly. The control protocols used in the computer networks are not directly applicable since embedded systems have resource and cost limitations. In this master's thesis the design and implementation of new loudspeaker control network protocols are presented. The protocol stack was required to be reliable, have short response times, configure the network automatically and support the dynamic addition and removal of loudspeakers. The implemented protocol stack was also required to be as efficient and lightweight as possible because the network nodes are fairly simple and lack processing power. The protocol stack was thoroughly tested, validated and verified. The protocols were formally described using LOTOS (Language of Temporal Ordering Specifications) and verified using reachability analysis. A prototype of the loudspeaker network was built and used for testing the operation and the performance of the control protocols. The implemented control protocol stack met the design specifications and proved to be highly reliable and efficient.
Resumo:
La International Network for the Availability of Scientific Publications (INASP) es una ONG de cooperación al desarrollo, que persigue la mejora del acceso, producción y uso de la información científica, de forma que cada país cuente con recursos en este ámbito para afrontar sus propios retos de desarrollo. Surgió en 1992 por iniciativa del ICSU (International Council of Scientific Unions, actualmente International Council for Science) y contó con el apoyo de UNESCO, TWAS-The Academy of Sciences for the Developing World y AAAS (American Association for the Advancement of Science).
Resumo:
PURPOSE: Small cell carcinomas of the bladder (SCCB) account for fewer than 1% of all urinary bladder tumors. There is no consensus regarding the optimal treatment for SCCB. METHODS AND MATERIALS: Fifteen academic Rare Cancer Network medical centers contributed SCCB cases. The eligibility criteria were as follows: pure or mixed SCC; local, locoregional, and metastatic stages; and age ≥18 years. The overall survival (OS) and disease-free survival (DFS) were calculated from the date of diagnosis according to the Kaplan-Meier method. The log-rank and Wilcoxon tests were used to analyze survival as functions of clinical and therapeutic factors. RESULTS: The study included 107 patients (mean [±standard deviation, SD] age, 69.6 [±10.6] years; mean follow-up time, 4.4 years) with primary bladder SCC, with 66% of these patients having pure SCC. Seventy-two percent and 12% of the patients presented with T2-4N0M0 and T2-4N1-3M0 stages, respectively, and 16% presented with synchronous metastases. The most frequent curative treatments were radical surgery and chemotherapy, sequential chemotherapy and radiation therapy, and radical surgery alone. The median (interquartile range, IQR) OS and DFS times were 12.9 months (IQR, 7-32 months) and 9 months (IQR, 5-23 months), respectively. The metastatic, T2-4N0M0, and T2-4N1-3M0 groups differed significantly (P=.001) in terms of median OS and DFS. In a multivariate analysis, impaired creatinine clearance (OS and DFS), clinical stage (OS and DFS), a Karnofsky performance status <80 (OS), and pure SCC histology (OS) were independent and significant adverse prognostic factors. In the patients with nonmetastatic disease, the type of treatment (ie radical surgery with or without adjuvant chemotherapy vs conservative treatment) did not significantly influence OS or DFS (P=.7). CONCLUSIONS: The prognosis for SCCB remains poor. The finding that radical cystectomy did not influence DFS or OS in the patients with nonmetastatic disease suggests that conservative treatment is appropriate in this situation.