8 resultados para software implementation
em BORIS: Bern Open Repository and Information System - Berna - Suiça
Resumo:
Service providers make use of cost-effective wireless solutions to identify, localize, and possibly track users using their carried MDs to support added services, such as geo-advertisement, security, and management. Indoor and outdoor hotspot areas play a significant role for such services. However, GPS does not work in many of these areas. To solve this problem, service providers leverage available indoor radio technologies, such as WiFi, GSM, and LTE, to identify and localize users. We focus our research on passive services provided by third parties, which are responsible for (i) data acquisition and (ii) processing, and network-based services, where (i) and (ii) are done inside the serving network. For better understanding of parameters that affect indoor localization, we investigate several factors that affect indoor signal propagation for both Bluetooth and WiFi technologies. For GSM-based passive services, we developed first a data acquisition module: a GSM receiver that can overhear GSM uplink messages transmitted by MDs while being invisible. A set of optimizations were made for the receiver components to support wideband capturing of the GSM spectrum while operating in real-time. Processing the wide-spectrum of the GSM is possible using a proposed distributed processing approach over an IP network. Then, to overcome the lack of information about tracked devices’ radio settings, we developed two novel localization algorithms that rely on proximity-based solutions to estimate in real environments devices’ locations. Given the challenging indoor environment on radio signals, such as NLOS reception and multipath propagation, we developed an original algorithm to detect and remove contaminated radio signals before being fed to the localization algorithm. To improve the localization algorithm, we extended our work with a hybrid based approach that uses both WiFi and GSM interfaces to localize users. For network-based services, we used a software implementation of a LTE base station to develop our algorithms, which characterize the indoor environment before applying the localization algorithm. Experiments were conducted without any special hardware, any prior knowledge of the indoor layout or any offline calibration of the system.
Resumo:
Service providers make use of cost-effective wireless solutions to identify, localize, and possibly track users using their carried MDs to support added services, such as geo-advertisement, security, and management. Indoor and outdoor hotspot areas play a significant role for such services. However, GPS does not work in many of these areas. To solve this problem, service providers leverage available indoor radio technologies, such as WiFi, GSM, and LTE, to identify and localize users. We focus our research on passive services provided by third parties, which are responsible for (i) data acquisition and (ii) processing, and network-based services, where (i) and (ii) are done inside the serving network. For better understanding of parameters that affect indoor localization, we investigate several factors that affect indoor signal propagation for both Bluetooth and WiFi technologies. For GSM-based passive services, we developed first a data acquisition module: a GSM receiver that can overhear GSM uplink messages transmitted by MDs while being invisible. A set of optimizations were made for the receiver components to support wideband capturing of the GSM spectrum while operating in real-time. Processing the wide-spectrum of the GSM is possible using a proposed distributed processing approach over an IP network. Then, to overcome the lack of information about tracked devices’ radio settings, we developed two novel localization algorithms that rely on proximity-based solutions to estimate in real environments devices’ locations. Given the challenging indoor environment on radio signals, such as NLOS reception and multipath propagation, we developed an original algorithm to detect and remove contaminated radio signals before being fed to the localization algorithm. To improve the localization algorithm, we extended our work with a hybrid based approach that uses both WiFi and GSM interfaces to localize users. For network-based services, we used a software implementation of a LTE base station to develop our algorithms, which characterize the indoor environment before applying the localization algorithm. Experiments were conducted without any special hardware, any prior knowledge of the indoor layout or any offline calibration of the system.
Resumo:
Software dependencies play a vital role in programme comprehension, change impact analysis and other software maintenance activities. Traditionally, these activities are supported by source code analysis; however, the source code is sometimes inaccessible or difficult to analyse, as in hybrid systems composed of source code in multiple languages using various paradigms (e.g. object-oriented programming and relational databases). Moreover, not all stakeholders have adequate knowledge to perform such analyses. For example, non-technical domain experts and consultants raise most maintenance requests; however, they cannot predict the cost and impact of the requested changes without the support of the developers. We propose a novel approach to predicting software dependencies by exploiting the coupling present in domain-level information. Our approach is independent of the software implementation; hence, it can be used to approximate architectural dependencies without access to the source code or the database. As such, it can be applied to hybrid systems with heterogeneous source code or legacy systems with missing source code. In addition, this approach is based solely on information visible and understandable to domain users; therefore, it can be efficiently used by domain experts without the support of software developers. We evaluate our approach with a case study on a large-scale enterprise system, in which we demonstrate how up to 65 of the source code dependencies and 77% of the database dependencies are predicted solely based on domain information.
Resumo:
Software visualizations can provide a concise overview of a complex software system. Unfortunately, as software has no physical shape, there is no `natural' mapping of software to a two-dimensional space. As a consequence most visualizations tend to use a layout in which position and distance have no meaning, and consequently layout typically diverges from one visualization to another. We propose an approach to consistent layout for software visualization, called Software Cartography, in which the position of a software artifact reflects its vocabulary, and distance corresponds to similarity of vocabulary. We use Latent Semantic Indexing (LSI) to map software artifacts to a vector space, and then use Multidimensional Scaling (MDS) to map this vector space down to two dimensions. The resulting consistent layout allows us to develop a variety of thematic software maps that express very different aspects of software while making it easy to compare them. The approach is especially suitable for comparing views of evolving software, as the vocabulary of software artifacts tends to be stable over time. We present a prototype implementation of Software Cartography, and illustrate its use with practical examples from numerous open-source case studies.
Resumo:
Features encapsulate the domain knowledge of a software system and thus are valuable sources of information for a reverse engineer. When analyzing the evolution of a system, we need to know how and which features were modified to recover both the change intention and its extent, namely which source artifacts are affected. Typically, the implementation of a feature crosscuts a number of source artifacts. To obtain a mapping between features to the source artifacts, we exercise the features and capture their execution traces. However this results in large traces that are difficult to interpret. To tackle this issue we compact the traces into simple sets of source artifacts that participate in a feature's runtime behavior. We refer to these compacted traces as feature views. Within a feature view, we partition the source artifacts into disjoint sets of characterized software entities. The characterization defines the level of participation of a source entity in the features. We then analyze the features over several versions of a system and we plot their evolution to reveal how and hich features were affected by changes in the code. We show the usefulness of our approach by applying it to a case study where we address the problem of merging parallel development tracks of the same system.
Resumo:
As more and more open-source software components become available on the internet we need automatic ways to label and compare them. For example, a developer who searches for reusable software must be able to quickly gain an understanding of retrieved components. This understanding cannot be gained at the level of source code due to the semantic gap between source code and the domain model. In this paper we present a lexical approach that uses the log-likelihood ratios of word frequencies to automatically provide labels for software components. We present a prototype implementation of our labeling/comparison algorithm and provide examples of its application. In particular, we apply the approach to detect trends in the evolution of a software system.
Resumo:
In the long run, the widespread use of slide scanners by pathologists requires an adaptation of teaching methods in histology and cytology in order to target these new possibilities of image processing and presentation via the internet. Accordingly, we were looking for a tool with the possibility to teach microscopic anatomy, histology, and cytology of tissue samples which would be able to combine image data from light and electron microscopes independently of microscope suppliers. With the example of a section through the villus of jejunum, we describe here how to process image data from light and electron microscopes in order to get one image-stack which allows a correlation of structures from the microscopic anatomic to the cytological level. With commercially available image-presentation software that we adapted to our needs, we present here a platform which allows for the presentation of this new but also of older material independently of microscope suppliers.
Resumo:
Peritoneal transport characteristics and residual renal function require regular control and subsequent adjustment of the peritoneal dialysis (PD) prescription. Prescription models shall facilitate the prediction of the outcome of such adaptations for a given patient. In the present study, the prescription model implemented in the PatientOnLine software was validated in patients requiring a prescription change. This multicenter, international prospective cohort study with the aim to validate a PD prescription model included patients treated with continuous ambulatory peritoneal dialysis. Patients were examined with the peritoneal function test (PFT) to determine the outcome of their current prescription and the necessity for a prescription change. For these patients, a new prescription was modeled using the PatientOnLine software (Fresenius Medical Care, Bad Homburg, Germany). Two to four weeks after implementation of the new PD regimen, a second PFT was performed. The validation of the prescription model included 54 patients. Predicted and measured peritoneal Kt/V were 1.52 ± 0.31 and 1.66 ± 0.35, and total (peritoneal + renal) Kt/V values were 1.96 ± 0.48 and 2.06 ± 0.44, respectively. Predicted and measured peritoneal creatinine clearances were 42.9 ± 8.6 and 43.0 ± 8.8 L/1.73 m2/week and total creatinine clearances were 65.3 ± 26.0 and 63.3 ± 21.8 L/1.73 m2/week, respectively. The analysis revealed a Pearson's correlation coefficient for peritoneal Kt/V of 0.911 and Lin's concordance coefficient of 0.829. The value of both coefficients was 0.853 for peritoneal creatinine clearance. Predicted and measured daily net ultrafiltration was 0.77 ± 0.49 and 1.16 ± 0.63 L/24 h, respectively. Pearson's correlation and Lin's concordance coefficient were 0.518 and 0.402, respectively. Predicted and measured peritoneal glucose absorption was 125.8 ± 38.8 and 79.9 ± 30.7 g/24 h, respectively, and Pearson's correlation and Lin's concordance coefficient were 0.914 and 0.477, respectively. With good predictability of peritoneal Kt/V and creatinine clearance, the present model provides support for individual dialysis prescription in clinical practice. Peritoneal glucose absorption and ultrafiltration are less predictable and are likely to be influenced by additional clinical factors to be taken into consideration.