834 resultados para Network Analysis Methods
Resumo:
Current debate within forensic authorship analysis has tended to polarise those who argue that analysis methods should reflect a strong cognitive theory of idiolect and others who see less of a need to look behind the stylistic variation of the texts they are examining. This chapter examines theories of idiolect and asks how useful or necessary they are to the practice of forensic authorship analysis. Taking a specific text messaging case the chapter demonstrates that methodologically rigorous, theoretically informed authorship analysis need not appeal to cognitive theories of idiolect in order to be valid. By considering text messaging forensics, lessons will be drawn which can contribute to wider debates on the role of theories of idiolect in forensic casework.
Resumo:
This paper analyzes the theme of knowledge transfer in supply chain management. The aim of this study is to present the social network analysis (SNA) as an useful tool to study knowledge networks within supply chain, to monitor knowledge flows and to identify the accumulating knowledge nodes of the networks.
Resumo:
The article proposes the model of management of information about program flow analysis for conducting computer experiments with program transformations. It considers the architecture and context of the flow analysis subsystem within the framework of Specialized Knowledge Bank on Program Transformations and describes the language for presenting flow analysis methods in the knowledge bank.
Resumo:
Objective In this study, we have used a chemometrics-based method to correlate key liposomal adjuvant attributes with in-vivo immune responses based on multivariate analysis. Methods The liposomal adjuvant composed of the cationic lipid dimethyldioctadecylammonium bromide (DDA) and trehalose 6,6-dibehenate (TDB) was modified with 1,2-distearoyl-sn-glycero-3-phosphocholine at a range of mol% ratios, and the main liposomal characteristics (liposome size and zeta potential) was measured along with their immunological performance as an adjuvant for the novel, postexposure fusion tuberculosis vaccine, Ag85B-ESAT-6-Rv2660c (H56 vaccine). Partial least square regression analysis was applied to correlate and cluster liposomal adjuvants particle characteristics with in-vivo derived immunological performances (IgG, IgG1, IgG2b, spleen proliferation, IL-2, IL-5, IL-6, IL-10, IFN-γ). Key findings While a range of factors varied in the formulations, decreasing the 1,2-distearoyl-sn-glycero-3-phosphocholine content (and subsequent zeta potential) together built the strongest variables in the model. Enhanced DDA and TDB content (and subsequent zeta potential) stimulated a response skewed towards a cell mediated immunity, with the model identifying correlations with IFN-γ, IL-2 and IL-6. Conclusion This study demonstrates the application of chemometrics-based correlations and clustering, which can inform liposomal adjuvant design.
Resumo:
This paper seeks to advance the theory and practice of the dynamics of complex networks in relation to direct and indirect citations. It applies social network analysis (SNA) and the ordered weighted averaging operator (OWA) to study a patent citations network. So far the SNA studies investigating long chains of patents citations have rarely been undertaken and the importance of a node in a network has been associated mostly with its number of direct ties. In this research OWA is used to analyse complex networks, assess the role of indirect ties, and provide guidance to reduce complexity for decision makers and analysts. An empirical example of a set of European patents published in 2000 in the renewable energy industry is provided to show the usefulness of the proposed approach for the preference ranking of patent citations.
Resumo:
Purpose: To assess the validity and repeatability of objective compared to subjective contact lens fit analysis. Methods: Thirty-five subjects (aged 22.0. ±. 3.0 years) wore two different soft contact lens designs. Four lens fit variables: centration, horizontal lag, post-blink movement in up-gaze and push-up recovery speed were assessed subjectively (four observers) and objectively from slit-lamp biomicroscopy captured images and video. The analysis was repeated a week later. Results: The average of the four experienced observers was compared to objective measures, but centration, movement on blink, lag and push-up recovery speed all varied significantly between them (p <. 0.001). Horizontal lens centration was on average close to central as assessed both objectively and subjectively (p > 0.05). The 95% confidence interval of subjective repeatability was better than objective assessment (±0.128. mm versus ±0.168. mm, p = 0.417), but utilised only 78% of the objective range. Vertical centration assessed objectively showed a slight inferior decentration (0.371. ±. 0.381. mm) with good inter- and intrasession repeatability (p > 0.05). Movement-on-blink was lower estimated subjectively than measured objectively (0.269. ±. 0.179. mm versus 0.352. ±. 0.355. mm; p = 0.035), but had better repeatability (±0.124. mm versus ±0.314. mm 95% confidence interval) unless correcting for the smaller range (47%). Horizontal lag was lower estimated subjectively (0.562. ±. 0.259. mm) than measured objectively (0.708. ±. 0.374. mm, p <. 0.001), had poorer repeatability (±0.132. mm versus ±0.089. mm 95% confidence interval) and had a smaller range (63%). Subjective categorisation of push-up speed of recovery showed reasonable differentiation relative to objective measurement (p <. 0.001). Conclusions: The objective image analysis allows an accurate, reliable and repeatable assessment of soft contact lens fit characteristics, being a useful tool for research and optimisation of lens fit in clinical practice.
Resumo:
This paper looks at potential distribution network stability problems under the Smart Grid scenario. This is to consider distributed energy resources (DERs) e.g. renewable power generations and intelligent loads with power-electronic controlled converters. The background of this topic is introduced and potential problems are defined from conventional power system stability and power electronic system stability theories. Challenges are identified with possible solutions from steady-state limits, small-signal, and large-signal stability indexes and criteria. Parallel computation techniques might be included for simulation or simplification approaches are required for a largescale distribution network analysis.
Resumo:
Data fluctuation in multiple measurements of Laser Induced Breakdown Spectroscopy (LIBS) greatly affects the accuracy of quantitative analysis. A new LIBS quantitative analysis method based on the Robust Least Squares Support Vector Machine (RLS-SVM) regression model is proposed. The usual way to enhance the analysis accuracy is to improve the quality and consistency of the emission signal, such as by averaging the spectral signals or spectrum standardization over a number of laser shots. The proposed method focuses more on how to enhance the robustness of the quantitative analysis regression model. The proposed RLS-SVM regression model originates from the Weighted Least Squares Support Vector Machine (WLS-SVM) but has an improved segmented weighting function and residual error calculation according to the statistical distribution of measured spectral data. Through the improved segmented weighting function, the information on the spectral data in the normal distribution will be retained in the regression model while the information on the outliers will be restrained or removed. Copper elemental concentration analysis experiments of 16 certified standard brass samples were carried out. The average value of relative standard deviation obtained from the RLS-SVM model was 3.06% and the root mean square error was 1.537%. The experimental results showed that the proposed method achieved better prediction accuracy and better modeling robustness compared with the quantitative analysis methods based on Partial Least Squares (PLS) regression, standard Support Vector Machine (SVM) and WLS-SVM. It was also demonstrated that the improved weighting function had better comprehensive performance in model robustness and convergence speed, compared with the four known weighting functions.
Resumo:
Background: To examine the views and current practice of SMBG among Black Caribbean and South Asian individuals with non-insulin treated Type 2 diabetes mellitus. Methods: Twelve participants completed semi-structured interviews that were guided by the Health Belief Model and analyzed using thematic network analysis. Results: The frequency of monitoring among participants varied from several times a day to once per week. Most participants expressed similar experiences regarding their views and practices of SMBG. Minor differences across gender and culture were observed. All participants understood the benefits, but not all viewed SMBG as beneficial to their personal diabetes management. SMBG can facilitate a better understanding and maintenance of self-care behaviours. However, it can trigger both positive and negative emotional responses, such as a sense of disappointment when high readings are not anticipated, resulting in emotional distress. Health care professionals play a key role in the way SMBG is perceived and used by patients. Conclusion: While the majority of participants value SMBG as a self-management tool, barriers exist that impede its practice, particularly its cost. How individuals cope with these barriers is integral to understanding why some patients adopt SMBG more than others. © 2013 Gucciardi et al.; licensee BioMed Central Ltd.
Resumo:
2000 Mathematics Subject Classi cation: Primary 90C31. Secondary 62C12, 62P05, 93C41.
Resumo:
OBJECTIVES: To understand older adults' experiences of moving into extra care housing which offers enrichment activities alongside social and healthcare support. DESIGN: A longitudinal study was conducted which adopted a phenomenological approach to data generation and analysis. METHODS: Semi-structured interviews were conducted in the first 18 months of living in extra care housing. Interpretative phenomenological analysis was used because its commitment to idiography enabled an in-depth analysis of the subjective lived experience of moving into extra care housing. Themes generated inductively were examined against an existential-phenomenological theory of well-being. RESULTS: Learning to live in an extra care community showed negotiating new relationships was not straightforward; maintaining friendships outside the community became more difficult as capacity declined. In springboard for opportunity/confinement, living in extra care provided new opportunities for social engagement and a restored sense of self. Over time horizons began to shrink as incapacities grew. Seeking care illustrated reticence to seek care, due to embarrassment and a sense of duty to one's partner. Becoming aged presented an ontological challenge. Nevertheless, some showed a readiness for death, a sense of homecoming. CONCLUSIONS: An authentic later life was possible but residents required emotional and social support to live through the transition and challenges of becoming aged. Enhancement activities boosted residents' quality of life but the range of activities could be extended to cater better for quieter, smaller scale events within the community; volunteer activity facilitators could be used here. Peer mentoring may help build new relationships and opportunities for interactive stimulation. Acknowledging the importance of feeling-empathic imagination-in caregiving may help staff and residents relate better to each other, thus helping individuals to become ontologically secure and live well to the end.
Resumo:
A tárgyalófelek elé kitett mobiltelefon alkalmazása előrejelzi a beszélgetőpartnerek versenyképességét a versenyképesség-mutatók alapján, javaslatokat adva a tárgyalás további menetére. Ez a vízió nyilván még futurisztikus, ám a csúcsvezetői nyilatkozatok rejtett szövegtartalma alapján következtetéseket levonni a képviselt szervezetek versenyképességi orientációira – ez már ma lehetőség. A GLOBE-projekt kultúrakutatási módszertanával, valamint szövegelemzési módszerekkel sikerült kimutatni a versenyképességet előrejelző hatalmi távolság és az intézményi kollektivizmus szövegbeli jeleit. Mindez eszközt jelenthet egyebek mellett a szervezetfejlesztéssel, hírszerzéssel, HR-gazdálkodással foglalkozó szakembereknek is. _______ The use of the mobile telephones laid in front of the negotiators during their conversations forecasts their indicators of competitiveness and gives suggestions for the further course of negotiation. This is obviously a futuristic vision, but drawing conclusions from the hidden content of top management narratives concerning the competitive cultural orientations of the represented organizations is a possibility that is already available. Using the culture research methodology of the GLOBE project as well as text analysis methods, it was possible to reveal narrative patterns both of the power distance, forecasting competitiveness, and of institutional collectivism. These findings may be useful tools for professionals, among others of organizational development, intelligence service and HR management.
Resumo:
Több mint száz éve született meg Henry Gantt (Gantt, 1910) sávos ütemterve, Kelley (Kelley, 1961) és Walker (Walker, 1959) is több mint hatvan éve publikálta kritikus út módszerét. Az ezekre épülő költség- és erőforrás- tervezési módszerek vajon alkalmasak-e a ma kihívásaira? Az olvasó ebben a tanulmányban többéves kutatómunka gyümölcsét láthatja. A kutatás során az egyik legfontosabb cél annak vizsgálata volt, hogy a meglévő projekttervezési eszközök mennyiben felelnek meg a mai projektek kihívásainak; hol és milyen területen van szükség e módszerek továbbfejlesztésére, esetleg meghaladására. Ebben a tanulmányban a szerző olyan módszereket mutat be, amelyek messze túlvezetnek bennünket a projekttervezés eddig elsősorban operatív feladatokra szorítkozó módszereitől, és olyan kérdések megválaszolására fordítja figyelmünket, mint pl. milyen tevékenységeket, projekteket valósítsunk meg; melyeket hagyjuk el vagy ütemezzük be egy későbbi projektbe; hogyan rangsoroljuk, priorizáljuk a projektek megvalósítását, fontosságát? ______ Gantt chart (Gantt, 1910) was born by Henry Gantt more than a hundred years ago. Kelley and Walker published their critical planning method more than a 60 years ago (see i.e. Kelley-Walker, 1959). Can we use methods based on network planning methods for the challenges of 21st century? In this paper the author can see the results of the recent researches. In this study with their colleagues he investigated which project planning methods can be used in challenges of the 21st century and where and how to improve them. In these researches new matrix-based project planning methods are specified, where they can deal not only operative but strategic questions: which subprojects/tasks should be completed, how to treat priorities of completion in case of defining logic planning, how to support not only traditional but agile project management approaches.In this paper he introduces a new matrix-based method, which can be used for ranking project or multi project scenarios with different kinds of target functions. The author shows methods that are used in an expert module. He shows how to integrate this expert module into the traditional PMS system.
Resumo:
With the latest development in computer science, multivariate data analysis methods became increasingly popular among economists. Pattern recognition in complex economic data and empirical model construction can be more straightforward with proper application of modern softwares. However, despite the appealing simplicity of some popular software packages, the interpretation of data analysis results requires strong theoretical knowledge. This book aims at combining the development of both theoretical and applicationrelated data analysis knowledge. The text is designed for advanced level studies and assumes acquaintance with elementary statistical terms. After a brief introduction to selected mathematical concepts, the highlighting of selected model features is followed by a practice-oriented introduction to the interpretation of SPSS1 outputs for the described data analysis methods. Learning of data analysis is usually time-consuming and requires efforts, but with tenacity the learning process can bring about a significant improvement of individual data analysis skills.
Resumo:
The Internet has become an integral part of our nation’s critical socio-economic infrastructure. With its heightened use and growing complexity however, organizations are at greater risk of cyber crimes. To aid in the investigation of crimes committed on or via the Internet, a network forensics analysis tool pulls together needed digital evidence. It provides a platform for performing deep network analysis by capturing, recording and analyzing network events to find out the source of a security attack or other information security incidents. Existing network forensics work has been mostly focused on the Internet and fixed networks. But the exponential growth and use of wireless technologies, coupled with their unprecedented characteristics, necessitates the development of new network forensic analysis tools. This dissertation fostered the emergence of a new research field in cellular and ad-hoc network forensics. It was one of the first works to identify this problem and offer fundamental techniques and tools that laid the groundwork for future research. In particular, it introduced novel methods to record network incidents and report logged incidents. For recording incidents, location is considered essential to documenting network incidents. However, in network topology spaces, location cannot be measured due to absence of a ‘distance metric’. Therefore, a novel solution was proposed to label locations of nodes within network topology spaces, and then to authenticate the identity of nodes in ad hoc environments. For reporting logged incidents, a novel technique based on Distributed Hash Tables (DHT) was adopted. Although the direct use of DHTs for reporting logged incidents would result in an uncontrollably recursive traffic, a new mechanism was introduced that overcome this recursive process. These logging and reporting techniques aided forensics over cellular and ad-hoc networks, which in turn increased their ability to track and trace attacks to their source. These techniques were a starting point for further research and development that would result in equipping future ad hoc networks with forensic components to complement existing security mechanisms.