929 resultados para Remote Data Acquisition and Storage


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This dissertation is a discourse on the capital market and its interactive framework of acquisition and issuance of financial assets that drive the economy from both sides—investors/lenders and issuers/users of capital assets. My work consists of four essays in financial economics that offer a spectrum of revisions to this significant area of study. The first essay is a delineation of the capital market over the past half a century and major developments on capital markets on issues that pertain to the investor's opportunity set and the corporation's capital-raising availability set. This chapter should have merits on two counts: (i) a comprehensive account of capital markets and return-generating assets and (ii) a backdrop against which I present my findings in Chapters 2 through 4. ^ In Chapter 2, I rework on the Markowitz-Roy-Tobin structure of the efficient frontier and of the Separation Theorem. Starting off with a 2-asset portfolio and extending the paradigm to an n-asset portfolio, I bring out the optimal choice of assets for an investor under constrained utility maximization. In this chapter, I analyze the selection and revision-theoretic construct and bring out optimum choices. The effect of a change in perceived risk or return in the mind of an investor is ascertained on the portfolio composition. ^ Chapter 3 takes a look into corporations that issue market securities. The question of how a corporation decides what kinds of securities it should issue in the marketplace to raise funds brings out the classic value invariance proposition of Modigliani and Miller and fills the gap that existed in the literature for almost half a century. I question the general validity in the classic results of Modigliani and Miller and modify the existing literature on the celebrated value invariance proposition. ^ Chapter 4 takes the Modigliani-Miller regime to its correct prescription in the presence of corporate and personal taxes. I show that Modigliani-Miller's age-old proposition needs corrections and extensions, which I derive. ^ My dissertation overall brings all of these corrections and extensions to the existing literature as my findings, showing that capital markets are in an ever-changing state of necessary revision. ^

Relevância:

100.00% 100.00%

Publicador:

Resumo:

With the explosive growth of the volume and complexity of document data (e.g., news, blogs, web pages), it has become a necessity to semantically understand documents and deliver meaningful information to users. Areas dealing with these problems are crossing data mining, information retrieval, and machine learning. For example, document clustering and summarization are two fundamental techniques for understanding document data and have attracted much attention in recent years. Given a collection of documents, document clustering aims to partition them into different groups to provide efficient document browsing and navigation mechanisms. One unrevealed area in document clustering is that how to generate meaningful interpretation for the each document cluster resulted from the clustering process. Document summarization is another effective technique for document understanding, which generates a summary by selecting sentences that deliver the major or topic-relevant information in the original documents. How to improve the automatic summarization performance and apply it to newly emerging problems are two valuable research directions. To assist people to capture the semantics of documents effectively and efficiently, the dissertation focuses on developing effective data mining and machine learning algorithms and systems for (1) integrating document clustering and summarization to obtain meaningful document clusters with summarized interpretation, (2) improving document summarization performance and building document understanding systems to solve real-world applications, and (3) summarizing the differences and evolution of multiple document sources.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Due to the rapid advances in computing and sensing technologies, enormous amounts of data are being generated everyday in various applications. The integration of data mining and data visualization has been widely used to analyze these massive and complex data sets to discover hidden patterns. For both data mining and visualization to be effective, it is important to include the visualization techniques in the mining process and to generate the discovered patterns for a more comprehensive visual view. In this dissertation, four related problems: dimensionality reduction for visualizing high dimensional datasets, visualization-based clustering evaluation, interactive document mining, and multiple clusterings exploration are studied to explore the integration of data mining and data visualization. In particular, we 1) propose an efficient feature selection method (reliefF + mRMR) for preprocessing high dimensional datasets; 2) present DClusterE to integrate cluster validation with user interaction and provide rich visualization tools for users to examine document clustering results from multiple perspectives; 3) design two interactive document summarization systems to involve users efforts and generate customized summaries from 2D sentence layouts; and 4) propose a new framework which organizes the different input clusterings into a hierarchical tree structure and allows for interactive exploration of multiple clustering solutions.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

long-term research on freshwater ecosystems provides insights that can be difficult to obtain from other approaches. Widespread monitoring of ecologically relevant water-quality parameters spanning decades can facilitate important tests of ecological principles. Unique long-term data sets and analytical tools are increasingly available, allowing for powerful and synthetic analyses across sites. long-term measurements or experiments in aquatic systems can catch rare events, changes in highly variable systems, time-lagged responses, cumulative effects of stressors, and biotic responses that encompass multiple generations. Data are available from formal networks, local to international agencies, private organizations, various institutions, and paleontological and historic records; brief literature surveys suggest much existing data are not synthesized. Ecological sciences will benefit from careful maintenance and analyses of existing long-term programs, and subsequent insights can aid in the design of effective future long-term experimental and observational efforts. long-term research on freshwaters is particularly important because of their value to humanity.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper details the research methods an introductory qualitative research class used to both study an issue related to race and identity, and to familiarize themselves with data collection strategies. Throughout the paper the authors attempt to capture the challenges, disagreements, and consensus building that marked this unusual research endeavor.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Carbon capture and storage (CCS) can contribute significantly to addressing the global greenhouse gas (GHG) emissions problem. Despite widespread political support, CCS remains unknown to the general public. Public perception researchers have found that, when asked, the public is relatively unfamiliar with CCS yet many individuals voice specific safety concerns regarding the technology. We believe this leads many stakeholders conflate CCS with the better-known and more visible technology hydraulic fracturing (fracking). We support this with content analysis of media coverage, web analytics, and public lobbying records. Furthermore, we present results from a survey of United States residents. This first-of-its-kind survey assessed participants’ knowledge, opinions and support of CCS and fracking technologies. The survey showed that participants had more knowledge of fracking than CCS, and that knowledge of fracking made participants less willing to support CCS projects. Additionally, it showed that participants viewed the two technologies as having similar risks and similar risk intensities. In the CCS stakeholder literature, judgment and decision-making (JDM) frameworks are noticeably absent, and public perception is not discussed using any cognitive biases as a way of understanding or explaining irrational decisions, yet these survey results show evidence of both anchoring bias and the ambiguity effect. Public acceptance of CCS is essential for a national low-carbon future plan. In conclusion, we propose changes in communications and incentives as programs to increase support of CCS.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This dissertation presents a calibration procedure for a pressure velocity probe. The dissertation is divided into four main chapters. The first chapter is divided into six main sections. In the firsts two, the wave equation in fluids and the velocity of sound in gases are calculated, the third section contains a general solution of the wave equation in the case of plane acoustic waves. Section four and five report the definition of the acoustic impedance and admittance, and the practical units the sound level is measured with, i.e. the decibel scale. Finally, the last section of the chapter is about the theory linked to the frequency analysis of a sound wave and includes the analysis of sound in bands and the discrete Fourier analysis, with the definition of some important functions. The second chapter describes different reference field calibration procedures that are used to calibrate the P-V probes, between them the progressive plane wave method, which is that has been used in this work. Finally, the last section of the chapter contains a description of the working principles of the two transducers that have been used, with a focus on the velocity one. The third chapter of the dissertation is devoted to the explanation of the calibration set up and the instruments used for the data acquisition and analysis. Since software routines were extremely important, this chapter includes a dedicated section on them and the proprietary routines most used are thoroughly explained. Finally, there is the description of the work that has been done, which is identified with three different phases, where the data acquired and the results obtained are presented. All the graphs and data reported were obtained through the Matlab® routine. As for the last chapter, it briefly presents all the work that has been done as well as an excursus on a new probe and on the way the procedure implemented in this dissertation could be applied in the case of a general field.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The program PanTool was developed as a tool box like a Swiss Army Knife for data conversion and recalculation, written to harmonize individual data collections to standard import format used by PANGAEA. The format of input files the program PanTool needs is a tabular saved in plain ASCII. The user can create this files with a spread sheet program like MS-Excel or with the system text editor. PanTool is distributed as freeware for the operating systems Microsoft Windows, Apple OS X and Linux.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Research on temporal-order perception uses temporal-order judgment (TOJ) tasks or synchrony judgment (SJ) tasks in their binary SJ2 or ternary SJ3 variants. In all cases, two stimuli are presented with some temporal delay, and observers judge the order of presentation. Arbitrary psychometric functions are typically fitted to obtain performance measures such as sensitivity or the point of subjective simultaneity, but the parameters of these functions are uninterpretable. We describe routines in MATLAB and R that fit model-based functions whose parameters are interpretable in terms of the processes underlying temporal-order and simultaneity judgments and responses. These functions arise from an independent-channels model assuming arrival latencies with exponential distributions and a trichotomous decision space. Different routines fit data separately for SJ2, SJ3, and TOJ tasks, jointly for any two tasks, or also jointly for the three tasks (for common cases in which two or even the three tasks were used with the same stimuli and participants). Additional routines provide bootstrap p-values and confidence intervals for estimated parameters. A further routine is included that obtains performance measures from the fitted functions. An R package for Windows and source code of the MATLAB and R routines are available as Supplementary Files.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Funding and trial registration: Scottish Government Chief Scientist Office grant CZH/3/17. ClinicalTrials.gov registration NCT01602705.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this work, we present an adaptive unequal loss protection (ULP) scheme for H264/AVC video transmission over lossy networks. This scheme combines erasure coding, H.264/AVC error resilience techniques and importance measures in video coding. The unequal importance of the video packets is identified in the group of pictures (GOP) and the H.264/AVC data partitioning levels. The presented method can adaptively assign unequal amount of forward error correction (FEC) parity across the video packets according to the network conditions, such as the available network bandwidth, packet loss rate and average packet burst loss length. A near optimal algorithm is developed to deal with the FEC assignment for optimization. The simulation results show that our scheme can effectively utilize network resources such as bandwidth, while improving the quality of the video transmission. In addition, the proposed ULP strategy ensures graceful degradation of the received video quality as the packet loss rate increases. © 2010 IEEE.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The realization of an energy future based on safe, clean, sustainable, and economically viable technologies is one of the grand challenges facing modern society. Electrochemical energy technologies underpin the potential success of this effort to divert energy sources away from fossil fuels, whether one considers alternative energy conversion strategies through photoelectrochemical (PEC) production of chemical fuels or fuel cells run with sustainable hydrogen, or energy storage strategies, such as in batteries and supercapacitors. This dissertation builds on recent advances in nanomaterials design, synthesis, and characterization to develop novel electrodes that can electrochemically convert and store energy.

Chapter 2 of this dissertation focuses on refining the properties of TiO2-based PEC water-splitting photoanodes used for the direct electrochemical conversion of solar energy into hydrogen fuel. The approach utilized atomic layer deposition (ALD); a growth process uniquely suited for the conformal and uniform deposition of thin films with angstrom-level thickness precision. ALD’s thickness control enabled a better understanding of how the effects of nitrogen doping via NH3 annealing treatments, used to reduce TiO2’s bandgap, can have a strong dependence on TiO2’s thickness and crystalline quality. In addition, it was found that some of the negative effects on the PEC performance typically associated with N-doped TiO2 could be mitigated if the NH3-annealing was directly preceded by an air-annealing step, especially for ultrathin (i.e., < 10 nm) TiO2 films. ALD was also used to conformally coat an ultraporous conductive fluorine-doped tin oxide nanoparticle (nanoFTO) scaffold with an ultrathin layer of TiO2. The integration of these ultrathin films and the oxide nanoparticles resulted in a heteronanostructure design with excellent PEC water oxidation photocurrents (0.7 mA/cm2 at 0 V vs. Ag/AgCl) and charge transfer efficiency.

In Chapter 3, two innovative nanoarchitectures were engineered in order to enhance the pseudocapacitive energy storage of next generation supercapacitor electrodes. The morphology and quantity of MnO2 electrodeposits was controlled by adjusting the density of graphene foliates on a novel graphenated carbon nanotube (g-CNT) scaffold. This control enabled the nanocomposite supercapacitor electrode to reach a capacitance of 640 F/g, under MnO2 specific mass loading conditions (2.3 mg/cm2) that are higher than previously reported. In the second engineered nanoarchitecture, the electrochemical energy storage properties of a transparent electrode based on a network of solution-processed Cu/Ni cores/shell nanowires (NWs) were activated by electrochemically converting the Ni metal shell into Ni(OH)2. Furthermore, an adjustment of the molar percentage of Ni plated onto the Cu NWs was found to result in a tradeoff between capacitance, transmittance, and stability of the resulting nickel hydroxide-based electrode. The nominal area capacitance and power performance results obtained for this Cu/Ni(OH)2 transparent electrode demonstrates that it has significant potential as a hybrid supercapacitor electrode for integration into cutting edge flexible and transparent electronic devices.