10 resultados para scale-free networks
em CaltechTHESIS
Resumo:
There are two competing models of our universe right now. One is Big Bang with inflation cosmology. The other is the cyclic model with ekpyrotic phase in each cycle. This paper is divided into two main parts according to these two models. In the first part, we quantify the potentially observable effects of a small violation of translational invariance during inflation, as characterized by the presence of a preferred point, line, or plane. We explore the imprint such a violation would leave on the cosmic microwave background anisotropy, and provide explicit formulas for the expected amplitudes $\langle a_{lm}a_{l'm'}^*\rangle$ of the spherical-harmonic coefficients. We then provide a model and study the two-point correlation of a massless scalar (the inflaton) when the stress tensor contains the energy density from an infinitely long straight cosmic string in addition to a cosmological constant. Finally, we discuss if inflation can reconcile with the Liouville's theorem as far as the fine-tuning problem is concerned. In the second part, we find several problems in the cyclic/ekpyrotic cosmology. First of all, quantum to classical transition would not happen during an ekpyrotic phase even for superhorizon modes, and therefore the fluctuations cannot be interpreted as classical. This implies the prediction of scale-free power spectrum in ekpyrotic/cyclic universe model requires more inspection. Secondly, we find that the usual mechanism to solve fine-tuning problems is not compatible with eternal universe which contains infinitely many cycles in both direction of time. Therefore, all fine-tuning problems including the flatness problem still asks for an explanation in any generic cyclic models.
Resumo:
In four chapters various aspects of earthquake source are studied.
Chapter I
Surface displacements that followed the Parkfield, 1966, earthquakes were measured for two years with six small-scale geodetic networks straddling the fault trace. The logarithmic rate and the periodic nature of the creep displacement recorded on a strain meter made it possible to predict creep episodes on the San Andreas fault. Some individual earthquakes were related directly to surface displacement, while in general, slow creep and aftershock activity were found to occur independently. The Parkfield earthquake is interpreted as a buried dislocation.
Chapter II
The source parameters of earthquakes between magnitude 1 and 6 were studied using field observations, fault plane solutions, and surface wave and S-wave spectral analysis. The seismic moment, MO, was found to be related to local magnitude, ML, by log MO = 1.7 ML + 15.1. The source length vs magnitude relation for the San Andreas system found to be: ML = 1.9 log L - 6.7. The surface wave envelope parameter AR gives the moment according to log MO = log AR300 + 30.1, and the stress drop, τ, was found to be related to the magnitude by τ = 0.54 M - 2.58. The relation between surface wave magnitude MS and ML is proposed to be MS = 1.7 ML - 4.1. It is proposed to estimate the relative stress level (and possibly the strength) of a source-region by the amplitude ratio of high-frequency to low-frequency waves. An apparent stress map for Southern California is presented.
Chapter III
Seismic triggering and seismic shaking are proposed as two closely related mechanisms of strain release which explain observations of the character of the P wave generated by the Alaskan earthquake of 1964, and distant fault slippage observed after the Borrego Mountain, California earthquake of 1968. The Alaska, 1964, earthquake is shown to be adequately described as a series of individual rupture events. The first of these events had a body wave magnitude of 6.6 and is considered to have initiated or triggered the whole sequence. The propagation velocity of the disturbance is estimated to be 3.5 km/sec. On the basis of circumstantial evidence it is proposed that the Borrego Mountain, 1968, earthquake caused release of tectonic strain along three active faults at distances of 45 to 75 km from the epicenter. It is suggested that this mechanism of strain release is best described as "seismic shaking."
Chapter IV
The changes of apparent stress with depth are studied in the South American deep seismic zone. For shallow earthquakes the apparent stress is 20 bars on the average, the same as for earthquakes in the Aleutians and on Oceanic Ridges. At depths between 50 and 150 km the apparent stresses are relatively high, approximately 380 bars, and around 600 km depth they are again near 20 bars. The seismic efficiency is estimated to be 0.1. This suggests that the true stress is obtained by multiplying the apparent stress by ten. The variation of apparent stress with depth is explained in terms of the hypothesis of ocean floor consumption.
Resumo:
The dissertation studies the general area of complex networked systems that consist of interconnected and active heterogeneous components and usually operate in uncertain environments and with incomplete information. Problems associated with those systems are typically large-scale and computationally intractable, yet they are also very well-structured and have features that can be exploited by appropriate modeling and computational methods. The goal of this thesis is to develop foundational theories and tools to exploit those structures that can lead to computationally-efficient and distributed solutions, and apply them to improve systems operations and architecture.
Specifically, the thesis focuses on two concrete areas. The first one is to design distributed rules to manage distributed energy resources in the power network. The power network is undergoing a fundamental transformation. The future smart grid, especially on the distribution system, will be a large-scale network of distributed energy resources (DERs), each introducing random and rapid fluctuations in power supply, demand, voltage and frequency. These DERs provide a tremendous opportunity for sustainability, efficiency, and power reliability. However, there are daunting technical challenges in managing these DERs and optimizing their operation. The focus of this dissertation is to develop scalable, distributed, and real-time control and optimization to achieve system-wide efficiency, reliability, and robustness for the future power grid. In particular, we will present how to explore the power network structure to design efficient and distributed market and algorithms for the energy management. We will also show how to connect the algorithms with physical dynamics and existing control mechanisms for real-time control in power networks.
The second focus is to develop distributed optimization rules for general multi-agent engineering systems. A central goal in multiagent systems is to design local control laws for the individual agents to ensure that the emergent global behavior is desirable with respect to the given system level objective. Ideally, a system designer seeks to satisfy this goal while conditioning each agent’s control on the least amount of information possible. Our work focused on achieving this goal using the framework of game theory. In particular, we derived a systematic methodology for designing local agent objective functions that guarantees (i) an equivalence between the resulting game-theoretic equilibria and the system level design objective and (ii) that the resulting game possesses an inherent structure that can be exploited for distributed learning, e.g., potential games. The control design can then be completed by applying any distributed learning algorithm that guarantees convergence to the game-theoretic equilibrium. One main advantage of this game theoretic approach is that it provides a hierarchical decomposition between the decomposition of the systemic objective (game design) and the specific local decision rules (distributed learning algorithms). This decomposition provides the system designer with tremendous flexibility to meet the design objectives and constraints inherent in a broad class of multiagent systems. Furthermore, in many settings the resulting controllers will be inherently robust to a host of uncertainties including asynchronous clock rates, delays in information, and component failures.
Resumo:
This thesis belongs to the growing field of economic networks. In particular, we develop three essays in which we study the problem of bargaining, discrete choice representation, and pricing in the context of networked markets. Despite analyzing very different problems, the three essays share the common feature of making use of a network representation to describe the market of interest.
In Chapter 1 we present an analysis of bargaining in networked markets. We make two contributions. First, we characterize market equilibria in a bargaining model, and find that players' equilibrium payoffs coincide with their degree of centrality in the network, as measured by Bonacich's centrality measure. This characterization allows us to map, in a simple way, network structures into market equilibrium outcomes, so that payoffs dispersion in networked markets is driven by players' network positions. Second, we show that the market equilibrium for our model converges to the so called eigenvector centrality measure. We show that the economic condition for reaching convergence is that the players' discount factor goes to one. In particular, we show how the discount factor, the matching technology, and the network structure interact in a very particular way in order to see the eigenvector centrality as the limiting case of our market equilibrium.
We point out that the eigenvector approach is a way of finding the most central or relevant players in terms of the “global” structure of the network, and to pay less attention to patterns that are more “local”. Mathematically, the eigenvector centrality captures the relevance of players in the bargaining process, using the eigenvector associated to the largest eigenvalue of the adjacency matrix of a given network. Thus our result may be viewed as an economic justification of the eigenvector approach in the context of bargaining in networked markets.
As an application, we analyze the special case of seller-buyer networks, showing how our framework may be useful for analyzing price dispersion as a function of sellers and buyers' network positions.
Finally, in Chapter 3 we study the problem of price competition and free entry in networked markets subject to congestion effects. In many environments, such as communication networks in which network flows are allocated, or transportation networks in which traffic is directed through the underlying road architecture, congestion plays an important role. In particular, we consider a network with multiple origins and a common destination node, where each link is owned by a firm that sets prices in order to maximize profits, whereas users want to minimize the total cost they face, which is given by the congestion cost plus the prices set by firms. In this environment, we introduce the notion of Markovian traffic equilibrium to establish the existence and uniqueness of a pure strategy price equilibrium, without assuming that the demand functions are concave nor imposing particular functional forms for the latency functions. We derive explicit conditions to guarantee existence and uniqueness of equilibria. Given this existence and uniqueness result, we apply our framework to study entry decisions and welfare, and establish that in congested markets with free entry, the number of firms exceeds the social optimum.
Resumo:
Underlying matter and light are their building blocks of tiny atoms and photons. The ability to control and utilize matter-light interactions down to the elementary single atom and photon level at the nano-scale opens up exciting studies at the frontiers of science with applications in medicine, energy, and information technology. Of these, an intriguing front is the development of quantum networks where N >> 1 single-atom nodes are coherently linked by single photons, forming a collective quantum entity potentially capable of performing quantum computations and simulations. Here, a promising approach is to use optical cavities within the setting of cavity quantum electrodynamics (QED). However, since its first realization in 1992 by Kimble et al., current proof-of-principle experiments have involved just one or two conventional cavities. To move beyond to N >> 1 nodes, in this thesis we investigate a platform born from the marriage of cavity QED and nanophotonics, where single atoms at ~100 nm near the surfaces of lithographically fabricated dielectric photonic devices can strongly interact with single photons, on a chip. Particularly, we experimentally investigate three main types of devices: microtoroidal optical cavities, optical nanofibers, and nanophotonic crystal based structures. With a microtoroidal cavity, we realized a robust and efficient photon router where single photons are extracted from an incident coherent state of light and redirected to a separate output with high efficiency. We achieved strong single atom-photon coupling with atoms located ~100 nm near the surface of a microtoroid, which revealed important aspects in the atom dynamics and QED of these systems including atom-surface interaction effects. We present a method to achieve state-insensitive atom trapping near optical nanofibers, critical in nanophotonic systems where electromagnetic fields are tightly confined. We developed a system that fabricates high quality nanofibers with high controllability, with which we experimentally demonstrate a state-insensitive atom trap. We present initial investigations on nanophotonic crystal based structures as a platform for strong atom-photon interactions. The experimental advances and theoretical investigations carried out in this thesis provide a framework for and open the door to strong single atom-photon interactions using nanophotonics for chip-integrated quantum networks.
Resumo:
The geology and structure of two crustal scale shear zones were studied to understand the partitioning of strain within intracontinental orogenic belts. Movement histories and regional tectonic implications are deduced from observational data. The two widely separated study areas bear the imprint of intense Late Mesozoic through Middle Cenozoic tectonic activity. A regional transition from Late Cretaceous-Early Tertiary plutonism, metamorphism, and shortening strain to Middle Tertiary extension and magmatism is preserved in each area, with contrasting environments and mechanisms. Compressional phases of this tectonic history are better displayed in the Rand Mountains, whereas younger extensional structures dominate rock fabrics in the Magdalena area.
In the northwestern Mojave desert, the Rand Thrust Complex reveals a stack of four distinctive tectonic plates offset along the Garlock Fault. The lowermost plate, Rand Schist, is composed of greenschist facies metagraywacke, metachert, and metabasalt. Rand Schist is structurally overlain by Johannesburg Gneiss (= garnet-amphibolite grade orthogneisses, marbles and quartzites), which in turn is overlain by a Late Cretaceous hornblende-biotite granodiorite. Biotite granite forms the fourth and highest plate. Initial assembly of the tectonic stack involved a Late Cretaceous? south or southwest vergent overthrusting event in which Johannesburg Gneiss was imbricated and attenuated between Rand Schist and hornblende-biotite granodiorite. Thrusting postdated metamorphism and deformation of the lower two plates in separate environments. A post-kinematic stock, the Late Cretaceous Randsburg Granodiorite, intrudes deep levels of the complex and contains xenoliths of both Rand Schist and mylonitized Johannesburg? gneiss. Minimum shortening implied by the map patterns is 20 kilometers.
Some low angle faults of the Rand Thrust Complex formed or were reactivated between Late Cretaceous and Early Miocene time. South-southwest directed mylonites derived from Johannesburg Gneiss are commonly overprinted by less penetrative north-northeast vergent structures. Available kinematic information at shallower structural levels indicates that late disturbance(s) culminated in northward transport of the uppermost plate. Persistence of brittle fabrics along certain structural horizons suggests a possible association of late movement(s) with regionally known detachment faults. The four plates were juxtaposed and significant intraplate movements had ceased prior to Early Miocene emplacement of rhyolite porphyry dikes.
In the Magdalena region of north central Sonora, components of a pre-Middle Cretaceous stratigraphy are used as strain markers in tracking the evolution of a long lived orogenic belt. Important elements of the tectonic history include: (1) Compression during the Late Cretaceous and Early Tertiary, accompanied by plutonism, metamorphism, and ductile strain at depth, and thrust driven? syntectonic sedimentation at the surface. (2) Middle Tertiary transition to crustal extension, initially recorded by intrusion of leucogranites, inflation of the previously shortened middle and upper crustal section, and surface volcanism. (3) Gravity induced development of a normal sense ductile shear zone at mid crustal levels, with eventual detachment and southwestward displacement of the upper crustal stratigraphy by Early Miocene time.
Elucidation of the metamorphic core complex evolution just described was facilitated by fortuitous preservation of a unique assemblage of rocks and structures. The "type" stratigraphy utilized for regional correlation and strain analysis includes a Jurassic volcanic arc assemblage overlain by an Upper Jurassic-Lower Cretaceous quartz pebble conglomerate, in turn overlain by marine strata with fossiliferous Aptian-Albian limestones. The Jurassic strata, comprised of (a) rhyolite porphyries interstratified with quartz arenites, (b) rhyolite cobble conglomerate, and (c) intrusive granite porphyries, are known to rest on Precambrian basement north and east of the study area. The quartz pebble conglomerate is correlated with the Glance Conglomerate of southeastern Arizona and northeastern Sonora. The marine sequence represents part of an isolated arm? of the Bisbee Basin.
Crosscutting structural relationships between the pre-Middle Cretaceous supracrustal section, younger plutons, and deformational fabrics allow the tectonic sequence to be determined. Earliest phases of a Late Cretaceous-Early Tertiary orogeny are marked by emplacement of the 78 ± 3 Ma Guacomea Granodiorite (U/Pb zircon, Anderson et al., 1980) as a sill into deep levels of the layered Jurassic series. Subsequent regional metamorphism and ductile strain is recorded by a penetrative schistosity and lineation, and east-west trending folds. These fabrics are intruded by post-kinematic Early Tertiary? two mica granites. At shallower crustal levels, the orogeny is represented by north directed thrust faulting, formation of a large intermontane basin, and development of a pronounced unconformity. A second important phase of ductile strain followed Middle Tertiary? emplacement of leucogranites as sills and northwest trending dikes into intermediate levels of the deformed section (surficial volcanism was also active during this transitional period to regional extension). Gravitational instabilities resulting from crustal swelling via intrusion and thermal expansion led to development of a ductile shear zone within the stratigraphic horizon occupied by a laterally extensive leucogranite sill. With continued extension, upper crustal brittle normal faults (detachment faults) enhanced the uplift and tectonic denudation of this mylonite zone, ultimately resulting in southwestward displacement of the upper crustal stratigraphy.
Strains associated with the two ductile deformation events have been successfully partitioned through a multifaceted analysis. R_f/Ø measurements on various markers from the "type" stratigraphy allow a gradient representing cumulative strain since Middle Cretaceous time to be determined. From this gradient, noncoaxial strains accrued since emplacement of the leucogranites may be removed. Irrotational components of the postleucogranite strain are measured from quartz grain shapes in deformed granites; rotational components (shear strains) are determined from S-C fabrics and from restoration of rotated dike and vein networks. Structural observations and strain data are compatable with a deformation path of: (1) coaxial strain (pure shear?), followed by (2) injection of leucogranites as dikes (perpendicular to the minimum principle stress) and sills (parallel to the minimum principle stress), then (3) southwest directed simple shear. Modeling the late strain gradient as a simple shear zone permits a minimum displacement of 10 kilometers on the Magdalena mylonite zone/detachment fault system. Removal of the Middle Tertiary noncoaxial strains yields a residual (or pre-existing) strain gradient representative of the Late Cretaceous-Early Tertiary deformation. Several partially destrained cross sections, restored to the time of leucogranite emplacement, illustrate the idea that the upper plate of the core complex bas been detached from a region of significant topographic relief. 50% to 100% bulk extension across a 50 kilometer wide corridor is demonstrated.
Late Cenozoic tectonics of the Magdalena region are dominated by Basin and Range style faulting. Northeast and north-northwest trending high angle normal faults have interacted to extend the crust in an east-west direction. Net extension for this period is minor (10% to 15%) in comparison to the Middle Tertiary detachment related extensional episode.
Resumo:
Smartphones and other powerful sensor-equipped consumer devices make it possible to sense the physical world at an unprecedented scale. Nearly 2 million Android and iOS devices are activated every day, each carrying numerous sensors and a high-speed internet connection. Whereas traditional sensor networks have typically deployed a fixed number of devices to sense a particular phenomena, community networks can grow as additional participants choose to install apps and join the network. In principle, this allows networks of thousands or millions of sensors to be created quickly and at low cost. However, making reliable inferences about the world using so many community sensors involves several challenges, including scalability, data quality, mobility, and user privacy.
This thesis focuses on how learning at both the sensor- and network-level can provide scalable techniques for data collection and event detection. First, this thesis considers the abstract problem of distributed algorithms for data collection, and proposes a distributed, online approach to selecting which set of sensors should be queried. In addition to providing theoretical guarantees for submodular objective functions, the approach is also compatible with local rules or heuristics for detecting and transmitting potentially valuable observations. Next, the thesis presents a decentralized algorithm for spatial event detection, and describes its use detecting strong earthquakes within the Caltech Community Seismic Network. Despite the fact that strong earthquakes are rare and complex events, and that community sensors can be very noisy, our decentralized anomaly detection approach obtains theoretical guarantees for event detection performance while simultaneously limiting the rate of false alarms.
Resumo:
An exciting frontier in quantum information science is the integration of otherwise "simple'' quantum elements into complex quantum networks. The laboratory realization of even small quantum networks enables the exploration of physical systems that have not heretofore existed in the natural world. Within this context, there is active research to achieve nanoscale quantum optical circuits, for which atoms are trapped near nano-scopic dielectric structures and "wired'' together by photons propagating through the circuit elements. Single atoms and atomic ensembles endow quantum functionality for otherwise linear optical circuits and thereby enable the capability of building quantum networks component by component. Toward these goals, we have experimentally investigated three different systems, from conventional to rather exotic systems : free-space atomic ensembles, optical nano fibers, and photonics crystal waveguides. First, we demonstrate measurement-induced quadripartite entanglement among four quantum memories. Next, following the landmark realization of a nanofiber trap, we demonstrate the implementation of a state-insensitive, compensated nanofiber trap. Finally, we reach more exotic systems based on photonics crystal devices. Beyond conventional topologies of resonators and waveguides, new opportunities emerge from the powerful capabilities of dispersion and modal engineering in photonic crystal waveguides. We have implemented an integrated optical circuit with a photonics crystal waveguide capable of both trapping and interfacing atoms with guided photons, and have observed the collective effect, superradiance, mediated by the guided photons. These advances provide an important capability for engineered light-matter interactions, enabling explorations of novel quantum transport and quantum many-body phenomena.
Resumo:
Climate change is arguably the most critical issue facing our generation and the next. As we move towards a sustainable future, the grid is rapidly evolving with the integration of more and more renewable energy resources and the emergence of electric vehicles. In particular, large scale adoption of residential and commercial solar photovoltaics (PV) plants is completely changing the traditional slowly-varying unidirectional power flow nature of distribution systems. High share of intermittent renewables pose several technical challenges, including voltage and frequency control. But along with these challenges, renewable generators also bring with them millions of new DC-AC inverter controllers each year. These fast power electronic devices can provide an unprecedented opportunity to increase energy efficiency and improve power quality, if combined with well-designed inverter control algorithms. The main goal of this dissertation is to develop scalable power flow optimization and control methods that achieve system-wide efficiency, reliability, and robustness for power distribution networks of future with high penetration of distributed inverter-based renewable generators.
Proposed solutions to power flow control problems in the literature range from fully centralized to fully local ones. In this thesis, we will focus on the two ends of this spectrum. In the first half of this thesis (chapters 2 and 3), we seek optimal solutions to voltage control problems provided a centralized architecture with complete information. These solutions are particularly important for better understanding the overall system behavior and can serve as a benchmark to compare the performance of other control methods against. To this end, we first propose a branch flow model (BFM) for the analysis and optimization of radial and meshed networks. This model leads to a new approach to solve optimal power flow (OPF) problems using a two step relaxation procedure, which has proven to be both reliable and computationally efficient in dealing with the non-convexity of power flow equations in radial and weakly-meshed distribution networks. We will then apply the results to fast time- scale inverter var control problem and evaluate the performance on real-world circuits in Southern California Edison’s service territory.
The second half (chapters 4 and 5), however, is dedicated to study local control approaches, as they are the only options available for immediate implementation on today’s distribution networks that lack sufficient monitoring and communication infrastructure. In particular, we will follow a reverse and forward engineering approach to study the recently proposed piecewise linear volt/var control curves. It is the aim of this dissertation to tackle some key problems in these two areas and contribute by providing rigorous theoretical basis for future work.
Resumo:
Single-cell functional proteomics assays can connect genomic information to biological function through quantitative and multiplex protein measurements. Tools for single-cell proteomics have developed rapidly over the past 5 years and are providing unique opportunities. This thesis describes an emerging microfluidics-based toolkit for single cell functional proteomics, focusing on the development of the single cell barcode chips (SCBCs) with applications in fundamental and translational cancer research.
The microchip designed to simultaneously quantify a panel of secreted, cytoplasmic and membrane proteins from single cells will be discussed at the beginning, which is the prototype for subsequent proteomic microchips with more sophisticated design in preclinical cancer research or clinical applications. The SCBCs are a highly versatile and information rich tool for single-cell functional proteomics. They are based upon isolating individual cells, or defined number of cells, within microchambers, each of which is equipped with a large antibody microarray (the barcode), with between a few hundred to ten thousand microchambers included within a single microchip. Functional proteomics assays at single-cell resolution yield unique pieces of information that significantly shape the way of thinking on cancer research. An in-depth discussion about analysis and interpretation of the unique information such as functional protein fluctuations and protein-protein correlative interactions will follow.
The SCBC is a powerful tool to resolve the functional heterogeneity of cancer cells. It has the capacity to extract a comprehensive picture of the signal transduction network from single tumor cells and thus provides insight into the effect of targeted therapies on protein signaling networks. We will demonstrate this point through applying the SCBCs to investigate three isogenic cell lines of glioblastoma multiforme (GBM).
The cancer cell population is highly heterogeneous with high-amplitude fluctuation at the single cell level, which in turn grants the robustness of the entire population. The concept that a stable population existing in the presence of random fluctuations is reminiscent of many physical systems that are successfully understood using statistical physics. Thus, tools derived from that field can probably be applied to using fluctuations to determine the nature of signaling networks. In the second part of the thesis, we will focus on such a case to use thermodynamics-motivated principles to understand cancer cell hypoxia, where single cell proteomics assays coupled with a quantitative version of Le Chatelier's principle derived from statistical mechanics yield detailed and surprising predictions, which were found to be correct in both cell line and primary tumor model.
The third part of the thesis demonstrates the application of this technology in the preclinical cancer research to study the GBM cancer cell resistance to molecular targeted therapy. Physical approaches to anticipate therapy resistance and to identify effective therapy combinations will be discussed in detail. Our approach is based upon elucidating the signaling coordination within the phosphoprotein signaling pathways that are hyperactivated in human GBMs, and interrogating how that coordination responds to the perturbation of targeted inhibitor. Strongly coupled protein-protein interactions constitute most signaling cascades. A physical analogy of such a system is the strongly coupled atom-atom interactions in a crystal lattice. Similar to decomposing the atomic interactions into a series of independent normal vibrational modes, a simplified picture of signaling network coordination can also be achieved by diagonalizing protein-protein correlation or covariance matrices to decompose the pairwise correlative interactions into a set of distinct linear combinations of signaling proteins (i.e. independent signaling modes). By doing so, two independent signaling modes – one associated with mTOR signaling and a second associated with ERK/Src signaling have been resolved, which in turn allow us to anticipate resistance, and to design combination therapies that are effective, as well as identify those therapies and therapy combinations that will be ineffective. We validated our predictions in mouse tumor models and all predictions were borne out.
In the last part, some preliminary results about the clinical translation of single-cell proteomics chips will be presented. The successful demonstration of our work on human-derived xenografts provides the rationale to extend our current work into the clinic. It will enable us to interrogate GBM tumor samples in a way that could potentially yield a straightforward, rapid interpretation so that we can give therapeutic guidance to the attending physicians within a clinical relevant time scale. The technical challenges of the clinical translation will be presented and our solutions to address the challenges will be discussed as well. A clinical case study will then follow, where some preliminary data collected from a pediatric GBM patient bearing an EGFR amplified tumor will be presented to demonstrate the general protocol and the workflow of the proposed clinical studies.