930 resultados para Paleolithic period -- Mathematical models
Resumo:
This thesis is concerned with the study of a non-sequential identification technique, so that it may be applied to the identification of process plant mathematical models from process measurements with the greatest degree of accuracy and reliability. In order to study the accuracy of the technique under differing conditions, simple mathematical models were set up on a parallel hybrid. computer and these models identified from input/output measurements by a small on-line digital computer. Initially, the simulated models were identified on-line. However, this method of operation was found not suitable for a thorough study of the technique due to equipment limitations. Further analysis was carried out in a large off-line computer using data generated by the small on-line computer. Hence identification was not strictly on-line. Results of the work have shovm that the identification technique may be successfully applied in practice. An optimum sampling period is suggested, together with noise level limitations for maximum accuracy. A description of a double-effect evaporator is included in this thesis. It is proposed that the next stage in the work will be the identification of a mathematical model of this evaporator using the teclmique described.
Resumo:
The literature on heat and mass transfer mechanisms in the convective drying of thick beds of solids has been critically reviewed. Related mathematical models of heat transfer are also considered. Experimental and theoretical studies were made of the temperature distribution within beds, and of drying rates, with various materials undergoing convective drying. The experimental work covered thick beds of hygroscopic and non-hygroscopic materials (glass beads of different diameters, polystyrene pellets, activated alumina and wood powder) at air temperatures of 54°C to 84°C. Tests were carried out in a laboratory drying apparatus comprising a wind tunnel through which the air, of controlled temperature and humidity, was passed over a sample suspended from a balance. Thermocouples were inserted at different depths within the sample bed. The temperature distribution profiles for both hygroscopic and non-hygroscopic beds exhibited a clear difference between the temperatures at the surface and bottom during the constant rate period. An effective method was introduced for predicting the critical moisture content. During the falling rate the profiles showed the existence of a receding evaporation plane; this divided the system into a hotter dry zone in the upper section and a wet zone near the bottom. A graphical procedure was established to predict accurately the position of the receding evaporation front at any time. A new mathematical model, based on the receding evaporation front phenomenon, was proposed to predict temperature distributions throughout a bed during drying. Good agreement was obtained when the model was validated by comparing its predictions with experimental data. The model was also able to predict the duration of each drying stage. In experiments using sample trays of different diameters, the drying rate was found to increase with a decrease in the effective length of the bed surface. During the constant rate period with trays of a small effective length, i.e. less than 0.08 m, an 'inversion' in temperature distribution occurred in the bed; the bottom temperature increased and became greater than that of the surface. Experimental measurements were verified in several ways to ensure this phenomenon was real. Theoretical explanations are given for both the effective length and temperature inversion phenomena.
Resumo:
* This paper was made according to the program of fundamental scientific research of the Presidium of the Russian Academy of Sciences «Mathematical simulation and intellectual systems», the project "Theoretical foundation of the intellectual systems based on ontologies for intellectual support of scientific researches".
Resumo:
For metal and metal halide vapor lasers excited by high frequency pulsed discharge, the thermal effect mainly caused by the radial temperature distribution is of considerable importance for stable laser operation and improvement of laser output characteristics. A short survey of the obtained analytical and numerical-analytical mathematical models of the temperature profile in a high-powered He-SrBr2 laser is presented. The models are described by the steady-state heat conduction equation with mixed type nonlinear boundary conditions for the arbitrary form of the volume power density. A complete model of radial heat flow between the two tubes is established for precise calculating the inner wall temperature. The models are applied for simulating temperature profiles for newly designed laser. The author’s software prototype LasSim is used for carrying out the mathematical models and simulations.
Resumo:
Nel presente lavoro, ho studiato e trovato le soluzioni esatte di un modello matematico applicato ai recettori cellulari della famiglia delle integrine. Nel modello le integrine sono considerate come un sistema a due livelli, attivo e non attivo. Quando le integrine si trovano nello stato inattivo possono diffondere nella membrana, mentre quando si trovano nello stato attivo risultano cristallizzate nella membrana, incapaci di diffondere. La variazione di concentrazione nella superficie cellulare di una sostanza chiamata attivatore dà luogo all’attivazione delle integrine. Inoltre, questi eterodimeri possono legare una molecola inibitrice con funzioni di controllo e regolazione, che chiameremo v, la quale, legandosi al recettore, fa aumentare la produzione della sostanza attizzatrice, che chiameremo u. In questo modo si innesca un meccanismo di retroazione positiva. L’inibitore v regola il meccanismo di produzione di u, ed assume, pertanto, il ruolo di modulatore. Infatti, grazie a questo sistema di fine regolazione il meccanismo di feedback positivo è in grado di autolimitarsi. Si costruisce poi un modello di equazioni differenziali partendo dalle semplici reazioni chimiche coinvolte. Una volta che il sistema di equazioni è impostato, si possono desumere le soluzioni per le concentrazioni dell’inibitore e dell’attivatore per un caso particolare dei parametri. Infine, si può eseguire un test per vedere cosa predice il modello in termini di integrine. Per farlo, ho utilizzato un’attivazione del tipo funzione gradino e l’ho inserita nel sistema, valutando la dinamica dei recettori. Si ottiene in questo modo un risultato in accordo con le previsioni: le integrine legate si trovano soprattutto ai limiti della zona attivata, mentre le integrine libere vengono a mancare nella zona attivata.
Resumo:
In perifusion cell cultures, the culture medium flows continuously through a chamber containing immobilized cells and the effluent is collected at the end. In our main applications, gonadotropin releasing hormone (GnRH) or oxytocin is introduced into the chamber as the input. They stimulate the cells to secrete luteinizing hormone (LH), which is collected in the effluent. To relate the effluent LH concentration to the cellular processes producing it, we develop and analyze a mathematical model consisting of coupled partial differential equations describing the intracellular signaling and the movement of substances in the cell chamber. We analyze three different data sets and give cellular mechanisms that explain the data. Our model indicates that two negative feedback loops, one fast and one slow, are needed to explain the data and we give their biological bases. We demonstrate that different LH outcomes in oxytocin and GnRH stimulations might originate from different receptor dynamics. We analyze the model to understand the influence of parameters, like the rate of the medium flow or the fraction collection time, on the experimental outcomes. We investigate how the rate of binding and dissociation of the input hormone to and from its receptor influence its movement down the chamber. Finally, we formulate and analyze simpler models that allow us to predict the distortion of a square pulse due to hormone-receptor interactions and to estimate parameters using perifusion data. We show that in the limit of high binding and dissociation the square pulse moves as a diffusing Gaussian and in this limit the biological parameters can be estimated.
Resumo:
Uncertainty quantification (UQ) is both an old and new concept. The current novelty lies in the interactions and synthesis of mathematical models, computer experiments, statistics, field/real experiments, and probability theory, with a particular emphasize on the large-scale simulations by computer models. The challenges not only come from the complication of scientific questions, but also from the size of the information. It is the focus in this thesis to provide statistical models that are scalable to massive data produced in computer experiments and real experiments, through fast and robust statistical inference.
Chapter 2 provides a practical approach for simultaneously emulating/approximating massive number of functions, with the application on hazard quantification of Soufri\`{e}re Hills volcano in Montserrate island. Chapter 3 discusses another problem with massive data, in which the number of observations of a function is large. An exact algorithm that is linear in time is developed for the problem of interpolation of Methylation levels. Chapter 4 and Chapter 5 are both about the robust inference of the models. Chapter 4 provides a new criteria robustness parameter estimation criteria and several ways of inference have been shown to satisfy such criteria. Chapter 5 develops a new prior that satisfies some more criteria and is thus proposed to use in practice.
Resumo:
Fire is a form of uncontrolled combustion which generates heat, smoke, toxic and irritant gases. All of these products are harmful to man and account for the heavy annual cost of 800 lives and £1,000,000,000 worth of property damage in Britain alone. The new discipline of Fire Safety Engineering has developed as a means of reducing these unacceptable losses. One of the main tools of Fire Safety Engineering is the mathematical model and over the past 15 years a number of mathematical models have emerged to cater for the needs of this discipline. Part of the difficulty faced by the Fire Safety Engineer is the selection of the most appropriate modelling tool to use for the job. To make an informed choice it is essential to have a good understanding of the various modelling approaches, their capabilities and limitations. In this paper some of the fundamental modelling tools used to predict fire and evacuation are investigated as are the issues associated with their use and recent developments in modelling technology.
Resumo:
The analysis of steel and composite frames has traditionally been carried out by idealizing beam-to-column connections as either rigid or pinned. Although some advanced analysis methods have been proposed to account for semi-rigid connections, the performance of these methods strongly depends on the proper modeling of connection behavior. The primary challenge of modeling beam-to-column connections is their inelastic response and continuously varying stiffness, strength, and ductility. In this dissertation, two distinct approaches—mathematical models and informational models—are proposed to account for the complex hysteretic behavior of beam-to-column connections. The performance of the two approaches is examined and is then followed by a discussion of their merits and deficiencies. To capitalize on the merits of both mathematical and informational representations, a new approach, a hybrid modeling framework, is developed and demonstrated through modeling beam-to-column connections. Component-based modeling is a compromise spanning two extremes in the field of mathematical modeling: simplified global models and finite element models. In the component-based modeling of angle connections, the five critical components of excessive deformation are identified. Constitutive relationships of angles, column panel zones, and contact between angles and column flanges, are derived by using only material and geometric properties and theoretical mechanics considerations. Those of slip and bolt hole ovalization are simplified by empirically-suggested mathematical representation and expert opinions. A mathematical model is then assembled as a macro-element by combining rigid bars and springs that represent the constitutive relationship of components. Lastly, the moment-rotation curves of the mathematical models are compared with those of experimental tests. In the case of a top-and-seat angle connection with double web angles, a pinched hysteretic response is predicted quite well by complete mechanical models, which take advantage of only material and geometric properties. On the other hand, to exhibit the highly pinched behavior of a top-and-seat angle connection without web angles, a mathematical model requires components of slip and bolt hole ovalization, which are more amenable to informational modeling. An alternative method is informational modeling, which constitutes a fundamental shift from mathematical equations to data that contain the required information about underlying mechanics. The information is extracted from observed data and stored in neural networks. Two different training data sets, analytically-generated and experimental data, are tested to examine the performance of informational models. Both informational models show acceptable agreement with the moment-rotation curves of the experiments. Adding a degradation parameter improves the informational models when modeling highly pinched hysteretic behavior. However, informational models cannot represent the contribution of individual components and therefore do not provide an insight into the underlying mechanics of components. In this study, a new hybrid modeling framework is proposed. In the hybrid framework, a conventional mathematical model is complemented by the informational methods. The basic premise of the proposed hybrid methodology is that not all features of system response are amenable to mathematical modeling, hence considering informational alternatives. This may be because (i) the underlying theory is not available or not sufficiently developed, or (ii) the existing theory is too complex and therefore not suitable for modeling within building frame analysis. The role of informational methods is to model aspects that the mathematical model leaves out. Autoprogressive algorithm and self-learning simulation extract the missing aspects from a system response. In a hybrid framework, experimental data is an integral part of modeling, rather than being used strictly for validation processes. The potential of the hybrid methodology is illustrated through modeling complex hysteretic behavior of beam-to-column connections. Mechanics-based components of deformation such as angles, flange-plates, and column panel zone, are idealized to a mathematical model by using a complete mechanical approach. Although the mathematical model represents envelope curves in terms of initial stiffness and yielding strength, it is not capable of capturing the pinching effects. Pinching is caused mainly by separation between angles and column flanges as well as slip between angles/flange-plates and beam flanges. These components of deformation are suitable for informational modeling. Finally, the moment-rotation curves of the hybrid models are validated with those of the experimental tests. The comparison shows that the hybrid models are capable of representing the highly pinched hysteretic behavior of beam-to-column connections. In addition, the developed hybrid model is successfully used to predict the behavior of a newly-designed connection.
Resumo:
Neste trabalho apresentamos a teoria da análise de correlação canónica, uma técnica de análise estatística multivariada para o estudo da relação, simultânea, entre dois, três ou mais grupos de variáveis. Descrevemos a natureza da correlação canónica com três ou mais variáveis, com modelos matemáticos, fazendo uma síntese dos métodos de generalização de correlação canónica nomeadamente o método Ssqcor, método Sumcor, método Ecart, método Maxvar, método Minvar, e o método de Carroll. Apresentamos uma aplicação utilizando dados provenientes do cálculo do Índice de Preços no Consumidor IPC, produzido pelo INE - STP (Instituto Nacional de Estatística de São Tomé e Príncipe), referente ao período 2010 a 2014. Estamos interessados em conhecer as correlações canónicas entre grupos de variáveis relacionadas com o cabaz de produtos pré-estabelecido para o cálculo do índice de preços no consumidor, concretamente os produtos alimentares (PA), produtos para bebidas (PB) e produtos não alimentares (PNA), constituindo assim os três grandes grupos de variáveis da nossa pesquisa.
Resumo:
Dissertação de Mestrado, Gestão Empresarial, Faculdade de Economia, Universidade do Algarve, 2015
Resumo:
With progressive climate change, the preservation of biodiversity is becoming increasingly important. Only if the gene pool is large enough and requirements of species are diverse, there will be species that can adapt to the changing circumstances. To maintain biodiversity, we must understand the consequences of the various strategies. Mathematical models of population dynamics could provide prognoses. However, a model that would reproduce and explain the mechanisms behind the diversity of species that we observe experimentally and in nature is still needed. A combination of theoretical models with detailed experiments is needed to test biological processes in models and compare predictions with outcomes in reality. In this thesis, several food webs are modeled and analyzed. Among others, models are formulated of laboratory experiments performed in the Zoological Institute of the University of Cologne. Numerical data of the simulations is in good agreement with the real experimental results. Via numerical simulations it can be demonstrated that few assumptions are necessary to reproduce in a model the sustained oscillations of the population size that experiments show. However, analysis indicates that species "thrown together by chance" are not very likely to survive together over long periods. Even larger food nets do not show significantly different outcomes and prove how extraordinary and complicated natural diversity is. In order to produce such a coexistence of randomly selected species—as the experiment does—models require additional information about biological processes or restrictions on the assumptions. Another explanation for the observed coexistence is a slow extinction that takes longer than the observation time. Simulated species survive a comparable period of time before they die out eventually. Interestingly, it can be stated that the same models allow the survival of several species in equilibrium and thus do not follow the so-called competitive exclusion principle. This state of equilibrium is more fragile, however, to changes in nutrient supply than the oscillating coexistence. Overall, the studies show, that having a diverse system means that population numbers are probably oscillating, and on the other hand oscillating population numbers stabilize a food web both against demographic noise as well as against changes of the habitat. Model predictions can certainly not be converted at their face value into policies for real ecosystems. But the stabilizing character of fluctuations should be considered in the regulations of animal populations.
Resumo:
In this thesis, we investigate the role of applied physics in epidemiological surveillance through the application of mathematical models, network science and machine learning. The spread of a communicable disease depends on many biological, social, and health factors. The large masses of data available make it possible, on the one hand, to monitor the evolution and spread of pathogenic organisms; on the other hand, to study the behavior of people, their opinions and habits. Presented here are three lines of research in which an attempt was made to solve real epidemiological problems through data analysis and the use of statistical and mathematical models. In Chapter 1, we applied language-inspired Deep Learning models to transform influenza protein sequences into vectors encoding their information content. We then attempted to reconstruct the antigenic properties of different viral strains using regression models and to identify the mutations responsible for vaccine escape. In Chapter 2, we constructed a compartmental model to describe the spread of a bacterium within a hospital ward. The model was informed and validated on time series of clinical measurements, and a sensitivity analysis was used to assess the impact of different control measures. Finally (Chapter 3) we reconstructed the network of retweets among COVID-19 themed Twitter users in the early months of the SARS-CoV-2 pandemic. By means of community detection algorithms and centrality measures, we characterized users’ attention shifts in the network, showing that scientific communities, initially the most retweeted, lost influence over time to national political communities. In the Conclusion, we highlighted the importance of the work done in light of the main contemporary challenges for epidemiological surveillance. In particular, we present reflections on the importance of nowcasting and forecasting, the relationship between data and scientific research, and the need to unite the different scales of epidemiological surveillance.
Resumo:
Artificial Intelligence (AI) and Machine Learning (ML) are novel data analysis techniques providing very accurate prediction results. They are widely adopted in a variety of industries to improve efficiency and decision-making, but they are also being used to develop intelligent systems. Their success grounds upon complex mathematical models, whose decisions and rationale are usually difficult to comprehend for human users to the point of being dubbed as black-boxes. This is particularly relevant in sensitive and highly regulated domains. To mitigate and possibly solve this issue, the Explainable AI (XAI) field became prominent in recent years. XAI consists of models and techniques to enable understanding of the intricated patterns discovered by black-box models. In this thesis, we consider model-agnostic XAI techniques, which can be applied to Tabular data, with a particular focus on the Credit Scoring domain. Special attention is dedicated to the LIME framework, for which we propose several modifications to the vanilla algorithm, in particular: a pair of complementary Stability Indices that accurately measure LIME stability, and the OptiLIME policy which helps the practitioner finding the proper balance among explanations' stability and reliability. We subsequently put forward GLEAMS a model-agnostic surrogate interpretable model which requires to be trained only once, while providing both Local and Global explanations of the black-box model. GLEAMS produces feature attributions and what-if scenarios, from both dataset and model perspective. Eventually, we argue that synthetic data are an emerging trend in AI, being more and more used to train complex models instead of original data. To be able to explain the outcomes of such models, we must guarantee that synthetic data are reliable enough to be able to translate their explanations to real-world individuals. To this end we propose DAISYnt, a suite of tests to measure synthetic tabular data quality and privacy.
Resumo:
Il primo modello matematico in grado di descrivere il prototipo di un sistema eccitabile assimilabile ad un neurone fu sviluppato da R. FitzHugh e J. Nagumo nel 1961. Tale modello, per quanto schematico, rappresenta un importante punto di partenza per la ricerca nell'ambito neuroscientifico delle dinamiche neuronali, ed è infatti capostipite di una serie di lavori che hanno puntato a migliorare l’accuratezza e la predicibilità dei modelli matematici per le scienze. L’elevato grado di complessità nello studio dei neuroni e delle dinamiche inter-neuronali comporta, tuttavia, che molte delle caratteristiche e delle potenzialità dell’ambito non siano ancora state comprese appieno. In questo lavoro verrà approfondito un modello ispirato al lavoro originale di FitzHugh e Nagumo. Tale modello presenta l’introduzione di un termine di self-coupling con ritardo temporale nel sistema di equazioni differenziali, diventa dunque rappresentativo di modelli di campo medio in grado di descrivere gli stati macroscopici di un ensemble di neuroni. L'introduzione del ritardo è funzionale ad una descrizione più realistica dei sistemi neuronali, e produce una dinamica più ricca e complessa rispetto a quella presente nella versione originale del modello. Sarà mostrata l'esistenza di una soluzione a ciclo limite nel modello che comprende il termine di ritardo temporale, ove tale soluzione non può essere interpretata nell’ambito delle biforcazioni di Hopf. Allo scopo di esplorare alcune delle caratteristiche basilari della modellizzazione del neurone, verrà principalmente utilizzata l’impostazione della teoria dei sistemi dinamici, integrando dove necessario con alcune nozioni provenienti dall’ambito fisiologico. In conclusione sarà riportata una sezione di approfondimento sulla integrazione numerica delle equazioni differenziali con ritardo.