900 resultados para Stochastic agent-based models


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Syttymistä ja palamisen etenemistä partikkelikerroksessa tutkitaan paloturvallisuuden parantamista sekä kiinteitä polttoaineita käyttävien polttolaitteiden toiminnan tuntemista ja kehittämistä varten. Tässä tutkimuksessa on tavoitteena kerätä yhteen syttymiseen ja liekkirintaman etenemiseen liittyviä kokeellisia ja teoreettisia tutkimustuloksia, jotka auttavat kiinteäkerrospoltto- ja -kaasutus-laitteiden kehittämisessä ja suunnittelussa. Työ on esitutkimus sitä seuraavalle kokeelliselle ja teoreettiselle osalle. Käsittelyssä keskitytään erityisesti puuperäisiin polttoaineisiin. Hiilidioksidipäästöjen vähentämistavoitteet sekä kiinteiden jätteiden energiakäytön lisääminen ja kaatopaikalle viennin vähentäminen aiheuttavat lähitulevaisuudessa kerrospolton lisääntymistä. Kuljetusmatkojen optimoinnin takia joudutaan rakentamaan melko pieniä polttolaitoksia, joissa kerrospolttotekniikka on edullisin vaihtoehto. Syttymispisteellä tarkoitetaan Semenovin määritelmän mukaan tilaa ja ajankohtaa, jolloin polttoaineen ja hapen reaktioissa muodostuva nettoenergia aikayksikössä on yhtäsuuri kuin ympäristöön siirtyvä nettoenergiavirta. Itsesyttyminen tarkoittaa syttymistä ympäristön lämpötilan tai paineen suurenemisen seurauksena. Pakotettu syttyminen tapahtuu, kun syttymispisteen läheisyydessä on esimerkiksi liekki tai hehkuva kiinteä kappale, joka aiheuttaa paikallisen syttymisen ja syttymisrintaman leviämisen muualle polttoaineeseen. Kokeellinen tutkimus on osoittanut tärkeimmiksi syttymiseen ja syttymisrintaman etenemiseen vaikuttaviksi tekijöiksi polttoaineen kosteuden, haihtuvien aineiden pitoisuuden ja lämpöarvon, partikkelikerroksen huokoisuuden, partikkelien koon ja muodon, polttoaineen pinnalle tulevan säteilylämpövirran tiheyden, kaasun virtausnopeuden kerroksessa, hapen osuuden ympäristössä sekä palamisilman esilämmityksen. Kosteuden lisääntyminen suurentaa syttymisenergiaa ja -lämpötilaa sekä pidentää syttymisaikaa. Mitä enemmän polttoaine sisältää haihtuvia aineita sitä pienemmässä lämpötilassa se syttyy. Syttyminen ja syttymisrintaman eteneminen ovat sitä nopeampia mitä suurempi on polttoaineen lämpöarvo. Kerroksen huokoisuuden kasvun on havaittu suurentavan palamisen etenemisnopeutta. Pienet partikkelit syttyvät yleensä nopeammin ja pienemmässä lämpötilassa kuin suuret. Syttymisrintaman eteneminen nopeutuu partikkelien pinta-ala - tilavuussuhteen kasvaessa. Säteilylämpövirran tiheys on useissa polttosovellutuksissa merkittävin lämmönsiirtotekijä, jonka kasvu luonnollisesti nopeuttaa syttymistä. Ilman ja palamiskaasujen virtausnopeus kerroksessa vaikuttaa konvektiiviseen lämmönsiirtoon ja hapen pitoisuuteen syttymisvyöhykkeellä. Ilmavirtaus voi jäähdyttää ja kuumankaasun virtaus lämmittää kerrosta. Hapen osuuden kasvaminen nopeuttaa syttymistä ja liekkirintaman etenemistä kunnes saavutetaan tila, jota suuremmilla virtauksilla ilma jäähdyttää ja laimentaa reaktiovyöhykettä. Palamisilman esilämmitys nopeuttaa syttymisrintaman etenemistä. Syttymistä ja liekkirintaman etenemistä kuvataan yleensä empiirisillä tai säilyvyysyhtälöihin perustuvilla malleilla. Empiiriset mallit perustuvat mittaustuloksista tehtyihin korrelaatioihin sekä joihinkin tunnettuihin fysikaalisiin lainalaisuuksiin. Säilyvyysyhtälöihin perustuvissa malleissa systeemille määritetään massan, energian, liikemäärän ja alkuaineiden säilymisyhtälöt, joiden nopeutta kuvaavien siirtoyhtälöiden muodostamiseen käytetään teoreettisella ja kokeellisella tutkimuksella saatuja yhtälöitä. Nämä mallinnusluokat ovat osittain päällekkäisiä. Pintojen syttymistä kuvataan usein säilyvyysyhtälöihin perustuvilla malleilla. Partikkelikerrosten mallinnuksessa tukeudutaan enimmäkseen empiirisiin yhtälöihin. Partikkelikerroksia kuvaavista malleista Xien ja Liangin hiilipartikkelikerroksen syttymiseen liittyvä tutkimus ja Gortin puun ja jätteen polttoon liittyvä reaktiorintaman etenemistutkimus ovat lähimpänä säilyvyysyhtälöihin perustuvaa mallintamista. Kaikissa malleissa joudutaan kuitenkin yksinkertaistamaan todellista tapausta esimerkiksi vähentämällä dimensioita, reaktioita ja yhdisteitä sekä eliminoimalla vähemmän merkittävät siirtomekanismit. Suoraan kerrospolttoa ja -kaasutusta palvelevia syttymisen ja palamisen etenemisen tutkimuksia on vähän. Muita tarkoituksia varten tehtyjen tutkimusten polttoaineet, kerrokset ja ympäristöolosuhteet poikkeavat yleensä selvästi polttolaitteiden vastaavista olosuhteista. Erikokoisten polttoainepartikkelien ja ominaisuuksiltaan erilaisten polttoaineiden seospolttoa ei ole tutkittu juuri ollenkaan. Polttoainepartikkelien muodon vaikutuksesta on vain vähän tutkimusta.Ilman kanavoitumisen vaikutuksista ei löytynyt tutkimuksia.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Simulation is a useful tool in cardiac SPECT to assess quantification algorithms. However, simple equation-based models are limited in their ability to simulate realistic heart motion and perfusion. We present a numerical dynamic model of the left ventricle, which allows us to simulate normal and anomalous cardiac cycles, as well as perfusion defects. Bicubic splines were fitted to a number of control points to represent endocardial and epicardial surfaces of the left ventricle. A transformation from each point on the surface to a template of activity was made to represent the myocardial perfusion. Geometry-based and patient-based simulations were performed to illustrate this model. Geometry-based simulations modeled ~1! a normal patient, ~2! a well-perfused patient with abnormal regional function, ~3! an ischaemic patient with abnormal regional function, and ~4! a patient study including tracer kinetics. Patient-based simulation consisted of a left ventricle including a realistic shape and motion obtained from a magnetic resonance study. We conclude that this model has the potential to study the influence of several physical parameters and the left ventricle contraction in myocardial perfusion SPECT and gated-SPECT studies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

One of the classic research topics in adaptive behavior is the collective displacement of groups of organisms such as flocks of birds, schools of fish, herds of mammals and crowds of people. However, most agent-based simulations of group behavior do not provide a quantitative index for determining the point at which the flock emerges. We have developed an index of the aggregation of moving individuals in a flock and have provided an example of how it can be used to quantify the degree to which a group of moving individuals actually forms a flock.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Statistical properties of binary complex networks are well understood and recently many attempts have been made to extend this knowledge to weighted ones. There are, however, subtle yet important considerations to be made regarding the nature of the weights used in this generalization. Weights can be either continuous or discrete magnitudes, and in the latter case, they can additionally have undistinguishable or distinguishable nature. This fact has not been addressed in the literature insofar and has deep implications on the network statistics. In this work we face this problem introducing multiedge networks as graphs where multiple (distinguishable) connections between nodes are considered. We develop a statistical mechanics framework where it is possible to get information about the most relevant observables given a large spectrum of linear and nonlinear constraints including those depending both on the number of multiedges per link and their binary projection. The latter case is particularly interesting as we show that binary projections can be understood from multiedge processes. The implications of these results are important as many real-agent-based problems mapped onto graphs require this treatment for a proper characterization of their collective behavior.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis focuses on the social-psychological factors that help coping with structural disadvantage, and specifically on the role of cohesive ingroups and the sense of connectedness and efficacy they entail in this process. It aims to complement existing group-based models of coping that are grounded in a categorization perspective to groups and consequently focus exclusively on the large-scale categories made salient in intergroup contexts of comparisons. The dissertation accomplishes this aim through a reconsideration of between-persons relational interdependence as a sufficient and independent antecedent of a sense of groupness, and the benefits that a sense of group connectedness in one's direct environment, regardless of the categorical or relational basis of groupness, might have in the everyday struggles of disadvantaged group members. The three empirical papers aim to validate this approach, outlined in the first theoretical introduction, by testing derived hypotheses. They are based on data collected with youth populations (15-30) from three institutions in French-speaking Switzerland within the context of a larger project on youth transitions. Methods of data collection are paper-pencil questionnaires and in-depth interviews with a selected sub-sample of participants. The key argument of the first paper is that members of socially disadvantaged categories face higher barriers to their life project and that a general sense of connectedness, either based on categorical identities or other proximal groups and relations, mitigates the feeling of powerlessness associated with this experience. The second paper develops and tests a model that defines individual needs satisfaction as antecedent of self-group bonds and the efficacy beliefs derived from these intragroup bonds as the mechanism underlining the role of ingroups in coping. The third paper highlights the complexities that might be associated with the construction of a sense of groupness directly from intergroup comparisons and categorization-based disadvantage, and points out a more subtle understanding of the processes underling the emergence of groupness out of the situation of structural disadvantage. Overall, the findings confirm the central role of ingroups in coping with structural disadvantage and the importance of an understanding of groupness and its role that goes beyond the dominant focus on intergroup contexts and categorization processes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Recent studies have demonstrated that the use of paramagnetic hepatobiliary contrast agents in the acquisition of magnetic resonance images remarkably improves the detection and differentiation of focal liver lesions, as compared with extracellular contrast agents. Paramagnetic hepatobiliary contrast agents initially show the perfusion of the lesions, as do extracellular agents, but delayed contrast-enhanced images can demonstrate contrast uptake by functional hepatocytes, providing further information for a better characterization of the lesions. Additionally, this intrinsic characteristic increases the accuracy in the detection of hepatocellular carcinomas and metastases, particularly the small-sized ones. Recently, a hepatobiliary contrast agent called gadolinium ethoxybenzyl dimeglumine, that is simply known as gadoxetic acid, was approved by the National Health Surveillance Agency for use in humans. The authors present a literature review and a practical approach of magnetic resonance imaging utilizing gadoxetic acid as contrast agent, based on patients' images acquired during their initial experiment.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Käyttöliittymä on rajapinta käyttäjän ja järjestelmän tarjoamien toimintojen välillä ja sen toimivuus vaikuttaa toimintojen suorittamiseen joko positiivisesti tai negatiivisesti. Täten sovelluksen suunnitteluvaiheessa on hyvä arvioida käyttöliittymän ja sen toimintojen laatua ja kokeilla ideoiden toimivuutta rakentamalla asiasta prototyyppejä. Prototypoinnilla voidaan tunnistaa ja korjata mahdolliset ongelmat jo suunnittelupöydällä. Tämä diplomityö käsittelee Web-sovelluksen kehityksen aikana toteutettua käyttöliittymän ja sen toimintojen prototypointia. Käyttöliittymien mallintamista voidaan toteuttaa erilaisilla menetelmillä, joita työssä käydään läpi teknologisista näkökulmista eli miten prototypointimenetelmiä voidaan soveltaa projektin eri vaiheissa. Prototypoinnin apuna käytettäviin työkaluihin luodaan lyhyt katsaus esitellen yleisellä tasolla muutamia eri sovelluskategorian ohjelmistoja ja lisäksi käsitellään suunnittelumallien hyödyntämistä. Työ osoittaa, että yleisiä prototypointimenetelmiä ja -periaatteita voidaan soveltaa Web-sovellusten prototypoinnissa. Prototypointi on hyödyllistä aloittaa luonnostelemalla ja jatkaa aikaisessa vaiheessa HTML-malleihin, joilla päästään lähelle toteutuksen teknologioita ja mallintamaan sovelluksen luonnetta, ilmettä, tuntumaa ja vuorovaikutusta. HTML-prototyypeistä voidaan jalostaa sekoitetun tarkkuuden malleja ja ne toimivat toteutuksen perustana. Jatkokehityksessä ideoita voidaan esittää useilla eri tarkkuuden tekniikoilla.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Työssä esitellään käytetyimpiä tuotantofilosofioita. Tuotantofilosofia on hyvin laaja käsite ja sen vuoksi myös jotkin esiteltävistä menetelmistä ovat hyvin kaukana toisistaan. Työ koostuu teoriaosiosta, jossa on esitelty kukin tuotantofilosofia ja lopuksi johtopäätöksiä-osiossa käsitellään sitä, kuinka menetelmät liittyvät toisiinsa. Työssä esitellään JIT/JOT-tuotanto, Lean-tuotanto, Monozukuri, Modulointi, Standardointi, Strategiatyö, Six Sigma, TQM, TPM, QFD, MFD, Simulointi, Digitaalinen valmistus, DFX ja ns. uudet tuotantofilosofiat. Eri menetelmistä löytyvää lähdemateriaalia on tarjolla monipuolisesti, josta johtuen menetelmistä on voitu esitellä vain pääpiirteet. Tuotantofilosofioiden avulla voidaan saavuttaa monia eri asioita. Osa menetelmistä on luotu tuotannon tehostamiseksi ja yksinkertaistamiseksi, osa puolestaan lisää tuotannon tai koko yrityksen laatutasoa ja osa puolestaan helpottaa tuotteiden suunnittelu-työtä. Moni esiteltävistä filosofioista ei istu yksinomaan yhteen edellä mainituista kategorioista vaan kattaa laajempia alueita pitäen sisällään jopa kaikkia kolmea mainittua tulosta. Näiden lisäksi työssä on esitelty lyhyesti uusia tuotantofilosofioita, jotka ovat hieman irrallisia kokonaisuuksia verrattuna muihin työssä esiteltäviin filosofioihin. Työn tarkoituksena on auttaa hahmottamaan suurta kokonaisuutta jonka tuotantofilosofiat tuottavat. On tärkeää osata hahmottaa filosofioiden riippuvuus toisistaan ja se, että otettaessa käyttöön jotain tuotantofilosofiaa, tarkoittaa se myös mahdollisesti monen muunkin asian huomioonottamista. Tätä näkökantaa selvennetään johtopäätöksissä.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this study was to simulate blood flow in thoracic human aorta and understand the role of flow dynamics in the initialization and localization of atherosclerotic plaque in human thoracic aorta. The blood flow dynamics in idealized and realistic models of human thoracic aorta were numerically simulated in three idealized and two realistic thoracic aorta models. The idealized models of thoracic aorta were reconstructed with measurements available from literature, and the realistic models of thoracic aorta were constructed by image processing Computed Tomographic (CT) images. The CT images were made available by South Karelia Central Hospital in Lappeenranta. The reconstruction of thoracic aorta consisted of operations, such as contrast adjustment, image segmentations, and 3D surface rendering. Additional design operations were performed to make the aorta model compatible for the numerical method based computer code. The image processing and design operations were performed with specialized medical image processing software. Pulsatile pressure and velocity boundary conditions were deployed as inlet boundary conditions. The blood flow was assumed homogeneous and incompressible. The blood was assumed to be a Newtonian fluid. The simulations with idealized models of thoracic aorta were carried out with Finite Element Method based computer code, while the simulations with realistic models of thoracic aorta were carried out with Finite Volume Method based computer code. Simulations were carried out for four cardiac cycles. The distribution of flow, pressure and Wall Shear Stress (WSS) observed during the fourth cardiac cycle were extensively analyzed. The aim of carrying out the simulations with idealized model was to get an estimate of flow dynamics in a realistic aorta model. The motive behind the choice of three aorta models with distinct features was to understand the dependence of flow dynamics on aorta anatomy. Highly disturbed and nonuniform distribution of velocity and WSS was observed in aortic arch, near brachiocephalic, left common artery, and left subclavian artery. On the other hand, the WSS profiles at the roots of branches show significant differences with geometry variation of aorta and branches. The comparison of instantaneous WSS profiles revealed that the model with straight branching arteries had relatively lower WSS compared to that in the aorta model with curved branches. In addition to this, significant differences were observed in the spatial and temporal profiles of WSS, flow, and pressure. The study with idealized model was extended to study blood flow in thoracic aorta under the effects of hypertension and hypotension. One of the idealized aorta models was modified along with the boundary conditions to mimic the thoracic aorta under the effects of hypertension and hypotension. The results of simulations with realistic models extracted from CT scans demonstrated more realistic flow dynamics than that in the idealized models. During systole, the velocity in ascending aorta was skewed towards the outer wall of aortic arch. The flow develops secondary flow patterns as it moves downstream towards aortic arch. Unlike idealized models, the distribution of flow was nonplanar and heavily guided by the artery anatomy. Flow cavitation was observed in the aorta model which was imaged giving longer branches. This could not be properly observed in the model with imaging containing a shorter length for aortic branches. The flow circulation was also observed in the inner wall of the aortic arch. However, during the diastole, the flow profiles were almost flat and regular due the acceleration of flow at the inlet. The flow profiles were weakly turbulent during the flow reversal. The complex flow patterns caused a non-uniform distribution of WSS. High WSS was distributed at the junction of branches and aortic arch. Low WSS was distributed at the proximal part of the junction, while intermedium WSS was distributed in the distal part of the junction. The pulsatile nature of the inflow caused oscillating WSS at the branch entry region and inner curvature of aortic arch. Based on the WSS distribution in the realistic model, one of the aorta models was altered to induce artificial atherosclerotic plaque at the branch entry region and inner curvature of aortic arch. Atherosclerotic plaque causing 50% blockage of lumen was introduced in brachiocephalic artery, common carotid artery, left subclavian artery, and aortic arch. The aim of this part of the study was first to study the effect of stenosis on flow and WSS distribution, understand the effect of shape of atherosclerotic plaque on flow and WSS distribution, and finally to investigate the effect of lumen blockage severity on flow and WSS distributions. The results revealed that the distribution of WSS is significantly affected by plaque with mere 50% stenosis. The asymmetric shape of stenosis causes higher WSS in branching arteries than in the cases with symmetric plaque. The flow dynamics within thoracic aorta models has been extensively studied and reported here. The effects of pressure and arterial anatomy on the flow dynamic were investigated. The distribution of complex flow and WSS is correlated with the localization of atherosclerosis. With the available results we can conclude that the thoracic aorta, with complex anatomy is the most vulnerable artery for the localization and development of atherosclerosis. The flow dynamics and arterial anatomy play a role in the localization of atherosclerosis. The patient specific image based models can be used to diagnose the locations in the aorta vulnerable to the development of arterial diseases such as atherosclerosis.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Due to the different dynamics required for organizations to serve the emerging market which contains billions of people at the bottom of the pyramid (BOP) coupled with the increasing desire for organizations to grow and be more multinational, organizations need to continually innovate. However, the tendency for large and established companies to ignore the BOP market and rather focus on existing markets, gives an indication of the existence of a vulnerability that potentially disruptive innovations from the BOP will not be recognized in good time for a counter measure. This can be deduced from the fact that good management practice advocates that managers should learn and listen to their customers. Therefore majority of the large existing companies continually focus on their main customer/market with sustaining innovations which leaves aspiring new entrants with an underserved BOP market to experiment with. With the aid of research interviews and an agent-based model (ABM) simulation, this thesis examines the attributes of BOP innovations that can qualify them as disruptive and the possibilities of tangible disruptive innovations arising from the bottom of the pyramid and their underlying drivers. The thesis Furthermore, examines the associated impact of such innovations on the future sustainability of established large companies that are operating in the developed world, particularly those with a primary focus which is targeted towards the market at the top of the pyramid (TOP). Additionally, with the use of a scenario planning model, the research provides an evaluation of the possible evolution and potential sustainability impacts that could emerge, from the interplay of innovations at the two pyramidal market levels and the chosen market focus of organizations – TOP or BOP. Using four scenario quadrants, the thesis demonstrates the resulting possibilities from the interaction between the rate of innovations and the segment focused on by organizations with disruptive era characterizing the paradigm shift quadrant. Furthermore, a mathematical model and two theoretical propositions are developed for further research. As recommendations, the thesis also extends the ambidextrous organizational theory, business model innovation and portfolio diversification as plausible recommendations to limit a catastrophic impact, resulting from disruptive innovations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Operating in business-to-business markets requires an in-depth understanding on business networks. Actions and reactions made to compete in markets are fundamentally based on managers‘ subjective perceptions of the network. However, an amalgamation of these individual perceptions, termed a network picture, to a common company level shared understanding on that network, known as network insight, is found to be a substantial challenge for companies. A company‘s capability to enhance common network insight is even argued to lead competitive advantage. Especially companies with value creating logics that require wide comprehension of and collaborating in networks, such as solution business, are necessitated to develop advanced network insight. According to the extant literature, dispersed pieces of atomized network pictures can be unified to a common network insight through a process of amalgamation that comprises barriers/drivers of multilateral exchange, manifold rationality, and recursive time. However, the extant body of literature appears to lack an understanding on the role of internal communication in the development of network insight. Nonetheless, the extant understanding on the amalgamation process indicates that internal communication plays a substantial role in the development of company level network insight. The purpose of the present thesis is to enhance understanding on internal communication in the amalgamation of network pictures to develop network insight in the solution business setting, which was chosen to represent business-to-business value creating logic that emphasizes the capability to understand and utilize networks. Thus, in solution business the role of succeeding in the amalgamation process is expected to emphasize. The study combines qualitative and quantitative research by means of various analytical methods including multiple case analysis, simulation, and social network analysis. Approaching the nascent research topic with differing perspectives and means provides a broader insight on the phenomenon. The study provides empirical evidence from Finnish business-to-business companies which operate globally. The empirical data comprise interviews (n=28) with managers of three case companies. In addition the data includes a questionnaire (n=23) collected mainly for the purpose of social network analysis. In addition, the thesis includes a simulation study more specifically achieved by means of agent based modeling. The findings of the thesis shed light on the role of internal communication in the amalgamation process, contributing to the emergent discussion of network insights and thus to the industrial marketing research. In addition, the thesis increases understanding on internal communication in the change process to solution business, a supplier‘s internal communication in its matrix organization structure during a project sales process, key barriers and drivers that influence internal communication in project sales networks, perceived power within industrial project sales, and the revisioning of network pictures. According to the findings, internal communication is found to play a substantial role in the amalgamation process. First, it is suggested that internal communication is a base of multilateral exchange. Second, it is suggested that internal communication intensifies and maintains manifold rationality. Third, internal communication is needed to explicate the usually differing time perspectives of others and thus it is suggested that internal communication has role as the explicator of recursive time. Furthermore, the role of an efficient amalgamation process is found to be emphasized in solutions business as it requires a more advanced network insight for cross-functional collaboration. Finally, the thesis offers several managerial implications for industrial suppliers to enhance the amalgamation process when operating in solution business.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This study examines the structure of the Russian Reflexive Marker ( ся/-сь) and offers a usage-based model building on Construction Grammar and a probabilistic view of linguistic structure. Traditionally, reflexive verbs are accounted for relative to non-reflexive verbs. These accounts assume that linguistic structures emerge as pairs. Furthermore, these accounts assume directionality where the semantics and structure of a reflexive verb can be derived from the non-reflexive verb. However, this directionality does not necessarily hold diachronically. Additionally, the semantics and the patterns associated with a particular reflexive verb are not always shared with the non-reflexive verb. Thus, a model is proposed that can accommodate the traditional pairs as well as for the possible deviations without postulating different systems. A random sample of 2000 instances marked with the Reflexive Marker was extracted from the Russian National Corpus and the sample used in this study contains 819 unique reflexive verbs. This study moves away from the traditional pair account and introduces the concept of Neighbor Verb. A neighbor verb exists for a reflexive verb if they share the same phonological form excluding the Reflexive Marker. It is claimed here that the Reflexive Marker constitutes a system in Russian and the relation between the reflexive and neighbor verbs constitutes a cross-paradigmatic relation. Furthermore, the relation between the reflexive and the neighbor verb is argued to be of symbolic connectivity rather than directionality. Effectively, the relation holding between particular instantiations can vary. The theoretical basis of the present study builds on this assumption. Several new variables are examined in order to systematically model variability of this symbolic connectivity, specifically the degree and strength of connectivity between items. In usage-based models, the lexicon does not constitute an unstructured list of items. Instead, items are assumed to be interconnected in a network. This interconnectedness is defined as Neighborhood in this study. Additionally, each verb carves its own niche within the Neighborhood and this interconnectedness is modeled through rhyme verbs constituting the degree of connectivity of a particular verb in the lexicon. The second component of the degree of connectivity concerns the status of a particular verb relative to its rhyme verbs. The connectivity within the neighborhood of a particular verb varies and this variability is quantified by using the Levenshtein distance. The second property of the lexical network is the strength of connectivity between items. Frequency of use has been one of the primary variables in functional linguistics used to probe this. In addition, a new variable called Constructional Entropy is introduced in this study building on information theory. It is a quantification of the amount of information carried by a particular reflexive verb in one or more argument constructions. The results of the lexical connectivity indicate that the reflexive verbs have statistically greater neighborhood distances than the neighbor verbs. This distributional property can be used to motivate the traditional observation that the reflexive verbs tend to have idiosyncratic properties. A set of argument constructions, generalizations over usage patterns, are proposed for the reflexive verbs in this study. In addition to the variables associated with the lexical connectivity, a number of variables proposed in the literature are explored and used as predictors in the model. The second part of this study introduces the use of a machine learning algorithm called Random Forests. The performance of the model indicates that it is capable, up to a degree, of disambiguating the proposed argument construction types of the Russian Reflexive Marker. Additionally, a global ranking of the predictors used in the model is offered. Finally, most construction grammars assume that argument construction form a network structure. A new method is proposed that establishes generalization over the argument constructions referred to as Linking Construction. In sum, this study explores the structural properties of the Russian Reflexive Marker and a new model is set forth that can accommodate both the traditional pairs and potential deviations from it in a principled manner.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In accordance with the Moore's law, the increasing number of on-chip integrated transistors has enabled modern computing platforms with not only higher processing power but also more affordable prices. As a result, these platforms, including portable devices, work stations and data centres, are becoming an inevitable part of the human society. However, with the demand for portability and raising cost of power, energy efficiency has emerged to be a major concern for modern computing platforms. As the complexity of on-chip systems increases, Network-on-Chip (NoC) has been proved as an efficient communication architecture which can further improve system performances and scalability while reducing the design cost. Therefore, in this thesis, we study and propose energy optimization approaches based on NoC architecture, with special focuses on the following aspects. As the architectural trend of future computing platforms, 3D systems have many bene ts including higher integration density, smaller footprint, heterogeneous integration, etc. Moreover, 3D technology can signi cantly improve the network communication and effectively avoid long wirings, and therefore, provide higher system performance and energy efficiency. With the dynamic nature of on-chip communication in large scale NoC based systems, run-time system optimization is of crucial importance in order to achieve higher system reliability and essentially energy efficiency. In this thesis, we propose an agent based system design approach where agents are on-chip components which monitor and control system parameters such as supply voltage, operating frequency, etc. With this approach, we have analysed the implementation alternatives for dynamic voltage and frequency scaling and power gating techniques at different granularity, which reduce both dynamic and leakage energy consumption. Topologies, being one of the key factors for NoCs, are also explored for energy saving purpose. A Honeycomb NoC architecture is proposed in this thesis with turn-model based deadlock-free routing algorithms. Our analysis and simulation based evaluation show that Honeycomb NoCs outperform their Mesh based counterparts in terms of network cost, system performance as well as energy efficiency.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Työn tavoitteena oli vastata ensisijaisesti kysymykseen, voidaanko projektiliiketoiminnan kassavirtoja ennustaa 3-15 kuukauden aikavälillä ja jos voidaan, niin miten ja millä tarkkuudella. Tutkimus toteutettiin teoriatutkimuksena aihepiiristä ja tutkimuksen pohjalta luotiin malli kassavirtojen ennustamiseen kohdeyritykselle 3-15 kuukauden aikavälille. Mallin laatimiseksi oli hyödynnettävissä viiden vuoden aineistot kohdeyrityksen kassavirroista, budjetista ja liiketoiminnan toteumatiedoista. Työn teoriaosiossa tutkittiin kirjallisuuden pohjalta projektiliiketoimintaa, budjetointia sekä kassavirtoja ja niiden ennustamista. Tämän jälkeen teorian pohjalta rakennettiin kohdeyritykselle historiatietoihin perustuva malli kassavirtojen ennustamiseksi. Mallia rakennettaessa määritettiin ensimmäiseksi merkittävimmät kassavirran komponentit, minkä jälkeen niille laadittiin ennustemenetelmät. Samalla arvioitiin millä tarkkuudella projektilähtöisen liiketoiminnan kassavirtoja pystytään ennustamaan. Tutkimuksen tuloksena oli historiatietoihin pohjautuva ennustemalli kohdeyritykselle. Mallilla tehtyjen testien pohjalta voitiin todeta, että projektilähtöisen liiketoiminnan kassavirtoja pystytään ennustamaan melko hyvällä tarkkuudella, ennustaminen ei kuitenkaan ole niin luotettavaa, kuin jos ennustettaisiin tasaisemmin kehittyvän liiketoiminnan kassavirtoja. Historiaan pohjautuvaa mallia käytettäessä pitää myös muistaa, että mikään ei takaa historian toistumista tulevaisuudessa.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis studies the development of service offering model that creates added-value for customers in the field of logistics services. The study focusses on offering classification and structures of model. The purpose of model is to provide value-added solutions for customers and enable superior service experience. The aim of thesis is to define what customers expect from logistics solution provider and what value customers appreciate so greatly that they could invest in value-added services. Value propositions, costs structures of offerings and appropriate pricing methods are studied. First, literature review of creating solution business model and customer value is conducted. Customer value is found out with customer interviews and qualitative empiric data is used. To exploit expertise knowledge of logistics, innovation workshop tool is utilized. Customers and experts are involved in the design process of model. As a result of thesis, three-level value-added service offering model is created based on empiric and theoretical data. Offerings with value propositions are proposed and the level of model reflects the deepness of customer-provider relationship and the amount of added value. Performance efficiency improvements and cost savings create the most added value for customers. Value-based pricing methods, such as performance-based models are suggested to apply. Results indicate the interest of benefitting networks and partnership in field of logistics services. Networks development is proposed to be investigated further.