34 resultados para Mathematics. Trigonometric Functions. Geogebra
em Doria (National Library of Finland DSpace Services) - National Library of Finland, Finland
Resumo:
This PhD thesis in Mathematics belongs to the field of Geometric Function Theory. The thesis consists of four original papers. The topic studied deals with quasiconformal mappings and their distortion theory in Euclidean n-dimensional spaces. This theory has its roots in the pioneering papers of F. W. Gehring and J. Väisälä published in the early 1960’s and it has been studied by many mathematicians thereafter. In the first paper we refine the known bounds for the so-called Mori constant and also estimate the distortion in the hyperbolic metric. The second paper deals with radial functions which are simple examples of quasiconformal mappings. These radial functions lead us to the study of the so-called p-angular distance which has been studied recently e.g. by L. Maligranda and S. Dragomir. In the third paper we study a class of functions of a real variable studied by P. Lindqvist in an influential paper. This leads one to study parametrized analogues of classical trigonometric and hyperbolic functions which for the parameter value p = 2 coincide with the classical functions. Gaussian hypergeometric functions have an important role in the study of these special functions. Several new inequalities and identities involving p-analogues of these functions are also given. In the fourth paper we study the generalized complete elliptic integrals, modular functions and some related functions. We find the upper and lower bounds of these functions, and those bounds are given in a simple form. This theory has a long history which goes back two centuries and includes names such as A. M. Legendre, C. Jacobi, C. F. Gauss. Modular functions also occur in the study of quasiconformal mappings. Conformal invariants, such as the modulus of a curve family, are often applied in quasiconformal mapping theory. The invariants can be sometimes expressed in terms of special conformal mappings. This fact explains why special functions often occur in this theory.
Resumo:
Selostus: Sian kasvuominaisuuksien perinnölliset tunnusluvut arvioituna kolmannen asteen polynomifunktion avulla
Resumo:
Abstract
Resumo:
Yrityksen sisäisten rajapintojen tunteminen mahdollistaa tiedonvaihdon hallinnan läpi organisaation. Idean muokkaaminen kannattavaksi innovaatioksi edellyttää organisaation eri osien läpi kulkevaa saumatonta prosessiketjua sekä tietovirtaa. Tutkielman tavoitteena oli mallintaa organisaation kahden toiminnallisesti erilaisen osan välinen tiedon vaihto. Tiedon vaihto kuvattiin rajapintana, tietoliittymänä. Kolmiulotteinen organisaatiomalli muodosti tutkimuksen pääteorian. Se kytkettiin yrityksen tuotanto- ja myyntiosiin, kuten myös BestServ-projektin kehittämään uuteen palvelujen kehittämisen prosessiin. Uutta palvelujen kehittämisen prosessia laajennettiin ISO/IEC 15288 standardin kuvaamalla prosessimallilla. Yritysarkkitehtuurikehikoita käytettiin mallintamisen perustana. Tietoliittymä nimenä kuvastaa näkemystä siitä, että tieto [tietämys] on olemukseltaan yksilöiden tai ryhmien välistä. Mallinnusmenetelmät eivät kuitenkaan vielä mahdollista tietoon [tietämykseen] liittyvien kaikkien ominaisuuksien mallintamista. Tietoliittymän malli koostuu kolmesta osasta, joista kaksi esitetään graafisessa muodossa ja yksi taulukkona. Mallia voidaan käyttää itsenäisesti tai osana yritysarkkitehtuuria. Teollisessa palveluliiketoiminnassa sekä tietoliittymän mallinnusmenetelmä että sillä luotu malli voivat auttaa konepajateollisuuden yritystä ymmärtämään yrityksen kehittämistarpeet ja -kohteet, kun se haluaa palvelujen tuottamisella suuremman roolin asiakasyrityksen liiketoiminnassa. Tietoliittymän mallia voidaan käyttää apuna organisaation tietovarannon ja tietämyksen mallintamisessa sekä hallinnassa ja näin pyrkiä yhdistämään ne yrityksen strategiaa palvelevaksi kokonaisuudeksi. Tietoliittymän mallinnus tarjoaa tietojohtamisen kauppatieteelliselle tutkimukselle menetelmällisyyden tutkia innovaatioiden hallintaa sekä organisaation uudistumiskykyä. Kumpikin tutkimusalue tarvitsevat tarkempaa tietoa ja mahdollisuuksia hallita tietovirtoja, tiedon vaihtoa sekä organisaation tietovarannon käyttöä.
Resumo:
The thesis studies role based access control and its suitability in the enterprise environment. The aim is to research how extensively role based access control can be implemented in the case organization and how it support organization’s business and IT functions. This study points out the enterprise’s needs for access control, factors of access control in the enterprise environment and requirements for implementation and the benefits and challenges it brings along. To find the scope how extensively role based access control can be implemented into the case organization, firstly is examined the actual state of access control. Secondly is defined a rudimentary desired state (how things should be) and thirdly completed it by using the results of the implementation of role based access control application. The study results the role model for case organization unit, and the building blocks and the framework for the organization wide implementation. Ultimate value for organization is delivered by facilitating the normal operations of the organization whilst protecting its information assets.
Resumo:
Recent advances in machine learning methods enable increasingly the automatic construction of various types of computer assisted methods that have been difficult or laborious to program by human experts. The tasks for which this kind of tools are needed arise in many areas, here especially in the fields of bioinformatics and natural language processing. The machine learning methods may not work satisfactorily if they are not appropriately tailored to the task in question. However, their learning performance can often be improved by taking advantage of deeper insight of the application domain or the learning problem at hand. This thesis considers developing kernel-based learning algorithms incorporating this kind of prior knowledge of the task in question in an advantageous way. Moreover, computationally efficient algorithms for training the learning machines for specific tasks are presented. In the context of kernel-based learning methods, the incorporation of prior knowledge is often done by designing appropriate kernel functions. Another well-known way is to develop cost functions that fit to the task under consideration. For disambiguation tasks in natural language, we develop kernel functions that take account of the positional information and the mutual similarities of words. It is shown that the use of this information significantly improves the disambiguation performance of the learning machine. Further, we design a new cost function that is better suitable for the task of information retrieval and for more general ranking problems than the cost functions designed for regression and classification. We also consider other applications of the kernel-based learning algorithms such as text categorization, and pattern recognition in differential display. We develop computationally efficient algorithms for training the considered learning machines with the proposed kernel functions. We also design a fast cross-validation algorithm for regularized least-squares type of learning algorithm. Further, an efficient version of the regularized least-squares algorithm that can be used together with the new cost function for preference learning and ranking tasks is proposed. In summary, we demonstrate that the incorporation of prior knowledge is possible and beneficial, and novel advanced kernels and cost functions can be used in algorithms efficiently.
Resumo:
Probiotic lactobacilli and bifidobacteria in the mouth – in vitro studies on saliva-mediated functions and acid production Probiotics are viable bacteria which, when used in adequate amounts, are beneficial to the health of the host. Although most often related to intestinal health, probiotic bacteria can be found also in the mouth after consumption of products that contain them. This study aimed at evaluating the oral effects of probiotic bacteria already in commercial use. In a series of in vitro studies, the oral colonisation potential of different probiotic bacteria, their acid production and potential saliva-mediated effects on oral microbial ecology were investigated. The latter included effects on the salivary pellicle, the adhesion of other bacteria, and the activation of the peroxidase system. Streptococcus mutans, Streptococcus gordonii, Aggregatibacter actinomycetemcomitans and Helicobacter pylori were used as bacterial indicators of the studied phenomena. There were significant differences between the probiotic strains in their colonisation potential. They all were acidogenic, although using different sugars and sugar alcohols. However, their acid production could be inhibited by the peroxidase system. Based on the results, it can be suggested that probiotic bacteria might influence the oral microbiota by different, partly species or strain-specific means. These include the inhibition of bacterial adhesion, modification of the enamel pellicle, antimicrobial activity, and activation of the peroxidase system. To conclude, probiotic strains differed from each other in their colonisation potential and other oral effects as evaluated in vitro. Both positive and potentially harmful effects were observed, but the significance of the perceived results needs to be further evaluated in vivo.
Resumo:
The thesis discusses the regulation of foodstuffs and medicines, and particularly the regulation of functional foods. Legal systems investigated are the EU and China. Both are members of the WTO and Codex Alimentarius, which binds European and Chinese rules together. The study uses three Chinese berries as case examples of how product development faces regulation in practice. The berries have traditional uses as herbal medicines. Europe and China have similar nutrition problems to be resolved, such as obesity, cardiovascular disease, and diabetes. The three berries might be suitable raw materials for functional foods. Consumer products with health-enhancing functions, such as lowering blood pressure, might legally be classifi ed either as foodstuffs or medicines. The classifi cation will depend on functions and presentation of the product. In our opinion, food and medicine regulation should come closer together so the classifi cation issue would no longer be an issue. Safety of both foodstuffs and medicines is strictly regulated. With medicines, safety is a more relative concept, where benefi ts of the product are compared to side-effects in thorough scientifi c tests and trials. Foods, on the other hand, are not allowed to have side-effects. Hygiene rules and rules on the use of chemicals apply. In China, food safety is currently at focus as China has had several severe food scandals. Newly developed foods are called novel foods, and are specifi cally regulated. The current European novel food regulation from 1997 treats traditional third country products as novel. The Chinese regulation of 2007 also defi nes novel foods as something unfamiliar to a Chinese consumer. The concepts of novel food thus serve a protectionist purpose. As regards marketing, foods are allowed to bear health claims, whereas medicines bear medicinal claims. The separation is legally strict: foods are not to be presented as having medicinal functions. European nutrition and health claim regulation exists since 2006. China also has its regulation on health foods, listing the permitted claims and how to substantiate them. Health claims are allowed only on health foods. The European rules on medicines include separate categories for herbal medicines, traditional herbal medicines, and homeopathic medicines, where there are differing requirements for scientifi c substantiation. The scientifi c and political grounds for the separate categories provoke criticism. At surface, the Chinese legal system seems similar to the European one. To facilitate trade, China has enacted modern laws. Laws are needed as the country moves from planned economy to market economy: ‘rule of law’ needs to replace ‘rule of man’. Instead of being citizens, Chinese people long were subordinates to the Emperor. Confucius himself advised to avoid confl ict. Still, Chinese people do not and cannot always trust the legal system, as laws are enforced in an inconsistent manner, and courts are weak. In China, there have been problems with confl icting national and local laws. In Europe, the competence of the EU vs. the competence of the Member States is still not resolved, even though the European Commission often states that free trade requires harmonisation. Food and medicine regulation is created by international organisations, food and medicine control agencies, standards agencies, companies and their organisations. Regulation can be divided in ‘hard law’ and ‘soft law’. One might claim that hard law is in crisis, as soft law is gaining importance. If law is out of fashion, regulation certainly isn’t. In the future, ‘law’ might mean a process where rules and incentives are created by states, NGOs, companies, consumers, and other stakeholders. ‘Law’ might thus refer to a constant negotiation between public and private actors. Legal principles such as transparency, equal treatment, and the right to be heard would still be important.
Resumo:
Electricity distribution network operation (NO) models are challenged as they are expected to continue to undergo changes during the coming decades in the fairly developed and regulated Nordic electricity market. Network asset managers are to adapt to competitive technoeconomical business models regarding the operation of increasingly intelligent distribution networks. Factors driving the changes for new business models within network operation include: increased investments in distributed automation (DA), regulative frameworks for annual profit limits and quality through outage cost, increasing end-customer demands, climatic changes and increasing use of data system tools, such as Distribution Management System (DMS). The doctoral thesis addresses the questions a) whether there exist conditions and qualifications for competitive markets within electricity distribution network operation and b) if so, identification of limitations and required business mechanisms. This doctoral thesis aims to provide an analytical business framework, primarily for electric utilities, for evaluation and development purposes of dedicated network operation models to meet future market dynamics within network operation. In the thesis, the generic build-up of a business model has been addressed through the use of the strategicbusiness hierarchy levels of mission, vision and strategy for definition of the strategic direction of the business followed by the planning, management and process execution levels of enterprisestrategy execution. Research questions within electricity distribution network operation are addressed at the specified hierarchy levels. The results of the research represent interdisciplinary findings in the areas of electrical engineering and production economics. The main scientific contributions include further development of the extended transaction cost economics (TCE) for government decisions within electricity networks and validation of the usability of the methodology for the electricity distribution industry. Moreover, DMS benefit evaluations in the thesis based on the outage cost calculations propose theoretical maximum benefits of DMS applications equalling roughly 25% of the annual outage costs and 10% of the respective operative costs in the case electric utility. Hence, the annual measurable theoretical benefits from the use of DMS applications are considerable. The theoretical results in the thesis are generally validated by surveys and questionnaires.
Resumo:
This study presents the information required to describe the machine and device resources in the turret punch press environment which are needed for the development of the analysing method for automated production. The description of product and device resources and their interconnectedness is the starting point for method comparison the development of expenses, production planning and the performance of optimisation. The manufacturing method cannot be optimized unless the variables and their interdependence are known. Sheet metal parts in particular may then become remarkably complex, and their automatic manufacture may be difficult or, with some automatic equipment, even impossible if not know manufacturing properties. This thesis consists of three main elements, which constitute the triangulation. In the first phase of triangulation, the manufacture occuring on a turret punch press is examined in order to find the factors that affect the efficiency of production. In the second phase of triangulation, the manufacturability of products on turret punch presses is examined through a set of laboratory tests. The third phase oftriangulation involves an examination of five industry parts. The main key findings of this study are: all possible efficiency in high automation level machining cannot be achieved unless the raw materials used in production and the dependencies of the machine and tools are well known. Machine-specific manufacturability factors for turret punch presses were not taken into account in the industrial case samples. On the grounds of the performed tests and industrial case samples, the designer of a sheet metal product can directly influence the machining time, material loss, energy consumption and the number of tools required on a turret punch press by making decisions in the way presented in the hypothesis of thisstudy. The sheet metal parts to be produced can be optimised to bemanufactured on a turret punch press when the material to be used and the kinds of machine and tool options available are known. This provides in-depth knowledge of the machine and tool properties machine and tool-specifically. None of the optimisation starting points described here is a separate entity; instead, they are all connected to each other.
Resumo:
Programming and mathematics are core areas of computer science (CS) and consequently also important parts of CS education. Introductory instruction in these two topics is, however, not without problems. Studies show that CS students find programming difficult to learn and that teaching mathematical topics to CS novices is challenging. One reason for the latter is the disconnection between mathematics and programming found in many CS curricula, which results in students not seeing the relevance of the subject for their studies. In addition, reports indicate that students' mathematical capability and maturity levels are dropping. The challenges faced when teaching mathematics and programming at CS departments can also be traced back to gaps in students' prior education. In Finland the high school curriculum does not include CS as a subject; instead, focus is on learning to use the computer and its applications as tools. Similarly, many of the mathematics courses emphasize application of formulas, while logic, formalisms and proofs, which are important in CS, are avoided. Consequently, high school graduates are not well prepared for studies in CS. Motivated by these challenges, the goal of the present work is to describe new approaches to teaching mathematics and programming aimed at addressing these issues: Structured derivations is a logic-based approach to teaching mathematics, where formalisms and justifications are made explicit. The aim is to help students become better at communicating their reasoning using mathematical language and logical notation at the same time as they become more confident with formalisms. The Python programming language was originally designed with education in mind, and has a simple syntax compared to many other popular languages. The aim of using it in instruction is to address algorithms and their implementation in a way that allows focus to be put on learning algorithmic thinking and programming instead of on learning a complex syntax. Invariant based programming is a diagrammatic approach to developing programs that are correct by construction. The approach is based on elementary propositional and predicate logic, and makes explicit the underlying mathematical foundations of programming. The aim is also to show how mathematics in general, and logic in particular, can be used to create better programs.