68 resultados para Binary Image Representation
em Doria (National Library of Finland DSpace Services) - National Library of Finland, Finland
Resumo:
This thesis deals with distance transforms which are a fundamental issue in image processing and computer vision. In this thesis, two new distance transforms for gray level images are presented. As a new application for distance transforms, they are applied to gray level image compression. The new distance transforms are both new extensions of the well known distance transform algorithm developed by Rosenfeld, Pfaltz and Lay. With some modification their algorithm which calculates a distance transform on binary images with a chosen kernel has been made to calculate a chessboard like distance transform with integer numbers (DTOCS) and a real value distance transform (EDTOCS) on gray level images. Both distance transforms, the DTOCS and EDTOCS, require only two passes over the graylevel image and are extremely simple to implement. Only two image buffers are needed: The original gray level image and the binary image which defines the region(s) of calculation. No other image buffers are needed even if more than one iteration round is performed. For large neighborhoods and complicated images the two pass distance algorithm has to be applied to the image more than once, typically 3 10 times. Different types of kernels can be adopted. It is important to notice that no other existing transform calculates the same kind of distance map as the DTOCS. All the other gray weighted distance function, GRAYMAT etc. algorithms find the minimum path joining two points by the smallest sum of gray levels or weighting the distance values directly by the gray levels in some manner. The DTOCS does not weight them that way. The DTOCS gives a weighted version of the chessboard distance map. The weights are not constant, but gray value differences of the original image. The difference between the DTOCS map and other distance transforms for gray level images is shown. The difference between the DTOCS and EDTOCS is that the EDTOCS calculates these gray level differences in a different way. It propagates local Euclidean distances inside a kernel. Analytical derivations of some results concerning the DTOCS and the EDTOCS are presented. Commonly distance transforms are used for feature extraction in pattern recognition and learning. Their use in image compression is very rare. This thesis introduces a new application area for distance transforms. Three new image compression algorithms based on the DTOCS and one based on the EDTOCS are presented. Control points, i.e. points that are considered fundamental for the reconstruction of the image, are selected from the gray level image using the DTOCS and the EDTOCS. The first group of methods select the maximas of the distance image to new control points and the second group of methods compare the DTOCS distance to binary image chessboard distance. The effect of applying threshold masks of different sizes along the threshold boundaries is studied. The time complexity of the compression algorithms is analyzed both analytically and experimentally. It is shown that the time complexity of the algorithms is independent of the number of control points, i.e. the compression ratio. Also a new morphological image decompression scheme is presented, the 8 kernels' method. Several decompressed images are presented. The best results are obtained using the Delaunay triangulation. The obtained image quality equals that of the DCT images with a 4 x 4
Resumo:
Tässä työssä raportoidaan harjoitustyön kehittäminen ja toteuttaminen Aktiivisen- ja robottinäön kurssille. Harjoitustyössä suunnitellaan ja toteutetaan järjestelmä joka liikuttaa kappaleita robottikäsivarrella kolmiuloitteisessa avaruudessa. Kappaleidenpaikkojen määrittämiseen järjestelmä käyttää digitaalisia kuvia. Tässä työssä esiteltävässä harjoitustyötoteutuksessa käytettiin raja-arvoistusta HSV-väriavaruudessa kappaleiden segmentointiin kuvasta niiden värien perusteella. Segmentoinnin tuloksena saatavaa binäärikuvaa suodatettiin mediaanisuotimella kuvan häiriöiden poistamiseksi. Kappaleen paikkabinäärikuvassa määritettiin nimeämällä yhtenäisiä pikseliryhmiä yhtenäisen alueen nimeämismenetelmällä. Kappaleen paikaksi määritettiin suurimman nimetyn pikseliryhmän paikka. Kappaleiden paikat kuvassa yhdistettiin kolmiuloitteisiin koordinaatteihin kalibroidun kameran avulla. Järjestelmä liikutti kappaleita niiden arvioitujen kolmiuloitteisten paikkojen perusteella.
Resumo:
The aim of this study is to analyse the content of the interdisciplinary conversations in Göttingen between 1949 and 1961. The task is to compare models for describing reality presented by quantum physicists and theologians. Descriptions of reality indifferent disciplines are conditioned by the development of the concept of reality in philosophy, physics and theology. Our basic problem is stated in the question: How is it possible for the intramental image to match the external object?Cartesian knowledge presupposes clear and distinct ideas in the mind prior to observation resulting in a true correspondence between the observed object and the cogitative observing subject. The Kantian synthesis between rationalism and empiricism emphasises an extended character of representation. The human mind is not a passive receiver of external information, but is actively construing intramental representations of external reality in the epistemological process. Heidegger's aim was to reach a more primordial mode of understanding reality than what is possible in the Cartesian Subject-Object distinction. In Heidegger's philosophy, ontology as being-in-the-world is prior to knowledge concerning being. Ontology can be grasped only in the totality of being (Dasein), not only as an object of reflection and perception. According to Bohr, quantum mechanics introduces an irreducible loss in representation, which classically understood is a deficiency in knowledge. The conflicting aspects (particle and wave pictures) in our comprehension of physical reality, cannot be completely accommodated into an entire and coherent model of reality. What Bohr rejects is not realism, but the classical Einsteinian version of it. By the use of complementary descriptions, Bohr tries to save a fundamentally realistic position. The fundamental question in Barthian theology is the problem of God as an object of theological discourse. Dialectics is Barth¿s way to express knowledge of God avoiding a speculative theology and a human-centred religious self-consciousness. In Barthian theology, the human capacity for knowledge, independently of revelation, is insufficient to comprehend the being of God. Our knowledge of God is real knowledge in revelation and our words are made to correspond with the divine reality in an analogy of faith. The point of the Bultmannian demythologising programme was to claim the real existence of God beyond our faculties. We cannot simply define God as a human ideal of existence or a focus of values. The theological programme of Bultmann emphasised the notion that we can talk meaningfully of God only insofar as we have existential experience of his intervention. Common to all these twentieth century philosophical, physical and theological positions, is a form of anti-Cartesianism. Consequently, in regard to their epistemology, they can be labelled antirealist. This common insight also made it possible to find a common meeting point between the different disciplines. In this study, the different standpoints from all three areas and the conversations in Göttingen are analysed in the frameworkof realism/antirealism. One of the first tasks in the Göttingen conversations was to analyse the nature of the likeness between the complementary structures inquantum physics introduced by Niels Bohr and the dialectical forms in the Barthian doctrine of God. The reaction against epistemological Cartesianism, metaphysics of substance and deterministic description of reality was the common point of departure for theologians and physicists in the Göttingen discussions. In his complementarity, Bohr anticipated the crossing of traditional epistemic boundaries and the generalisation of epistemological strategies by introducing interpretative procedures across various disciplines.
Resumo:
I den första delen av den här avhandlingen presenteras en bildens genealogi. Den skildrar hur begreppen för bilden, seendet och jaget utvecklades i relation till varandra i en specifik vetenskaplig och filosofisk kontext. Berättelsen sträcker sig från den tidiga renässansen och det perspektivistiska måleriet, till fotografiets födelse och positivismen. Den här utvecklingen medförde en form av reduktionism i vilken jagets roll – betydelsen av den mänskliga psykologin, vårt omdöme, vår uppmärksamhet och vår vilja – blev förbisedd. Inom den här tanketraditionen uppstod en förskjutning, från en förståelse av bilden som en representation av det tredimensionella rummet på en tvådimensionell yta, till en uppfattning om bilden som en genomskinlig ruta, ett fönster ut mot världen. Idén om avbildningen som en neutral ”blick från ingenstans” kom att förstärka en skeptisk hållning till kommunikation, dialog och vittnesmål och därmed även undergräva vår tillit till varandra och följaktligen vår tillit till oss själva. I den andra delen erbjuder författaren ett alternativ till den tanketradition som behandlas i den första delen. Det som blev förbisett i uppfattningen om en blick från ingenstans var att bilden är ett hjälpmedel då vi bearbetar vårt synfält. Bilden hjälper oss att dela vår syn på saker. Genom den här uppgiften av att dela blir bilden riktningsgivande i våra försök att orientera oss i världen. Jag kan stå bredvid en annan människa och se vad hon ser, men jag vet inte nödvändigtvis hur hon uppfattar det vi ser. Bilden lägger till ett led i det här förhållandet eftersom den inte enbart visar vad den andra ser. När bilden fungerar som den skall visar den också hur den andra ser och på det här sättet blir bilden verksam. Den föreliggande avhandlingen kombinerar epistemologi med vetenskapshistoria och visuella kulturstudier, men dess huvudintresse är filosofiskt. Den befattar sig med filosofiska missförstånd angående avbildning som en mimetisk konstform, kunskap som domesticering och varseblivning som mottagning av data. ------------------------------------------------------ Tämän väitöskirjan ensimmäinen osa selvittää kuvakäsitteen genealogiaa. Se havainnollistaa miten kuvan, näkemisen ja minän käsitteet kehittyivät suhteessa toisiinsa. Kertomus ulottuu varhaisesta renessanssista ja perspektivistisestä maalaustaiteesta, positivismin aikakauteen ja valokuvan syntyyn. Tämä kehitys toi mukanaan reduktionismin jossa minän rooli – ihmisen psykologian merkitys, meidän arviointikyky, meidän huomiokyky sekä meidän tahtomme – vaipui unohduksiin. Ajatusmaailmassa tapahtui siirtymä, kuvan merkitys vaihtui käsityksestä jossa se on kolmiulotteisen tilan representaatio kaksiulotteisella pinnalla, käsitykseen jossa kuva on läpinäkyvä ruutu, ikkuna kohti maailmaa. Ajatus kuvasta neutraalin näkökulman kantajana vahvisti skeptistä suhtautumista kommunikaatiota, dialogisuutta ja subjektiivisuutta kohtaan. Tämä skeptisyys ilmentyi myös vahvana epäluottamuksena ihmiskeskeisyyttä ja toiseutta kohtaan. Toisessa osassa tekijä tarjoaa vaihtoehdon tälle skeptiselle ajatusmaailmalle jota tarkastellaan ensimmäisessä osassa. Kuva on myös väline joka auttaa meitä jäsentämään meidän näkökenttäämme. Se auttaa meitä jakamaan meidän käsityksiä toistemme kanssa. Tämä näkemisen jakamisen käytäntö on kuvan keskeinen tehtävä. Voin seistä toisen ihmisen vieressä ja nähdä samat asiat kuin hän, mutta en välttämättä ymmärrä miten hän näkee nämä asiat. Kuva lisää jotain olennaista tähän suhteeseen. Kun kuva toimii niin kun sen kuuluu toimia, se näyttää myös miten toinen näkee, tällä tavalla kuvasta tulee välittäjä. Tämä väitöskirja yhdistää epistemologiaa, tieteen historiaa ja visuaalisen kulttuurin tutkimusta, mutta sen pääasiallinen tavoite on filosofinen. Se käsittelee filosofisia väärinkäsityksiä koskien kuvan eideettisyyttä.
Resumo:
Feature extraction is the part of pattern recognition, where the sensor data is transformed into a more suitable form for the machine to interpret. The purpose of this step is also to reduce the amount of information passed to the next stages of the system, and to preserve the essential information in the view of discriminating the data into different classes. For instance, in the case of image analysis the actual image intensities are vulnerable to various environmental effects, such as lighting changes and the feature extraction can be used as means for detecting features, which are invariant to certain types of illumination changes. Finally, classification tries to make decisions based on the previously transformed data. The main focus of this thesis is on developing new methods for the embedded feature extraction based on local non-parametric image descriptors. Also, feature analysis is carried out for the selected image features. Low-level Local Binary Pattern (LBP) based features are in a main role in the analysis. In the embedded domain, the pattern recognition system must usually meet strict performance constraints, such as high speed, compact size and low power consumption. The characteristics of the final system can be seen as a trade-off between these metrics, which is largely affected by the decisions made during the implementation phase. The implementation alternatives of the LBP based feature extraction are explored in the embedded domain in the context of focal-plane vision processors. In particular, the thesis demonstrates the LBP extraction with MIPA4k massively parallel focal-plane processor IC. Also higher level processing is incorporated to this framework, by means of a framework for implementing a single chip face recognition system. Furthermore, a new method for determining optical flow based on LBPs, designed in particular to the embedded domain is presented. Inspired by some of the principles observed through the feature analysis of the Local Binary Patterns, an extension to the well known non-parametric rank transform is proposed, and its performance is evaluated in face recognition experiments with a standard dataset. Finally, an a priori model where the LBPs are seen as combinations of n-tuples is also presented
Resumo:
Diabetic retinopathy, age-related macular degeneration and glaucoma are the leading causes of blindness worldwide. Automatic methods for diagnosis exist, but their performance is limited by the quality of the data. Spectral retinal images provide a significantly better representation of the colour information than common grayscale or red-green-blue retinal imaging, having the potential to improve the performance of automatic diagnosis methods. This work studies the image processing techniques required for composing spectral retinal images with accurate reflection spectra, including wavelength channel image registration, spectral and spatial calibration, illumination correction, and the estimation of depth information from image disparities. The composition of a spectral retinal image database of patients with diabetic retinopathy is described. The database includes gold standards for a number of pathologies and retinal structures, marked by two expert ophthalmologists. The diagnostic applications of the reflectance spectra are studied using supervised classifiers for lesion detection. In addition, inversion of a model of light transport is used to estimate histological parameters from the reflectance spectra. Experimental results suggest that the methods for composing, calibrating and postprocessing spectral images presented in this work can be used to improve the quality of the spectral data. The experiments on the direct and indirect use of the data show the diagnostic potential of spectral retinal data over standard retinal images. The use of spectral data could improve automatic and semi-automated diagnostics for the screening of retinal diseases, for the quantitative detection of retinal changes for follow-up, clinically relevant end-points for clinical studies and development of new therapeutic modalities.
Resumo:
Summary
Resumo:
Selostus: Tasoskannerin ja digitaalisen kuva-analyysimenetelmän kalibrointi juurten morfologian kvantifioimiseksi
Resumo:
Abstract
Resumo:
Abstract
Resumo:
Image filtering is a highly demanded approach of image enhancement in digital imaging systems design. It is widely used in television and camera design technologies to improve the quality of an output image to avoid various problems such as image blurring problem thatgains importance in design of displays of large sizes and design of digital cameras. This thesis proposes a new image filtering method basedon visual characteristics of human eye such as MTF. In contrast to the traditional filtering methods based on human visual characteristics this thesis takes into account the anisotropy of the human eye vision. The proposed method is based on laboratory measurements of the human eye MTF and takes into account degradation of the image by the latter. This method improves an image in the way it will be degraded by human eye MTF to give perception of the original image quality. This thesis gives a basic understanding of an image filtering approach and the concept of MTF and describes an algorithm to perform an image enhancement based on MTF of human eye. Performed experiments have shown quite good results according to human evaluation. Suggestions to improve the algorithm are also given for the future improvements.
Resumo:
The topic of this thesis is studying how lesions in retina caused by diabetic retinopathy can be detected from color fundus images by using machine vision methods. Methods for equalizing uneven illumination in fundus images, detecting regions of poor image quality due toinadequate illumination, and recognizing abnormal lesions were developed duringthe work. The developed methods exploit mainly the color information and simpleshape features to detect lesions. In addition, a graphical tool for collecting lesion data was developed. The tool was used by an ophthalmologist who marked lesions in the images to help method development and evaluation. The tool is a general purpose one, and thus it is possible to reuse the tool in similar projects.The developed methods were tested with a separate test set of 128 color fundus images. From test results it was calculated how accurately methods classify abnormal funduses as abnormal (sensitivity) and healthy funduses as normal (specificity). The sensitivity values were 92% for hemorrhages, 73% for red small dots (microaneurysms and small hemorrhages), and 77% for exudates (hard and soft exudates). The specificity values were 75% for hemorrhages, 70% for red small dots, and 50% for exudates. Thus, the developed methods detected hemorrhages accurately and microaneurysms and exudates moderately.
Resumo:
Tämän tutkimuksen tavoitteena oli selvittää, vaikuttaako kansainvälisen opiskelijan kulttuuritausta opiskelijan odotetun ja koetun yliopistoimagon muodostumiseen. Jotta kulttuurin vaikutuksia yliopistoimagoon voitiin tutkia, tutkimuksessa tunnistettiin yliopistoimagon muodostumiseen oleellisesti vaikuttavat tekijät. Kulttuurin roolia organisaation imagon muodostumisessa ei ole tutkittu aiemmissa tieteellisissä julkaisuissa. Näin ollen tämän tutkimuksen voidaan katsoa edistäneen nykyistä imagotutkimusta. Tutkimuksen kohdeyliopistona oli Lappeenrannan teknillinen yliopisto (LTY). Tutkimuksen empiirinen osa toteutettiin kvantitatiivisena Internet - pohjaisena kyselytutkimuksena tilastollisen analyysin menetelmin. Otos (N=179) koostui kaikista Lappeenrannan teknillisessä yliopistossa lukuvuonna 2005-2006 opiskelleista kansainvälisistä opiskelijoista. Kyselyyn vastasi 68,7 % opiskelijoista. Johtopäätöksenä voidaan todeta, että kulttuurilla ei ole merkittävää vaikutusta yliopistoimagon muodostumiseen. Tutkimuksessa saatiin selville, että yliopiston Internet-sivujen laatu vaikuttaa positiivisesti odotetun yliopistoimagon muodostumiseen, kun taas koettuun yliopistoimagoon vaikuttavat positiivisesti odotettu yliopistoimago, pedagoginen laatu sekä opetusympäristö. Markkinoinnin näkökulmasta tulokset voidaan vetää yhteen toteamalla, että yliopistojen ei tarvitsisi räätälöidä tutkimuksessa tunnistettuja imagoon vaikuttavia tekijöitä eri kulttuureistatulevia opiskelijoita varten.