894 resultados para Text categorization
Resumo:
Social tagging has become very popular around the Internet as well as in research. The main idea behind tagging is to allow users to provide metadata to the web content from their perspective to facilitate categorization and retrieval. There are many factors that influence users' tag choice. Many studies have been conducted to reveal these factors by analysing tagging data. This paper uses two theories to identify these factors, namely the semiotics theory and activity theory. The former treats tags as signs and the latter treats tagging as an activity. The paper uses both theories to analyse tagging behaviour by explaining all aspects of a tagging system, including tags, tagging system components and the tagging activity. The theoretical analysis produced a framework that was used to identify a number of factors. These factors can be considered as categories that can be consulted to redirect user tagging choice in order to support particular tagging behaviour, such as cross-lingual tagging.
Resumo:
Tagging provides support for retrieval and categorization of online content depending on users' tag choice. A number of models of tagging behaviour have been proposed to identify factors that are considered to affect taggers, such as users' tagging history. In this paper, we use Semiotics Analysis and Activity theory, to study the effect the system designer has over tagging behaviour. The framework we use shows the components that comprise the tagging system and how they interact together to direct tagging behaviour. We analysed two collaborative tagging systems: CiteULike and Delicious by studying their components by applying our framework. Using datasets from both systems, we found that 35% of CiteULike users did not provide tags compared to only 0.1% of Delicious users. This was directly linked to the type of tools used by the system designer to support tagging.
Resumo:
Background: Research in aphasia has focused on acquired dyslexias at the single word level, with a paucity of assessment techniques and rehabilitation approaches for individuals with difficulty at the text level. A rich literature from research with paediatric populations and healthy non-brain damaged, skilled adult readers allows the component processes that are important for text reading to be defined and more appropriate assessments to be devised. Aims: To assess the component processes of text reading in a small group of individuals with aphasia who report difficulties reading at the text level. Do assessments of component processes in reading comprehension reveal distinct profiles of text comprehension? To what extent are text comprehension difficulties caused by underlying linguistic and/or cognitive deficits? Methods & Procedures: Four individuals with mild aphasia who reported difficulties reading at the text level took part in a case-series study. Published assessments were used to confirm the presence of text comprehension impairment. Participants completed a range of assessments to provide a profile of their linguistic and cognitive skills, focusing on processes known to be important for text comprehension. We identified the following areas for assessment: reading speed, language skills (single word and sentence), inferencing, working memory and metacognitive skills (monitoring and strategy use). Outcomes & Results: Performance was compared against age-matched adult control data. One participant presented with a trend for impaired abilities in inferencing, with all other assessed skills being within normal limits. The other three had identified linguistic and working memory difficulties. One presented with a residual deficit in accessing single word meaning that affected text comprehension. The other two showed no clear link between sentence processing difficulties and text comprehension impairments. Across these three, data suggested a link between verbal working memory capacity and specific inferencing skills. Conclusions: Successful text reading relies on a number of component processes. In this paper we have made a start in defining those component processes and devising tasks suitable to assess them. From our results, assessment of verbal working memory and inferencing appears to be critical for understanding text comprehension impairments in aphasia. It is possible that rehabilitation input can capitalize on key meta-cognitive skills (monitoring, strategy use) to support functional reading in the face of existing linguistic, text comprehension and memory impairments.
Resumo:
This study explores how the typographic layout of information influences readers' impressions of magazine contents pages. Thirteen descriptors were used in a paired comparison procedure that assessed whether participants' rhetorical impressions of a set of six controlled documents change in relation to variations in layout. The combinations of layout attributes tested were derived from the structural attributes associated with three patterns of typographic differentiation (high, moderate, and low) described in a previous study (see Moys, 2014). The content and the range of stylistic attributes applied to the test material were controlled in order to focus on layout attributes. Triangulation of the quantitative and qualitative data indicates that, even within the experimental confines of limited stylistic differentiation, the layout attributes associated with patterns of high, moderate, and low typographic differentiation do influence readers' rhetorical judgments. In addition, the findings emphasize the importance of considering inter-relationships between clusters of typographic attributes rather than testing isolated variables.
Resumo:
This paper focuses on the language shift phenomenon in Singapore as a consequence of the top-town policies. By looking at bilingual family language policies it examines the characteristics of Singapore’s multilingual nature and cultural diversity. Specifically, it looks at what languages are practiced and how family language policies are enacted in Singaporean English-Chinese bilingual families, and to what extend macro language policies – i.e. national and educational language policies influence and interact with family language policies. Involving 545 families and including parents and grandparents as participants, the study traces the trajectory of the policy history. Data sources include 2 parts: 1) a prescribed linguistic practices survey; and 2) participant observation of actual negotiation of FLP in face-to-face social interaction in bilingual English-Chinese families. The data provides valuable information on how family language policy is enacted and language practices are negotiated, and what linguistic practices have been changed and abandoned against the background of the Speaking Mandarin Campaign and the current bilingual policy implemented in the 1970s. Importantly, the detailed face-to-face interactions and linguistics practices are able to enhance our understanding of the subtleties and processes of language (dis)continuity in relation to policy interventions. The study also discusses the reality of language management measures in contrast to the government’s ‘separate bilingualism’ (Creese & Blackledge, 2011) expectations with regard to ‘striking a balance’ between Asian and Western culture (Curdt-Christiansen & Silver 2013; Shepherd, 2005) and between English and mother tongue languages (Curdt-Christiansen, 2014). Demonstrating how parents and children negotiate their family language policy through translanguaging or heteroglossia practices (Canagarajah, 2013; Garcia & Li Wei, 2014), this paper argues that ‘striking a balance’ as a political ideology places emphasis on discrete and separate notions of cultural and linguistic categorization and thus downplays the significant influences from historical, political and sociolinguistic contexts in which people find themselves. This simplistic view of culture and linguistic code will inevitably constrain individuals’ language expression as it regards code switching and translanguaging as delimited and incompetent language behaviour.
Resumo:
This paper explores the relationship between discourse and action in practices involved in making and consuming texts. Texts are produced through the process of ‘entextualization’ in which strips of action and discourse are extracted from their original contexts and recontextualized into other situations. Different technologies for turning actions into texts affect the kinds of social actions and social identities that are made possible both at moments of entextualization and at future moments of recontextualization. In particular, I focus on how digital technologies affect the practices and participation structures around entextualization. Digital photography and video have had a profound effect on social practices and relationships around the making of texts. Specifically, they have made processes of entextualization more immediate, more contingent and more communal. Implications of these features of digital text making are discussed in light of previous work on literacy and orality.
Resumo:
Televideo cybersex provides a unique example of the ways meanings are instantiated, identities constructed, and relationships negotiated across different semiotic modes. This article explores the role of verbal messages in these multimodal exchanges, examining the specific interactional functions they perform. Text, it is argued, plays a rather unique role in this particular kind of interaction. Unlike ordinary face-to-face conversation, in which the body (posture, gestures, gaze) usually plays more of an ancillary role, in televideo cybersex, the bodily performance is primary, with verbal messages functioning to contextualize physical actions. Text is used to help increase the sense of ‘presence’ participants feel, to regulate the rhythm of the unfolding interaction, to help manage the orderly exchange of information, and to create narrative frames within which bodily displays can be interpreted and made coherent.
Resumo:
The aim of this research is to exhibit how literary playtexts can evoke multisensory trends prevalent in 21st century theatre. In order to do so, it explores a range of practical forms and theoretical contexts for creating participatory, site-specific and immersive theatre. With reference to literary theory, specifically to semiotics, reader-response theory, postmodernism and deconstruction, it attempts to revise dramatic theory established by Aristotle’s Poetics. Considering Gertrude Stein’s essay, Plays (1935), and relevant trends in theatre and performance, shaped by space, technology and the everchanging role of the audience member, a postdramatic poetics emerges from which to analyze the plays of Mac Wellman and Suzan-Lori Parks. Distinguishing the two textual lives of a play as the performance playtext and the literary playtext, it examines the conventions of the printed literary playtext, with reference to models of practice that radicalize the play form, including works by Mabou Mines, The Living Theatre and Fiona Templeton. The arguments of this practice-led Ph.D. developed out of direct engagement with the practice project, which explores the multisensory potential of written language when combined with hypermedia. The written thesis traces the development process of a new play, Rumi High, which is presented digitally as a ‘hyper(play)text,’ accessible through the Internet at www.RumiHigh.org. Here, ‘playwrighting’ practice is expanded spatially, collaboratively and textually. Plays are built, designed and crafted with many layers of meaning that explore both linguistic and graphic modes of poetic expression. The hyper(play)text of Rumi High establishes playwrighting practice as curatorial, where performance and literary playtexts are in a reciprocal relationship. This thesis argues that digital writing and reading spaces enable new approaches to expressing the many languages of performance, while expanding the collaborative network that produces the work. It questions how participatory forms of immersive and site-specific theatre can be presented as interactive literary playtexts, which enable the reader to have a multisensory experience. Through a reflection on process and an evaluation of the practice project, this thesis problematizes notions of authorship and text.
Resumo:
Traditional content-based image retrieval (CBIR) systems use low-level features such as colors, shapes, and textures of images. Although, users make queries based on semantics, which are not easily related to such low-level characteristics. Recent works on CBIR confirm that researchers have been trying to map visual low-level characteristics and high-level semantics. The relation between low-level characteristics and image textual information has motivated this article which proposes a model for automatic classification and categorization of words associated to images. This proposal considers a self-organizing neural network architecture, which classifies textual information without previous learning. Experimental results compare the performance results of the text-based approach to an image retrieval system based on low-level features. (c) 2008 Wiley Periodicals, Inc.
Resumo:
Automatic summarization of texts is now crucial for several information retrieval tasks owing to the huge amount of information available in digital media, which has increased the demand for simple, language-independent extractive summarization strategies. In this paper, we employ concepts and metrics of complex networks to select sentences for an extractive summary. The graph or network representing one piece of text consists of nodes corresponding to sentences, while edges connect sentences that share common meaningful nouns. Because various metrics could be used, we developed a set of 14 summarizers, generically referred to as CN-Summ, employing network concepts such as node degree, length of shortest paths, d-rings and k-cores. An additional summarizer was created which selects the highest ranked sentences in the 14 systems, as in a voting system. When applied to a corpus of Brazilian Portuguese texts, some CN-Summ versions performed better than summarizers that do not employ deep linguistic knowledge, with results comparable to state-of-the-art summarizers based on expensive linguistic resources. The use of complex networks to represent texts appears therefore as suitable for automatic summarization, consistent with the belief that the metrics of such networks may capture important text features. (c) 2008 Elsevier Inc. All rights reserved.
Resumo:
Is it really possible to create a (literary) text? It is actually impossible as to create an authentic text we need an authentic context which is impossible to create. So all the literary txets we have are not authentic and are not created authentically.