12 resultados para ICARUS search and rescue case
em Digital Commons at Florida International University
Resumo:
Most advertising research has focussed at examining effects of advertising on attitudinal responses or brand preference and choice. However, in a natural environment, the time period between advertising exposure and purchase decision is filled with prepurchase search. Prepurchase external search refers to information search from sources other than memory, prior to making a purchase decision. Usually consumers access only a small subset of available information and base their choice decisions on it. Prepurchase search therefore acts as a filter and, the final choice depends critically on the small subset of potential inputs the consumer notes in the environment and integrates into the decision. Previous research has identified a variety of factors that affect consumers' prepurchase search behavior. However, there is little understanding of how specific advertisements designed by marketers impact consumers' prepurchase search. A marketer would like consumers to search information that reflects favorably on his/her brand. Hence, s/he would attempt to influence the brands and attributes on which consumers seek information prior to making a choice. The dissertation investigates the process by which a particular marketer's advertising influences consumers' search on available brands, i.e., the marketer's brand and other competing brands. The dissertation considers a situation where exposure to advertising occurs prior to seeking information from any other source. Hence, the impact of advertising on subsequent search behavior is the topic of interest. The dissertation develops a conceptual model of advertising effects on brand search and conducts two experiments to test the tenets of this model. Specifically, the dissertation demonstrates that attitudinal responses generated by advertising mediate advertising effects on search attitudes and behaviors. The dissertation goes on to examine how attitudinal responses generated by advertising and subsequent effects on search alter brand preference and choice. ^
Resumo:
The United States has been increasingly concerned with the transnational threat posed by infectious diseases. Effective policy implementation to contain the spread of these diseases requires active engagement and support of the American public. To influence American public opinion and enlist support for related domestic and foreign policies, both domestic agencies and international organizations have framed infectious diseases as security threats, human rights disasters, economic risks, and as medical dangers. This study investigates whether American attitudes and opinions about infectious diseases are influenced by how the issue is framed. It also asks which issue frame has been most influential in shaping public opinion about global infectious diseases when people are exposed to multiple frames. The impact of media frames on public perception of infectious diseases is examined through content analysis of newspaper reports. Stories on SARS, avian flu, and HIV/AIDS were sampled from coverage in The New York Times and The Washington Post between 1999 and 2007. Surveys of public opinion on infectious diseases in the same time period were also drawn from databases like Health Poll Search and iPoll. Statistical analysis tests the relationship between media framing of diseases and changes in public opinion. Results indicate that no one frame was persuasive across all diseases. The economic frame had a significant effect on public opinion about SARS, as did the biomedical frame in the case of avian flu. Both the security and human rights frames affected opinion and increased public support for policies intended to prevent or treat HIV/AIDS. The findings also address the debate on the role and importance of domestic public opinion as a factor in domestic and foreign policy decisions of governments in an increasingly interconnected world. The public is able to make reasonable evaluations of the frames and the domestic and foreign policy issues emphasized in the frames.
Resumo:
The search-experience-credence framework from economics of information, the human-environment relations models from environmental psychology, and the consumer evaluation process from services marketing provide a conceptual basis for testing the model of "Pre-purchase Information Utilization in Service Physical Environments." The model addresses the effects of informational signs, as a dimension of the service physical environment, on consumers' perceptions (perceived veracity and perceived performance risk), emotions (pleasure) and behavior (willingness to buy). The informational signs provide attribute quality information (search and experience) through non-personal sources of information (simulated word-of-mouth and non-personal advocate sources).^ This dissertation examines: (1) the hypothesized relationships addressed in the model of "Pre-purchase Information Utilization in Service Physical Environments" among informational signs, perceived veracity, perceived performance risk, pleasure, and willingness to buy, and (2) the effects of attribute quality information and sources of information on consumers' perceived veracity and perceived performance risk.^ This research is the first in-depth study about the role and effects of information in service physical environments. Using a 2 x 2 between subjects experimental research procedure, undergraduate students were exposed to the informational signs in a simulated service physical environment. The service physical environments were simulated through color photographic slides.^ The results of the study suggest that: (1) the relationship between informational signs and willingness to buy is mediated by perceived veracity, perceived performance risk and pleasure, (2) experience attribute information shows higher perceived veracity and lower perceived performance risk when compared to search attribute information, and (3) information provided through simulated word-of-mouth shows higher perceived veracity and lower perceived performance risk when compared to information provided through non-personal advocate sources. ^
Resumo:
Graph-structured databases are widely prevalent, and the problem of effective search and retrieval from such graphs has been receiving much attention recently. For example, the Web can be naturally viewed as a graph. Likewise, a relational database can be viewed as a graph where tuples are modeled as vertices connected via foreign-key relationships. Keyword search querying has emerged as one of the most effective paradigms for information discovery, especially over HTML documents in the World Wide Web. One of the key advantages of keyword search querying is its simplicity—users do not have to learn a complex query language, and can issue queries without any prior knowledge about the structure of the underlying data. The purpose of this dissertation was to develop techniques for user-friendly, high quality and efficient searching of graph structured databases. Several ranked search methods on data graphs have been studied in the recent years. Given a top-k keyword search query on a graph and some ranking criteria, a keyword proximity search finds the top-k answers where each answer is a substructure of the graph containing all query keywords, which illustrates the relationship between the keyword present in the graph. We applied keyword proximity search on the web and the page graph of web documents to find top-k answers that satisfy user’s information need and increase user satisfaction. Another effective ranking mechanism applied on data graphs is the authority flow based ranking mechanism. Given a top- k keyword search query on a graph, an authority-flow based search finds the top-k answers where each answer is a node in the graph ranked according to its relevance and importance to the query. We developed techniques that improved the authority flow based search on data graphs by creating a framework to explain and reformulate them taking in to consideration user preferences and feedback. We also applied the proposed graph search techniques for Information Discovery over biological databases. Our algorithms were experimentally evaluated for performance and quality. The quality of our method was compared to current approaches by using user surveys.
Resumo:
Parallel processing is prevalent in many manufacturing and service systems. Many manufactured products are built and assembled from several components fabricated in parallel lines. An example of this manufacturing system configuration is observed at a manufacturing facility equipped to assemble and test web servers. Characteristics of a typical web server assembly line are: multiple products, job circulation, and paralleling processing. The primary objective of this research was to develop analytical approximations to predict performance measures of manufacturing systems with job failures and parallel processing. The analytical formulations extend previous queueing models used in assembly manufacturing systems in that they can handle serial and different configurations of paralleling processing with multiple product classes, and job circulation due to random part failures. In addition, appropriate correction terms via regression analysis were added to the approximations in order to minimize the gap in the error between the analytical approximation and the simulation models. Markovian and general type manufacturing systems, with multiple product classes, job circulation due to failures, and fork and join systems to model parallel processing were studied. In the Markovian and general case, the approximations without correction terms performed quite well for one and two product problem instances. However, it was observed that the flow time error increased as the number of products and net traffic intensity increased. Therefore, correction terms for single and fork-join stations were developed via regression analysis to deal with more than two products. The numerical comparisons showed that the approximations perform remarkably well when the corrections factors were used in the approximations. In general, the average flow time error was reduced from 38.19% to 5.59% in the Markovian case, and from 26.39% to 7.23% in the general case. All the equations stated in the analytical formulations were implemented as a set of Matlab scripts. By using this set, operations managers of web server assembly lines, manufacturing or other service systems with similar characteristics can estimate different system performance measures, and make judicious decisions - especially setting delivery due dates, capacity planning, and bottleneck mitigation, among others.
Resumo:
In a post-Cold War, post-9/11 world, the advent of US global supremacy resulted in the installation, perpetuation, and dissemination of an Absolutist Security Agenda (hereinafter, ASA). The US ASA explicitly and aggressively articulates and equates US national security interests with the security of all states in the international system, and replaced the bipolar, Cold War framework that defined international affairs from 1945-1992. Since the collapse of the USSR and the 11 September 2001 terrorist attacks, the US has unilaterally defined, implemented, and managed systemic security policy. The US ASA is indicative of a systemic category of knowledge (security) anchored in variegated conceptual and material components, such as morality, philosophy, and political rubrics. The US ASA is based on a logic that involves the following security components: (1) hyper militarization, (2) intimidation,(3) coercion, (4) criminalization, (5) panoptic surveillance, (6) plenary security measures, and (7) unabashed US interference in the domestic affairs of select states. Such interference has produced destabilizing tensions and conflicts that have, in turn, produced resistance, revolutions, proliferation, cults of personality, and militarization. This is the case because the US ASA rests on the notion that the international system of states is an extension, instrument of US power, rather than a system and/or society of states comprised of functionally sovereign entities. To analyze the US ASA, this study utilizes: (1) official government statements, legal doctrines, treaties, and policies pertaining to US foreign policy; (2) militarization rationales, budgets, and expenditures; and (3) case studies of rogue states. The data used in this study are drawn from information that is publicly available (academic journals, think-tank publications, government publications, and information provided by international organizations). The data supports the contention that global security is effectuated via a discrete set of hegemonic/imperialistic US values and interests, finding empirical expression in legal acts (USA Patriot ACT 2001) and the concept of rogue states. Rogue states, therefore, provide test cases to clarify the breadth, depth, and consequentialness of the US ASA in world affairs vis-à-vis the relationship between US security and global security.
Resumo:
This thesis attempted to explain society's worldview of Santeria and its practice of animal sacrifice, and the breakdown between the federal and local government after a 1993 Supreme Court ruling affirming their right to engage in this sacred ritual. Santeria practitioners are harassed and prosecuted for exercising their right to practice animal sacrifice. The research was intended to present the cosmology of the Lukumi tradition, the intellectual framework explored, a review of Freedom of Religion and the case of Lukumi v. Hialeah, and finally the media's role in shaping the worldview of Santeria that have perpetuated this breakdown. The thesis consisted of 87 research items, a community survey, interviews, a Santeria divination, and review of case law, books,newspaper and online journals. These findings demonstrated that freedom of religion is not so free in the U.S., and exists only to the extent the media and municipal laws choose to allow.
Low-fat food consumption by people with diabetes decreases fat saturated fat, and cholesterol intake
Resumo:
This study investigated the effect of providing free-access to several fat-modified foods on dietary energy and fat intake in free-living individuals with and without diabetes mellitus. Five low/no-fat products or their regular-fat versions were provided to volunteers to take home and use for 3 days. Energy and nutrient intakes of all foods consumed were determined through a weighed food diary and by weighing the food provided before and after consumption. Fifteen individuals with diabetes and 15 case-matched controls without diabetes participated in the study. Individuals with diabetes and controls responded similarly to the fat-modified foods. In both groups there was a significant reduction in the percent of kcals and grams of fat consumed during the low-fat condition compared to the regular-fat condition (p
Resumo:
Parallel processing is prevalent in many manufacturing and service systems. Many manufactured products are built and assembled from several components fabricated in parallel lines. An example of this manufacturing system configuration is observed at a manufacturing facility equipped to assemble and test web servers. Characteristics of a typical web server assembly line are: multiple products, job circulation, and paralleling processing. The primary objective of this research was to develop analytical approximations to predict performance measures of manufacturing systems with job failures and parallel processing. The analytical formulations extend previous queueing models used in assembly manufacturing systems in that they can handle serial and different configurations of paralleling processing with multiple product classes, and job circulation due to random part failures. In addition, appropriate correction terms via regression analysis were added to the approximations in order to minimize the gap in the error between the analytical approximation and the simulation models. Markovian and general type manufacturing systems, with multiple product classes, job circulation due to failures, and fork and join systems to model parallel processing were studied. In the Markovian and general case, the approximations without correction terms performed quite well for one and two product problem instances. However, it was observed that the flow time error increased as the number of products and net traffic intensity increased. Therefore, correction terms for single and fork-join stations were developed via regression analysis to deal with more than two products. The numerical comparisons showed that the approximations perform remarkably well when the corrections factors were used in the approximations. In general, the average flow time error was reduced from 38.19% to 5.59% in the Markovian case, and from 26.39% to 7.23% in the general case. All the equations stated in the analytical formulations were implemented as a set of Matlab scripts. By using this set, operations managers of web server assembly lines, manufacturing or other service systems with similar characteristics can estimate different system performance measures, and make judicious decisions - especially setting delivery due dates, capacity planning, and bottleneck mitigation, among others.
Resumo:
The objective of this thesis was to investigate the effects of the built environment on the outcome of young patients. This investigation included recent innovations in children's hospitals that integrated both medical and architectural case studies as part of their design issues. In addition, the intervention responded to man-made conditions and natural elements of the site. The thesis project, a Children's Rehabilitation Hospital, is located at 1500 N.W. River Drive in Miami, Florida. The thesis intervention emerged from a site analysis that focused on the shifting of the urban grid, the variation in scale of the immediate context and the visual-physical connection to the river's edge. Furthermore, it addressed the issues of overnight accommodation for patient's families, as well as sound control through the use of specific materials in space enclosures and open courtyards. The key to the success of this intervention lies in the special attention given to the integration between nature and the built environment. Issues such as the incorporation of nature within a building through the use of vistas and the exploitation of natural light through windows and skylights, were pivotal in the creation of a pleasant environment for visitors, employees and young patients.
Resumo:
In a post-Cold War, post-9/11 world, the advent of US global supremacy resulted in the installation, perpetuation, and dissemination of an Absolutist Security Agenda (hereinafter, ASA). The US ASA explicitly and aggressively articulates and equates US national security interests with the security of all states in the international system, and replaced the bipolar, Cold War framework that defined international affairs from 1945-1992. Since the collapse of the USSR and the 11 September 2001 terrorist attacks, the US has unilaterally defined, implemented, and managed systemic security policy. The US ASA is indicative of a systemic category of knowledge (security) anchored in variegated conceptual and material components, such as morality, philosophy, and political rubrics. The US ASA is based on a logic that involves the following security components: 1., hyper militarization, 2., intimidation, 3., coercion, 4., criminalization, 5., panoptic surveillance, 6., plenary security measures, and 7., unabashed US interference in the domestic affairs of select states. Such interference has produced destabilizing tensions and conflicts that have, in turn, produced resistance, revolutions, proliferation, cults of personality, and militarization. This is the case because the US ASA rests on the notion that the international system of states is an extension, instrument of US power, rather than a system and/or society of states comprised of functionally sovereign entities. To analyze the US ASA, this study utilizes: 1., official government statements, legal doctrines, treaties, and policies pertaining to US foreign policy; 2., militarization rationales, budgets, and expenditures; and 3., case studies of rogue states. The data used in this study are drawn from information that is publicly available (academic journals, think-tank publications, government publications, and information provided by international organizations). The data supports the contention that global security is effectuated via a discrete set of hegemonic/imperialistic US values and interests, finding empirical expression in legal acts (USA Patriot ACT 2001) and the concept of rogue states. Rogue states, therefore, provide test cases to clarify the breadth, depth, and consequentialness of the US ASA in world affairs vis-a-vis the relationship between US security and global security.
Resumo:
Thanks to the advanced technologies and social networks that allow the data to be widely shared among the Internet, there is an explosion of pervasive multimedia data, generating high demands of multimedia services and applications in various areas for people to easily access and manage multimedia data. Towards such demands, multimedia big data analysis has become an emerging hot topic in both industry and academia, which ranges from basic infrastructure, management, search, and mining to security, privacy, and applications. Within the scope of this dissertation, a multimedia big data analysis framework is proposed for semantic information management and retrieval with a focus on rare event detection in videos. The proposed framework is able to explore hidden semantic feature groups in multimedia data and incorporate temporal semantics, especially for video event detection. First, a hierarchical semantic data representation is presented to alleviate the semantic gap issue, and the Hidden Coherent Feature Group (HCFG) analysis method is proposed to capture the correlation between features and separate the original feature set into semantic groups, seamlessly integrating multimedia data in multiple modalities. Next, an Importance Factor based Temporal Multiple Correspondence Analysis (i.e., IF-TMCA) approach is presented for effective event detection. Specifically, the HCFG algorithm is integrated with the Hierarchical Information Gain Analysis (HIGA) method to generate the Importance Factor (IF) for producing the initial detection results. Then, the TMCA algorithm is proposed to efficiently incorporate temporal semantics for re-ranking and improving the final performance. At last, a sampling-based ensemble learning mechanism is applied to further accommodate the imbalanced datasets. In addition to the multimedia semantic representation and class imbalance problems, lack of organization is another critical issue for multimedia big data analysis. In this framework, an affinity propagation-based summarization method is also proposed to transform the unorganized data into a better structure with clean and well-organized information. The whole framework has been thoroughly evaluated across multiple domains, such as soccer goal event detection and disaster information management.