892 resultados para Year 2000 date conversion (Computer systems)


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This edition of the Bulletin provides a report on the Conference on the Effect of the Year 2000 Problem (Y2K) on Transport Systems in the Western Hemisphere, which was held in Cancún, Mexico, in May 1999. Also included is an analysis of sea transport in South America, prepared and presented by the Uruguayan delegation at the Fifth Ordinary Meeting of the Conference of Ministers of Transport, Communications and Public Works of South America, held in Cochabamba, Bolivia, in April 1999.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The paper presents the radiometric parameters determined by the medical physicist during routine radiotherapy planning service in cases of breast cancer . The contours of the breast volume in patients undergoing radiation breast tumors at the Hospital das Clinicas, Faculty of Medicine , UNESP, Botucatu ( HCFMB ) during the year 2012 were analyzed . In order to analyze the influence of physical and radiometric parameters for the determination of the dose distribution of irradiated breast volume , four measurements of isodose curves were prepared in four different heights breast , and compared with the isodose curves plotted computationally . In the routine of planning , the medical physicist must determine the isodose curve that gives the best dose distribution homogeneity in the irradiated volume . The choice of the treatment plan can be done by dedicated computer systems , which require significantly costly investments available services having better financial support . In the Service of Medical Physics , Department of Radiotherapy , HC FMB , we use a two-dimensional software for determination of isodose curves , however , this software is out of date and frequently becomes inoperable due to the lack of maintenance and it is a closed system without feasibility of interference from computer professionals . This fact requires manual preparation of isodose curves , which are subject to uncertainties due to the subjectivity in the clinical interpretation of medical radiation oncologist and medical physicist responsible for planning , plus dispendiar significant calculation time . The choice of the optimal isodose curve depends on the energy of the radiation beam , the geometry and dimensions of the irradiated area . The contours of the breast studied in this work evaluations showed that , for a given energy input , such as the energy of 1.25 MeV of gamma radiation Unit Telecobaltoterapia , the determination of the percentage depth dose ( PDP ) ...

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Animal production, hay production and feeding, winter forage composition changes, and summer pasture yields and nutrient composition of a year-round grazing system for spring-calving and fall-calving cows were compared to those of a conventional, minimal land system. Cows in the year-round and minimal land systems grazed forage from smooth bromegrassorchardgrass-birdsfoot trefoil (SB-O-T) pastures at 1.67 and 3.33 acres, respectively, per cow in the summer. During the summer, SB-O-T pastures in the year-round grazing system also were grazed by stockers at 1.67 stockers per acre, and spring-calving and fall-calving cows grazed smooth bromegrass–red clover (SB-RC) and endophyte-free tall fescue–red clover (TF-RC) at 2.5 acres per cow for approximately 45 days in midsummer. In the year-round grazing system, spring-calving cows grazed corn crop residues at 2.5 acres per cow and stockpiled SB-RC pastures at 2.5 acres per cow; fallcalving cows grazed stockpiled TF-RC pastures at 2.5 acres per cow during winter. In the minimal land system, in winter, cows were maintained in a drylot on first-cutting hay harvested from 62.5–75% of the pasture acres during summer. Hay was fed to maintain a body condition score of 5 on a 9-point scale for springcalving cows in both systems and a body condition score of 3 for fall-calving cows in the year-round system. Over 3 years, mean body weights of fall-calving cows in the year-round system did not differ from the body weights of spring-calving cows in either system, but fall-calving cows had higher (P < .05) body condition scores compared to spring-calving cows in either system. There were no differences among all groups of cows in body condition score changes over the winter grazing season (P > .05). During the summer grazing season, fall-calving cows in the year- round system and springcalving cows in the minimal land system gained more body condition and more weight (P < .05) than springcalving cows in the year-round grazing system. Fall calves in the year-round system had higher birth weights, lower weaning weights, and lower average preweaning daily gains compared to either group of spring calves (P < .05). However, there were no significant differences for birth weights, weaning weights, or average pre-weaning daily gains between spring calves in either system over the 3-year experiment (P > .05). The amount of total growing animal production (calves and stockers) per acre for each system did not differ in any year (P > .05). Over the 3-year experiment, 1.9 ton more hay was fed per cow and 1 ton more hay was fed per cow–calf pair in the minimal land system compared to the year-round grazing system (P < .05).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Queueing theory is an effective tool in the analysis of canputer camrunication systems. Many results in queueing analysis have teen derived in the form of Laplace and z-transform expressions. Accurate inversion of these transforms is very important in the study of computer systems, but the inversion is very often difficult. In this thesis, methods for solving some of these queueing problems, by use of digital signal processing techniques, are presented. The z-transform of the queue length distribution for the Mj GY jl system is derived. Two numerical methods for the inversion of the transfom, together with the standard numerical technique for solving transforms with multiple queue-state dependence, are presented. Bilinear and Poisson transform sequences are presented as useful ways of representing continuous-time functions in numerical computations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Intrusion Detection Systems (IDSs) provide an important layer of security for computer systems and networks, and are becoming more and more necessary as reliance on Internet services increases and systems with sensitive data are more commonly open to Internet access. An IDS’s responsibility is to detect suspicious or unacceptable system and network activity and to alert a systems administrator to this activity. The majority of IDSs use a set of signatures that define what suspicious traffic is, and Snort is one popular and actively developing open-source IDS that uses such a set of signatures known as Snort rules. Our aim is to identify a way in which Snort could be developed further by generalising rules to identify novel attacks. In particular, we attempted to relax and vary the conditions and parameters of current Snort rules, using a similar approach to classic rule learning operators such as generalisation and specialisation. We demonstrate the effectiveness of our approach through experiments with standard datasets and show that we are able to detect previously undetected variants of various attacks. We conclude by discussing the general effectiveness and appropriateness of generalisation in Snort based IDS rule processing. Keywords: anomaly detection, intrusion detection, Snort, Snort rules

Relevância:

100.00% 100.00%

Publicador:

Resumo:

En el área de Aerofotogrametría Digital, el software comercial prevalente para postproceso presenta limitaciones debido a dos factores: (i) las legislaciones de cada país o región requieren diferentes convenciones, y (ii) las necesidades de las empresas son tan cambiantes que no justifican la compra de software de alto rendimiento, que puede quedar sin utilizar debido a un viraje del mercado -- El presente proyecto se ha desarrollado para atender necesidades de procesamiento automático de planos (partición, detección y corrección de errores, etc.), así como módulos de importación – exportación paquete a paquete, trazado de rutas e interacción con GPS -- Este artículo informa de los dos últimos aspectos -- Debido a necesidades de los clientes, los archivos entregados deben llevar un formato comercial (DWG, DXF), pero el procesamiento de los archivos debe ser hecho en paquetes y formatos diversos (DGN) -- Por lo tanto, fue necesario diseñar e implementar un formato acompañante que permitió llevar la información que se pierde al usar filtros comerciales (DGN a DXF/DWG) -- Asimismo se crearon módulos de importación y exportación redundantes, que hicieron efectivos dichos atributos -- En el aspecto de generación de rutas de vuelo, se reportan en este artículo la aplicación de algoritmos tradicionales de barrido (peinado) de áreas 2D, a los cuales se agregaron restricciones geométricas (puntos fijos, offsets, orden de los barridos de acuerdo a coordenadas del sitio de partida, etc.) -- Debido a los altos costos de equipos equivalentes, se decidió desarrollar software para traducción de rutas entre formatos GPS y formatos geográficos locales al país -- Ello permite la eliminación de fuentes de error y además facilita la carga del plan de vuelo, a costos mucho menores a los del hardware / software comercial

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The construction industry is categorised as being an information-intensive industry and described as one of the most important industries in any developed country, facing a period of rapid and unparalleled change (Industry Science Resources 1999) (Love P.E.D., Tucker S.N. et al. 1996). Project communications are becoming increasingly complex, with a growing need and fundamental drive to collaborate electronically at project level and beyond (Olesen K. and Myers M.D. 1999; Thorpe T. and Mead S. 2001; CITE 2003). Yet, the industry is also identified as having a considerable lack of knowledge and awareness about innovative information and communication technology (ICT) and web-based communication processes, systems and solutions which may prove beneficial in the procurement, delivery and life cycle of projects (NSW Government 1998; Kajewski S. and Weippert A. 2000). The Internet has debatably revolutionised the way in which information is stored, exchanged and viewed, opening new avenues for business, which only a decade ago were deemed almost inconceivable (DCITA 1998; IIB 2002). In an attempt to put these ‘new avenues of business’ into perspective, this report provides an overall ‘snapshot’ of current public and private construction industry sector opportunities and practices in the implementation and application of web-based ICT tools, systems and processes (e-Uptake). Research found that even with a reserved uptake, the construction industry and its participating organisations are making concerted efforts (fortunately with positive results) in taking up innovative forms of doing business via the internet, including e-Tendering (making it possible to manage the entire tender letting process electronically and online) (Anumba C.J. and Ruikar K. 2002; ITCBP 2003). Furthermore, Government (often a key client within the construction industry),and with its increased tendency to transact its business electronically, undoubtedly has an effect on how various private industry consultants, contractors, suppliers, etc. do business (Murray M. 2003) – by offering a wide range of (current and anticipated) e-facilities / services, including e-Tendering (Ecommerce 2002). Overall, doing business electronically is found to have a profound impact on the way today’s construction businesses operate - streamlining existing processes, with the growth in innovative tools, such as e-Tender, offering the construction industry new responsibilities and opportunities for all parties involved (ITCBP 2003). It is therefore important that these opportunities should be accessible to as many construction industry businesses as possible (The Construction Confederation 2001). Historically, there is a considerable exchange of information between various parties during a tendering process, where accuracy and efficiency of documentation is critical. Traditionally this process is either paper-based (involving large volumes of supporting tender documentation), or via a number of stand-alone, non-compatible computer systems, usually costly to both the client and contractor. As such, having a standard electronic exchange format that allows all parties involved in an electronic tender process to access one system only via the Internet, saves both time and money, eliminates transcription errors and increases speed of bid analysis (The Construction Confederation 2001). Supporting this research project’s aims and objectives, researchers set to determine today’s construction industry ‘current state-of-play’ in relation to e-Tendering opportunities. The report also provides brief introductions to several Australian and International e-Tender systems identified during this investigation. e-Tendering, in its simplest form, is described as the electronic publishing, communicating, accessing, receiving and submitting of all tender related information and documentation via the internet, thereby replacing the traditional paper-based tender processes, and achieving a more efficient and effective business process for all parties involved (NT Governement 2000; NT Government 2000; NSW Department of Commerce 2003; NSW Government 2003). Although most of the e-Tender websites investigated at the time, maintain their tendering processes and capabilities are ‘electronic’, research shows these ‘eTendering’ systems vary from being reasonably advanced to more ‘basic’ electronic tender notification and archiving services for various industry sectors. Research also indicates an e-Tender system should have a number of basic features and capabilities, including: • All tender documentation to be distributed via a secure web-based tender system – thereby avoiding the need for collating paperwork and couriers. • The client/purchaser should be able to upload a notice and/or invitation to tender onto the system. • Notification is sent out electronically (usually via email) for suppliers to download the information and return their responses electronically (online). • During the tender period, updates and queries are exchanged through the same e-Tender system. • The client/purchaser should only be able to access the tenders after the deadline has passed. • All tender related information is held in a central database, which should be easily searchable and fully audited, with all activities recorded. • It is essential that tender documents are not read or submitted by unauthorised parties. • Users of the e-Tender system are to be properly identified and registered via controlled access. In simple terms, security has to be as good as if not better than a manual tender process. Data is to be encrypted and users authenticated by means such as digital signatures, electronic certificates or smartcards. • All parties must be assured that no 'undetected' alterations can be made to any tender. • The tenderer should be able to amend the bid right up to the deadline – whilst the client/purchaser cannot obtain access until the submission deadline has passed. • The e-Tender system may also include features such as a database of service providers with spreadsheet-based pricing schedules, which can make it easier for a potential tenderer to electronically prepare and analyse a tender. Research indicates the efficiency of an e-Tender process is well supported internationally, with a significant number, yet similar, e-Tender benefits identified during this investigation. Both construction industry and Government participants generally agree that the implementation of an automated e-Tendering process or system enhances the overall quality, timeliness and cost-effectiveness of a tender process, and provides a more streamlined method of receiving, managing, and submitting tender documents than the traditional paper-based process. On the other hand, whilst there are undoubtedly many more barriers challenging the successful implementation and adoption of an e-Tendering system or process, researchers have also identified a range of challenges and perceptions that seem to hinder the uptake of this innovative approach to tendering electronically. A central concern seems to be that of security - when industry organisations have to use the Internet for electronic information transfer. As a result, when it comes to e-Tendering, industry participants insist these innovative tendering systems are developed to ensure the utmost security and integrity. Finally, if Australian organisations continue to explore the competitive ‘dynamics’ of the construction industry, without realising the current and future, trends and benefits of adopting innovative processes, such as e-Tendering, it will limit their globalising opportunities to expand into overseas markets and allow the continuation of international firms successfully entering local markets. As such, researchers believe increased knowledge, awareness and successful implementation of innovative systems and processes raises great expectations regarding their contribution towards ‘stimulating’ the globalisation of electronic procurement activities, and improving overall business and project performances throughout the construction industry sectors and overall marketplace (NSW Government 2002; Harty C. 2003; Murray M. 2003; Pietroforte R. 2003). Achieving the successful integration of an innovative e-Tender solution with an existing / traditional process can be a complex, and if not done correctly, could lead to failure (Bourn J. 2002).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Computer forensics is the process of gathering and analysing evidence from computer systems to aid in the investigation of a crime. Typically, such investigations are undertaken by human forensic examiners using purpose-built software to discover evidence from a computer disk. This process is a manual one, and the time it takes for a forensic examiner to conduct such an investigation is proportional to the storage capacity of the computer's disk drives. The heterogeneity and complexity of various data formats stored on modern computer systems compounds the problems posed by the sheer volume of data. The decision to undertake a computer forensic examination of a computer system is a decision to commit significant quantities of a human examiner's time. Where there is no prior knowledge of the information contained on a computer system, this commitment of time and energy occurs with little idea of the potential benefit to the investigation. The key contribution of this research is the design and development of an automated process to describe a computer system and its activity for the purposes of a computer forensic investigation. The term proposed for this process is computer profiling. A model of a computer system and its activity has been developed over the course of this research. Using this model a computer system, which is the subj ect of investigation, can be automatically described in terms useful to a forensic investigator. The computer profiling process IS resilient to attempts to disguise malicious computer activity. This resilience is achieved by detecting inconsistencies in the information used to infer the apparent activity of the computer. The practicality of the computer profiling process has been demonstrated by a proof-of concept software implementation. The model and the prototype implementation utilising the model were tested with data from real computer systems. The resilience of the process to attempts to disguise malicious activity has also been demonstrated with practical experiments conducted with the same prototype software implementation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Dynamic Data eXchange (DDX) is our third generation platform for building distributed robot controllers. DDX allows a coalition of programs to share data at run-time through an efficient shared memory mechanism managed by a store. Further, stores on multiple machines can be linked by means of a global catalog and data is moved between the stores on an as needed basis by multi-casting. Heterogeneous computer systems are handled. We describe the architecture of DDX and the standard clients we have developed that let us rapidly build complex control systems with minimal coding.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Music making affects relationships with self and others by generating a sense of belonging to a culture or ideology (Bamford, 2006; Barovick, 2001; Dillon & Stewart, 2006; Fiske, 2000; Hallam, 2001). Whilst studies from arts education research present compelling examples of these relationships, others argue that they do not present sufficiently validated evidence of a causal link between music making experiences and cognitive or social change (Winner & Cooper, 2000; Winner & Hetland, 2000a, 2000b, 2001). I have suggested elsewhere that this disconnection between compelling evidence and observations of the effects of music making are in part due to the lack of rigor in research and the incapacity of many methods to capture these experiences in meaningful ways (Dillon, 2006). Part of the answer to these questions about rigor and causality lay in the creative use of new media technologies that capture the results of relationships in music artefacts. Crucially, it is the effective management of these artefacts within computer systems that allows researchers and practitioners to collect, organize, analyse and then theorise such music making experiences.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Historically, cities as urban forms have been critical to human development. In 1950, 30% of the world’s population lived in major cities. By the year 2000 this had increased to 47% with further expected growth to 50% by the end of 2007. Projections suggest that city-based densities will edge towards 60% of the global total by 2030. Such rapidly increasing urbanisation, in both developed and developing economies, challenges options for governance and planning, as well as crisis and disaster management. A common issue to the livability of cities as urban forms through time has been access to clean and reliable water supply. This is an issue that is particularly important in countries with arid ecosystems, such as Australia. This paper examines preliminary aspects, and theoretical basis, of a study into the resilience of the (potable) water supply system in Southeast Queensland (SEQ), an area with one of the most significant urban growth rates in Australia. The first stage will be to assess needs and requirements for gauging resilience characteristics of a generic water supply system, consisting of supply catchment, storage reservoir/s and treatment plant/s. The second stage will extend the analysis to examine the resilience of the SEQ water supply system incorporating specific characteristics of the SEQ water grid made increasingly vulnerable due to climate variability and projected impacts on rainfall characteristics and compounded by increasing demands due to population growth. Longer-term findings will inform decision making based on the application of the concept of resilience to designing and operating stand-alone and networked water supply infrastructure systems as well as its application to water resource systems more generally.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Most current computer systems authorise the user at the start of a session and do not detect whether the current user is still the initial authorised user, a substitute user, or an intruder pretending to be a valid user. Therefore, a system that continuously checks the identity of the user throughout the session is necessary without being intrusive to end-user and/or effectively doing this. Such a system is called a continuous authentication system (CAS). Researchers have applied several approaches for CAS and most of these techniques are based on biometrics. These continuous biometric authentication systems (CBAS) are supplied by user traits and characteristics. One of the main types of biometric is keystroke dynamics which has been widely tried and accepted for providing continuous user authentication. Keystroke dynamics is appealing for many reasons. First, it is less obtrusive, since users will be typing on the computer keyboard anyway. Second, it does not require extra hardware. Finally, keystroke dynamics will be available after the authentication step at the start of the computer session. Currently, there is insufficient research in the CBAS with keystroke dynamics field. To date, most of the existing schemes ignore the continuous authentication scenarios which might affect their practicality in different real world applications. Also, the contemporary CBAS with keystroke dynamics approaches use characters sequences as features that are representative of user typing behavior but their selected features criteria do not guarantee features with strong statistical significance which may cause less accurate statistical user-representation. Furthermore, their selected features do not inherently incorporate user typing behavior. Finally, the existing CBAS that are based on keystroke dynamics are typically dependent on pre-defined user-typing models for continuous authentication. This dependency restricts the systems to authenticate only known users whose typing samples are modelled. This research addresses the previous limitations associated with the existing CBAS schemes by developing a generic model to better identify and understand the characteristics and requirements of each type of CBAS and continuous authentication scenario. Also, the research proposes four statistical-based feature selection techniques that have highest statistical significance and encompasses different user typing behaviors which represent user typing patterns effectively. Finally, the research proposes the user-independent threshold approach that is able to authenticate a user accurately without needing any predefined user typing model a-priori. Also, we enhance the technique to detect the impostor or intruder who may take over during the entire computer session.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Several approaches have been introduced in literature for active noise control (ANC) systems. Since FxLMS algorithm appears to be the best choice as a controller filter, researchers tend to improve performance of ANC systems by enhancing and modifying this algorithm. This paper proposes a new version of FxLMS algorithm. In many ANC applications an online secondary path modelling method using a white noise as a training signal is required to ensure convergence of the system. This paper also proposes a new approach for online secondary path modelling in feedfoward ANC systems. The proposed algorithm stops injection of the white noise at the optimum point and reactivate the injection during the operation, if needed, to maintain performance of the system. Benefiting new version of FxLMS algorithm and not continually injection of white noise makes the system more desirable and improves the noise attenuation performance. Comparative simulation results indicate effectiveness of the proposed approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper discusses computer mediated distance learning on a Master's level course in the UK and student perceptions of this as a quality learning environment.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Enterprise resource planning (ERP) systems are rapidly being combined with “big data” analytics processes and publicly available “open data sets”, which are usually outside the arena of the enterprise, to expand activity through better service to current clients as well as identifying new opportunities. Moreover, these activities are now largely based around relevant software systems hosted in a “cloud computing” environment. However, the over 50- year old phrase related to mistrust in computer systems, namely “garbage in, garbage out” or “GIGO”, is used to describe problems of unqualified and unquestioning dependency on information systems. However, a more relevant GIGO interpretation arose sometime later, namely “garbage in, gospel out” signifying that with large scale information systems based around ERP and open datasets as well as “big data” analytics, particularly in a cloud environment, the ability to verify the authenticity and integrity of the data sets used may be almost impossible. In turn, this may easily result in decision making based upon questionable results which are unverifiable. Illicit “impersonation” of and modifications to legitimate data sets may become a reality while at the same time the ability to audit any derived results of analysis may be an important requirement, particularly in the public sector. The pressing need for enhancement of identity, reliability, authenticity and audit services, including naming and addressing services, in this emerging environment is discussed in this paper. Some current and appropriate technologies currently being offered are also examined. However, severe limitations in addressing the problems identified are found and the paper proposes further necessary research work for the area. (Note: This paper is based on an earlier unpublished paper/presentation “Identity, Addressing, Authenticity and Audit Requirements for Trust in ERP, Analytics and Big/Open Data in a ‘Cloud’ Computing Environment: A Review and Proposal” presented to the Department of Accounting and IT, College of Management, National Chung Chen University, 20 November 2013.)