880 resultados para Homography constraint
Resumo:
In this paper, Bayesian decision procedures are developed for dose-escalation studies based on bivariate observations of undesirable events and signs of therapeutic benefit. The methods generalize earlier approaches taking into account only the undesirable outcomes. Logistic regression models are used to model the two responses, which are both assumed to take a binary form. A prior distribution for the unknown model parameters is suggested and an optional safety constraint can be included. Gain functions to be maximized are formulated in terms of accurate estimation of the limits of a therapeutic window or optimal treatment of the next cohort of subjects, although the approach could be applied to achieve any of a wide variety of objectives. The designs introduced are illustrated through simulation and retrospective implementation to a completed dose-escalation study. Copyright © 2006 John Wiley & Sons, Ltd.
Resumo:
It has recently been proposed that life-history evolution is subject to a fundamental size-dependent constraint. This constraint limits the rate at which biomass can be produced so that production per unit of body mass is inevitably slower in larger organisms than in smaller ones. Here we derive predictions for how changes in body size and production rates evolve in different lifestyles subject to this constraint. Predictions are tested by using data on the mass of neonate tissue produced per adult per year in 637 placental mammal species and are generally supported. Compared with terrestrial insectivores with generalized primitive traits, mammals that have evolved more specialized lifestyles have divergent massspecific production rates: (i) increased in groups that specialize on abundant and reliable foods: grazing and browsing herbivores (artiodactyls, lagomorphs, perissoclactyls, and folivorous rodents) and flesh-eating marine mammals (pinnipeds, cetaceans); and (ii) decreased in groups that have lifestyles with reduced death rates: bats, primates, arboreal, fossorial, and desert rodents, bears, elephants, and rhinos. Convergent evolution of groups with similar lifestyles is common, so patterns of productivity across mammalian taxa reflect both ecology and phylogeny. The overall result is that groups with different lifestyles have parallel but offset relationships between production rate and body size. These results shed light on the evolution of the fast-slow life-history continuum, suggesting that variation occurs along two axes corresponding to body size and lifestyle.
Resumo:
In this paper, Bayesian decision procedures are developed for dose-escalation studies based on binary measures of undesirable events and continuous measures of therapeutic benefit. The methods generalize earlier approaches where undesirable events and therapeutic benefit are both binary. A logistic regression model is used to model the binary responses, while a linear regression model is used to model the continuous responses. Prior distributions for the unknown model parameters are suggested. A gain function is discussed and an optional safety constraint is included. Copyright (C) 2006 John Wiley & Sons, Ltd.
Resumo:
The compound bis[1,1'-N,N'-(2-picolyl) aminomethyl] ferrocene, L-1, was synthesized. The protonation constants of this ligand and the stability constants of its complexes with Ni2+, Cu2+, Zn2+, Cd2+ and Pb2+ were determined in aqueous solution by potentiometric methods at 25degreesC and at ionic strength 0.10 mol dm(-3) in KNO3. The compound L-1 forms only 1:1 (M:L) complexes with Pb2+ and Cd2+ while with Ni2+ and Cu2+ species of 2:1 ratio were also found. The complexing behaviour of L-1 is regulated by the constraint imposed by the ferrocene in its backbone, leading to lower values of stability constants for complexes of the divalent first row transition metals when compared with related ligands. However, the differences in stability are smaller for the larger metal ions. The structure of the copper complex with L-1 was determined by single-crystal X-ray diffraction and shows that a species of 2:2 ratio is formed. The two copper centres display distorted octahedral geometries and are linked through the two L1 bridges at a long distance of 8.781(10) Angstrom. The electrochemical behaviour of L-1 was studied in the presence of Ni2+, Cu2+, Zn2+, Cd2+ and Pb2+, showing that upon complexation the ferrocene-ferrocenium half-wave potential shifts anodically in relation to that of the free ligand. The maximum electrochemical shift (DeltaE(1/2)) of 268 mV was found in the presence of Pb2+, followed by Cu2+ (218 mV), Ni2+ (152 mV), Zn2+ (111 mV) and Cd2+ (110 mV). Moreover, L-1 is able to electrochemically and selectively sense Cu2+ in the presence of a large excess of the other transition metal cations studied.
Resumo:
The solid-phase synthesis of a cyclic peptide containing the 21-residue epitope found in the A-B loop of the Cepsilon3 domain of human immunoglobulin E has been carried out. The key macrocyclization step to form the 65-membered ring is achieved in similar to15% yield via an "on-resin" Sonogashira coupling reaction which concomitantly installs a diphenylacetylene amino acid conformational constraint within the loop.
Resumo:
Design management research usually deals with the processes within the professional design team and yet, in the UK, the volume of the total project information produced by the specialist trade contractors equals or exceeds that produced by the design team. There is a need to understand the scale of this production task and to plan and manage it accordingly. The model of the process on which the plan is to be based, while generic, must be sufficiently robust to cover the majority of instances. An approach using design elements, in sufficient depth to possibly develop tools for a predictive model of the process, is described. The starting point is that each construction element and its components have a generic sequence of design activities. Specific requirements tailor the element's application to the building. Then there are the constraints produced due to the interaction with other elements. Therefore, the selection of a component within the element may impose a set of constraints that will affect the choice of other design elements. Thus, a design decision can be seen as an interrelated element-constraint-element (ECE) sub-net. To illustrate this approach, an example of the process within precast concrete cladding has been used.
Resumo:
Design management research usually deals with the processes within the professional design team and yet, in the UK, the volume of the total project information produced by the specialist trade contractors equals or exceeds that produced by the design team. There is a need to understand the scale of this production task and to plan and manage it accordingly. The model of the process on which the plan is to be based, while generic, must be sufficiently robust to cover the majority of instances. An approach using design elements, in sufficient depth to possibly develop tools for a predictive model of the process, is described. The starting point is that each construction element and its components have a generic sequence of design activities. Specific requirements tailor the element's application to the building. Then there are the constraints produced due to the interaction with other elements. Therefore, the selection of a component within the element may impose a set of constraints that will affect the choice of other design elements. Thus, a design decision can be seen as an interrelated element-constraint-element (ECE) sub-net. To illustrate this approach, an example of the process within precast concrete cladding has been used.
Resumo:
Four groups of second language (L2) learners of English from different language backgrounds (Chinese, Japanese, German, and Greek) and a group of native speaker controls participated in an online reading time experiment with sentences involving long-distance whdependencies. Although the native speakers showed evidence of making use of intermediate syntactic gaps during processing, the L2 learners appeared to associate the fronted wh-phrase directly with its lexical subcategorizer, regardless of whether the subjacency constraint was operative in their native language. This finding is argued to support the hypothesis that nonnative comprehenders underuse syntactic information in L2 processing.
Resumo:
Under the framework of the European Union Funded SAFEE project(1), this paper gives an overview of a novel monitoring and scene analysis system developed for use onboard aircraft in spatially constrained environments. The techniques discussed herein aim to warn on-board crew about pre-determined indicators of threat intent (such as running or shouting in the cabin), as elicited from industry and security experts. The subject matter experts believe that activities such as these are strong indicators of the beginnings of undesirable chains of events or scenarios, which should not be allowed to develop aboard aircraft. This project aimes to detect these scenarios and provide advice to the crew. These events may involve unruly passengers or be indicative of the precursors to terrorist threats. With a state of the art tracking system using homography intersections of motion images, and probability based Petri nets for scene understanding, the SAFEE behavioural analysis system automatically assesses the output from multiple intelligent sensors, and creates. recommendations that are presented to the crew using an integrated airborn user interface. Evaluation of the system is conducted within a full size aircraft mockup, and experimental results are presented, showing that the SAFEE system is well suited to monitoring people in confined environments, and that meaningful and instructive output regarding human actions can be derived from the sensor network within the cabin.
Resumo:
Typically, algorithms for generating stereo disparity maps have been developed to minimise the energy equation of a single image. This paper proposes a method for implementing cross validation in a belief propagation optimisation. When tested using the Middlebury online stereo evaluation, the cross validation improves upon the results of standard belief propagation. Furthermore, it has been shown that regions of homogeneous colour within the images can be used for enforcing the so-called "Segment Constraint". Developing from this, Segment Support is introduced to boost belief between pixels of the same image region and improve propagation into textureless regions.
Resumo:
Using the classical Parzen window (PW) estimate as the target function, the sparse kernel density estimator is constructed in a forward constrained regression manner. The leave-one-out (LOO) test score is used for kernel selection. The jackknife parameter estimator subject to positivity constraint check is used for the parameter estimation of a single parameter at each forward step. As such the proposed approach is simple to implement and the associated computational cost is very low. An illustrative example is employed to demonstrate that the proposed approach is effective in constructing sparse kernel density estimators with comparable accuracy to that of the classical Parzen window estimate.
Resumo:
Multiple cooperating robot systems may be required to take up a closely coupled configuration in order to perform a task. An example is extended baseline stereo (EBS), requiring that two robots must establish and maintain for a certain period of time a constrained kinematic relationship to each other. In this paper we report on the development of a networked robotics framework for modular, distributed robot systems that supports the creation of such configurations. The framework incorporates a query mechanism to locate modules distributed across the two robot systems. The work presented in this paper introduces special mechanisms to model the kinematic constraint and its instantiation. The EBS configuration is used as a case study and experimental implementation to demonstrate the approach.
Resumo:
This correspondence proposes a new algorithm for the OFDM joint data detection and phase noise (PHN) cancellation for constant modulus modulations. We highlight that it is important to address the overfitting problem since this is a major detrimental factor impairing the joint detection process. In order to attack the overfitting problem we propose an iterative approach based on minimum mean square prediction error (MMSPE) subject to the constraint that the estimated data symbols have constant power. The proposed constrained MMSPE algorithm (C-MMSPE) significantly improves the performance of existing approaches with little extra complexity being imposed. Simulation results are also given to verify the proposed algorithm.
Resumo:
Using the classical Parzen window (PW) estimate as the target function, the sparse kernel density estimator is constructed in a forward-constrained regression (FCR) manner. The proposed algorithm selects significant kernels one at a time, while the leave-one-out (LOO) test score is minimized subject to a simple positivity constraint in each forward stage. The model parameter estimation in each forward stage is simply the solution of jackknife parameter estimator for a single parameter, subject to the same positivity constraint check. For each selected kernels, the associated kernel width is updated via the Gauss-Newton method with the model parameter estimate fixed. The proposed approach is simple to implement and the associated computational cost is very low. Numerical examples are employed to demonstrate the efficacy of the proposed approach.
Resumo:
In this paper, data from spaceborne radar, lidar and infrared radiometers on the “A-Train” of satellites are combined in a variational algorithm to retrieve ice cloud properties. The method allows a seamless retrieval between regions where both radar and lidar are sensitive to the regions where one detects the cloud. We first implement a cloud phase identification method, including identification of supercooled water layers using the lidar signal and temperature to discriminate ice from liquid. We also include rigorous calculation of errors assigned in the variational scheme. We estimate the impact of the microphysical assumptions on the algorithm when radiances are not assimilated by evaluating the impact of the change in the area-diameter and the density-diameter relationships in the retrieval of cloud properties. We show that changes to these assumptions affect the radar-only and lidar-only retrieval more than the radar-lidar retrieval, although the lidar-only extinction retrieval is only weakly affected. We also show that making use of the molecular lidar signal beyond the cloud as a constraint on optical depth, when ice clouds are sufficiently thin to allow the lidar signal to penetrate them entirely, improves the retrieved extinction. When infrared radiances are available, they provide an extra constraint and allow the extinction-to-backscatter ratio to vary linearly with height instead of being constant, which improves the vertical distribution of retrieved cloud properties.