4 resultados para two-stage
em DRUM (Digital Repository at the University of Maryland)
Resumo:
Nonlinear thermo-mechanical properties of advanced polymers are crucial to accurate prediction of the process induced warpage and residual stress of electronics packages. The Fiber Bragg grating (FBG) sensor based method is advanced and implemented to determine temperature and time dependent nonlinear properties. The FBG sensor is embedded in the center of the cylindrical specimen, which deforms together with the specimen. The strains of the specimen at different loading conditions are monitored by the FBG sensor. Two main sources of the warpage are considered: curing induced warpage and coefficient of thermal expansion (CTE) mismatch induced warpage. The effective chemical shrinkage and the equilibrium modulus are needed for the curing induced warpage prediction. Considering various polymeric materials used in microelectronic packages, unique curing setups and procedures are developed for elastomers (extremely low modulus, medium viscosity, room temperature curing), underfill materials (medium modulus, low viscosity, high temperature curing), and epoxy molding compound (EMC: high modulus, high viscosity, high temperature pressure curing), most notably, (1) zero-constraint mold for elastomers; (2) a two-stage curing procedure for underfill materials and (3) an air-cylinder based novel setup for EMC. For the CTE mismatch induced warpage, the temperature dependent CTE and the comprehensive viscoelastic properties are measured. The cured cylindrical specimen with a FBG sensor embedded in the center is further used for viscoelastic property measurements. A uni-axial compressive loading is applied to the specimen to measure the time dependent Young’s modulus. The test is repeated from room temperature to the reflow temperature to capture the time-temperature dependent Young’s modulus. A separate high pressure system is developed for the bulk modulus measurement. The time temperature dependent bulk modulus is measured at the same temperatures as the Young’s modulus. The master curve of the Young’s modulus and bulk modulus of the EMC is created and a single set of the shift factors is determined from the time temperature superposition. The supplementary experiments are conducted to verify the validity of the assumptions associated with the linear viscoelasticity. The measured time-temperature dependent properties are further verified by a shadow moiré and Twyman/Green test.
Resumo:
The size of online image datasets is constantly increasing. Considering an image dataset with millions of images, image retrieval becomes a seemingly intractable problem for exhaustive similarity search algorithms. Hashing methods, which encodes high-dimensional descriptors into compact binary strings, have become very popular because of their high efficiency in search and storage capacity. In the first part, we propose a multimodal retrieval method based on latent feature models. The procedure consists of a nonparametric Bayesian framework for learning underlying semantically meaningful abstract features in a multimodal dataset, a probabilistic retrieval model that allows cross-modal queries and an extension model for relevance feedback. In the second part, we focus on supervised hashing with kernels. We describe a flexible hashing procedure that treats binary codes and pairwise semantic similarity as latent and observed variables, respectively, in a probabilistic model based on Gaussian processes for binary classification. We present a scalable inference algorithm with the sparse pseudo-input Gaussian process (SPGP) model and distributed computing. In the last part, we define an incremental hashing strategy for dynamic databases where new images are added to the databases frequently. The method is based on a two-stage classification framework using binary and multi-class SVMs. The proposed method also enforces balance in binary codes by an imbalance penalty to obtain higher quality binary codes. We learn hash functions by an efficient algorithm where the NP-hard problem of finding optimal binary codes is solved via cyclic coordinate descent and SVMs are trained in a parallelized incremental manner. For modifications like adding images from an unseen class, we propose an incremental procedure for effective and efficient updates to the previous hash functions. Experiments on three large-scale image datasets demonstrate that the incremental strategy is capable of efficiently updating hash functions to the same retrieval performance as hashing from scratch.
Resumo:
Nonpoint sources (NPS) pollution from agriculture is the leading source of water quality impairment in U.S. rivers and streams, and a major contributor to lakes, wetlands, estuaries and coastal waters (U.S. EPA 2016). Using data from a survey of farmers in Maryland, this dissertation examines the effects of a cost sharing policy designed to encourage adoption of conservation practices that reduce NPS pollution in the Chesapeake Bay watershed. This watershed is the site of the largest Total Maximum Daily Load (TMDL) implemented to date, making it an important setting in the U.S. for water quality policy. I study two main questions related to the reduction of NPS pollution from agriculture. First, I examine the issue of additionality of cost sharing payments by estimating the direct effect of cover crop cost sharing on the acres of cover crops, and the indirect effect of cover crop cost sharing on the acres of two other practices: conservation tillage and contour/strip cropping. A two-stage simultaneous equation approach is used to correct for voluntary self-selection into cost sharing programs and account for substitution effects among conservation practices. Quasi-random Halton sequences are employed to solve the system of equations for conservation practice acreage and to minimize the computational burden involved. By considering patterns of agronomic complementarity or substitution among conservation practices (Blum et al., 1997; USDA SARE, 2012), this analysis estimates water quality impacts of the crowding-in or crowding-out of private investment in conservation due to public incentive payments. Second, I connect the econometric behavioral results with model parameters from the EPA’s Chesapeake Bay Program to conduct a policy simulation on water quality effects. I expand the econometric model to also consider the potential loss of vegetative cover due to cropland incentive payments, or slippage (Lichtenberg and Smith-Ramirez, 2011). Econometric results are linked with the Chesapeake Bay Program watershed model to estimate the change in abatement levels and costs for nitrogen, phosphorus and sediment under various behavioral scenarios. Finally, I use inverse sampling weights to derive statewide abatement quantities and costs for each of these pollutants, comparing these with TMDL targets for agriculture in Maryland.
Resumo:
Matching theory and matching markets are a core component of modern economic theory and market design. This dissertation presents three original contributions to this area. The first essay constructs a matching mechanism in an incomplete information matching market in which the positive assortative match is the unique efficient and unique stable match. The mechanism asks each agent in the matching market to reveal her privately known type. Through its novel payment rule, truthful revelation forms an ex post Nash equilibrium in this setting. This mechanism works in one-, two- and many-sided matching markets, thus offering the first mechanism to unify these matching markets under a single mechanism design framework. The second essay confronts a problem of matching in an environment in which no efficient and incentive compatible matching mechanism exists due to matching externalities. I develop a two-stage matching game in which a contracting stage facilitates subsequent conditionally efficient and incentive compatible Vickrey auction stage. Infinite repetition of this two-stage matching game enforces the contract in every period. This mechanism produces inequitably distributed social improvement: parties to the contract receive all of the gains and then some. The final essay demonstrates the existence of prices which stably and efficiently partition a single set of agents into firms and workers, and match those two sets to each other. This pricing system extends Kelso and Crawford's general equilibrium results in a labor market matching model and links one- and two-sided matching markets as well.