Sangam: A Confluence of Knowledge Streams

OPTIMIZATION UNDER STOCHASTIC ENVIRONMENT

Show simple item record

dc.contributor Fu, Michael C
dc.contributor Digital Repository at the University of Maryland
dc.contributor University of Maryland (College Park, Md.)
dc.contributor Electrical Engineering
dc.creator Li, Yunchuan
dc.date 2020-10-10T05:33:44Z
dc.date 2020-10-10T05:33:44Z
dc.date 2020
dc.date.accessioned 2022-05-20T08:38:59Z
dc.date.available 2022-05-20T08:38:59Z
dc.identifier https://doi.org/10.13016/jfhk-sugx
dc.identifier http://hdl.handle.net/1903/26594
dc.identifier.uri http://localhost:8080/xmlui/handle/CUHPOERS/117677
dc.description Stochastic optimization (SO) is extensively studied in various fields, such as control engineering, operations research, and computer science. It has found wide applications ranging from path planning (civil engineering) and tool-life testing (industrial engineering) to Go-playing artificial intelligence (computer science). However, SO is usually a hard problem primarily because of the added complexity from random variables. The objective of this research is to investigate three types of SO problems: single-stage SO, multi-stage SO and fast real-time parameter estimation under stochastic environment.\par We first study the single-stage optimization problem. We propose Direct Gradient Augmented Response Surface Methodology (DiGARSM), a new sequential first-order method for optimizing a stochastic function. In this approach, gradients of the objective function with respect to the desired parameters are utilized in addition to response measurements. We intend to establish convergence of the proposed method, as well as traditional approaches which do not use gradients. We expect an improvement in convergence speed with the added derivative information. \par Second, we analyze a tree search problem with an underlying Markov decision process. Unlike traditional tree search algorithms where the goal is to maximize the cumulative reward in the learning process, the proposed method aims at identifying the best action at the root that achieves the highest reward. A new tree algorithm based on ranking and selection is proposed. The selection policy at each node aims at maximizing the probability of correctly selecting the best action. \par The third topic is motivated by problems arising in neuroscience, specifically, a Maximum Likelihood (ML) parameter estimation of linear models with noise-corrupted observations. We developed an optimization algorithm designed for non-convex, linear state-space model parameter estimation. The ML estimation is carried out by the Expectation-Maximization algorithm, which iteratively updates parameter estimates based on the previous estimates. Since the likelihood surface is in general non-convex, a model-based global optimization method called Model Reference Adaptive Search (MRAS) is applied.
dc.format application/pdf
dc.language en
dc.subject Electrical engineering
dc.subject Markov decision process
dc.subject Maximum likihood estimation
dc.subject Stochastic approximation
dc.subject Stochastic optimization
dc.title OPTIMIZATION UNDER STOCHASTIC ENVIRONMENT
dc.type Dissertation


Files in this item

Files Size Format View
Li_umd_0117E_21066.pdf 2.019Mb application/pdf View/Open

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse