Research Portfolio and Projects

The group’s research interests lie in the intersection of simulation, stochastic optimization, and probabilistic data analysis. We pursue creating robust and fast tools for analysis and decision-making under uncertainty. We are actively involved in improving the computing (cyber)infrastructure for optimization in stochastic settings. We are also invested in bridging between the Monte Carlo perspective and machine learning. 

All ongoing and prospective projects of the group can be viewed within two main thrusts:

(I) Efficiency and Reliability of Simulation(-based) Optimization Algorithms

Finding optimal solutions to problems represented with stochastic simulations is hard. We seek designing robust and efficient simulation optimization algorithms.

I.B) Experienced with various challenges in simulation optimization, we have actively engaged in establishing strong infrastructure to help the development and improvement of existing tools and solvers in the area. One such effort has been revitalizing “SimOpt”, an open-source object-oriented library, testbed, and benchmarking platform (recently funded by an NSF grant). The library enjoys new evaluation metrics carefully devised to address the solver’s probabilistic behavior in finite-time as well as their handling of stochastic constraints. Experimental design procedures tune hyper-parameters and can help in other developing directions.

(II) Monte Carlo Simulation Methods for Machine Learning and Big Data

Learning from Big Data is computationally challenging. Monte Carlo theories glean distributional insights for interpretable prediction / prevention of high-risk events.
II.A) For calibrating digital twins (a project supported by an NSF grant), we seek to enhance solvers’ ability to handle big data. This challenge, motivated by computer models that predict power generation in wind farms, has led to integrating the well-known variance reduction technique—stratified sampling—within continuous optimization. To maximally improve solver efficiency, how many samples is not the only main concern, but which samples also! And the goal is not to economize estimation of one function value, but rather the estimation of function value of the entire trajectory of incumbents. Therefore, both adaptive budget allocation and adaptive stratification may help with big data. We exploit dynamic partitioning via binary trees, guided by an in-depth study of score-based probabilistic trees, and a classical simulation method based on control variates for this purpose. We also prove the impact of well-implemented stratified adaptive sampling in the overall complexity (computational cost) of an optimization algorithm.