Blanchet, J. and Kang, Y. Sample-out-of-sample Inference Based on Wasserstein Distance. To appear in Operations Research.

 

Summary:

This paper was actually produced slightly before “Blanchet, J., Kang, Y., Murthy, K. Robust Wasserstein Profile Inference and Applications to Machine Learning. Journal of Applied Probability, 56, (2019), pp. 830-857.”. Our motivation is also DRO (Distributionally Robust Optimization) based on the Wasserstein distance. The cool thing is that the support of the alternative measures in the uncertainty set is chosen at random, based on a sample. I think this is very interesting because this is a way to inform the uncertainty set in a data-driven way. We actually use these results in the context of semi-supervised learning. But also the asymptotic statistics of the profile function are also very interesting. In this setting, in contrast to the unconstrained case (i.e. paper cited above), the rate of convergence is dimension dependent but never worse than the root of the unconstrained case.

 

Bibtex:

@Article{BK_2020,

    author = {J. Blanchet and Y. Kang},

    title = {Sample-out-of-sample Inference Based on Wasserstein Distance.},

    journal = {Operations Research},

    year = {2020},

    volume = {To appear},

%    pages = {271-276}

}