
Statistical Learning for Big and Complex Data
Saturday, Aug. 26, 2023




Zhejiang University
Title: Proximal Causal Learning of Heterogeneous Treatment Effects
Abstract: Efficiently and flexibly estimating treatment effect heterogeneity is an important task in a wide variety of settings ranging from medicine to marketing, and there are a considerable number of promising conditional average treatment effect estimators currently available. These, however, typically rely on the assumption that the measured covariates are enough to justify conditional exchangeability. We propose the P-learner, motivated by the R- and DR-learner, a tailored two-stage loss function for learning heterogeneous treatment effects in settings where exchangeability given observed covariates is an implausible assumption, and we wish to rely on proxy variables for causal inference. Our proposed estimator can be implemented by off-the-shelf loss-minimizing machine learning methods, which in the case of kernel regression satisfies an oracle bound on the estimated error as long as the nuisance components are estimated reasonably well.
Renmin University of China
Title: Learning Invariant Representations for Algorithmic Fairness and Domain Generalization with Minimax Optimality
Abstract: Machine learning methods often assume that the test data have the same distribution as the training data. However, this assumption may not hold due to multiple levels of heterogeneity. In this work, we address the problem of fair and generalizable machine learning in unseen environments. we first propose a training environment-based oracle, FAIRM, which has desirable fairness and domain generalization properties in unseen environments under a diversity-type of conditions. We then adapt the FAIRM framework to linear models and consider two classes of invariant representations: invariant features and invariant subspaces. We develop efficient algorithms for these models and provide finite sample guarantees. We evaluate our method on Color-MNIST data and a census income data set based on high-dimensional linear models and show that it outperforms ERM and Maxmin estimators.
Southwest University of Finance and Economics
Title: Ensemble Methods for Testing a Global Null
Abstract: Testing a global null is a canonical problem in statistics and has a wide range of applications. In view of the fact that no uniformly most powerful test exists, prior and/or domain knowledge are commonly used to focus on a certain class of alternatives to improve the testing power. However, it is generally challenging to develop tests that are particularly powerful against a certain class of alternatives. In this paper, motivated by the success of ensemble learning methods for prediction or classification, we propose an ensemble framework for testing that mimics the spirit of random forests to deal with the challenges. Our ensemble testing framework aggregates a collection of weak base tests to form a final ensemble test that maintains strong and robust power. We apply the framework to four problems about global testing in different classes of alternatives arising from Whole Genome Sequencing (WGS) association studies. Specific ensemble tests are proposed for each of these problems, and their theoretical optimality is established in terms of Bahadur efficiency. Extensive simulations and an analysis of a real WGS dataset are conducted to demonstrate the type I error control and/or power gain of the proposed ensemble tests.
Hong Kong University of Science and Technology
Title: Online Tensor Learning: Computational and Statistical Trade-Offs, Adaptivity and Optimal Regret
Abstract: Large tensor learning algorithms are usually computationally expensive and require storing an exceedingly large amount of data. In this paper, we propose a unified online Riemannian gradient descent (oRGrad) algorithm for tensor learning, which is computationally fast, consumes much less memory, and can deal with sequentially arrived data and make timely prediction. The algorithm is applicable to both linear and generalized linear models. We observe an intriguing trade-off between the computational convergence rate and statistical rate of oRGrad. Increasing the step size accelerates computational convergence but leads to larger statistical error, whereas a smaller step size yields smaller statistical error at the expense of slower computational convergence. If the time horizon T is known, oRGrad achieves statistical optimality by choosing an appropriate fixed step size. Besides the aforementioned benefits, we discover that noisy tensor completion is particularly blessed by online algorithms in that it avoids the trimming procedure and guarantees sharp entry-wise statistical error, which is usually deemed technically challenging for offline methods. Online algorithms render powerful martingale tools applicable for theoretical analysis. The regret of oRGrad is investigated and a fascinating trilemma concerning the computational convergence rate, statistical error, and regret is observed. By choosing an appropriate constant step size, oRGrad achieves an O(T^1/2) regret. We then introduce the adaptive-oRGrad algorithm, which can achieve the optimal O(log T ) regret by adaptively selecting the step sizes, regardless of whether the time horizon is known or unknown. The adaptive-oRGrad algorithm can also attain statistically optimal error without knowing the horizon. Comprehensive numerical simulation results corroborate our theoretical findings.