Skip to yearly menu bar Skip to main content


Poster

Regression Learning with Limited Observations of Multivariate Responses and Features

Yifan Sun · Grace Yi


Abstract: Multivariate linear regression models are broadly used to facilitate the relationship between multi-dimensional real-valued labels and features. However, their effectiveness is compromised by the presence of missing observations, a ubiquitous challenge in real-world applications. Considering a scenario where learners access only limited components for both responses and features, we develop efficient algorithms tailored for the least squares ($L_2$) and least absolute ($L_1$) loss functions, each coupled with a ridge or Lasso-type penalty, and establish rigorous error bounds for all proposed algorithms. Notably, our $L_2$ loss function algorithms are probably approximately correct (PAC), distinguishing them from their $L_1$ counterparts. Extensive numerical experiments unveil that our approach surpasses methods naively applying existing algorithms for univariate label to each coordinate of multivariate labels. Further, utilizing the $L_1$ loss function or introducing a Lasso-type penalty can enhance predictions in the presence of outliers or high dimensional features. This research contributes valuable insights into addressing the challenges posed by missing data.

Live content is unavailable. Log in and register to view live content