ORIGINAL RESEARCH article

Front. Appl. Math. Stat.

Sec. Optimization

Volume 11 - 2025 | doi: 10.3389/fams.2025.1593680

This article is part of the Research TopicLarge Tensor Analysis and ApplicationsView all articles

Expectation-Maximization Alternating Least Squares for Tensor Network Logistic Regression

Provisionally accepted
Naoya  YamauchiNaoya Yamauchi1Hidekata  HontaniHidekata Hontani1Tatsuya  YokotaTatsuya Yokota1,2*
  • 1Nagoya Institute of Technology, Nagoya, Japan
  • 2RIKEN Center for Advanced Intelligence Project (AIP), Chuo-ku, Tokyo, Japan

The final, formatted version of the article will be published soon.

In recent years, a learning method for classifiers using tensor networks (TNs) has attracted attention. When constructing a classification function for high-dimensional data using a basis function model, a huge number of basis functions and coefficients are generally required, but the TN model makes it possible to avoid the curse of dimensionality by representing the huge coefficients using TNs. However, there is a problem with TN learning, namely the gradient vanishing, and learning using the gradient method cannot be performed efficiently. In this study, we propose a novel optimization algorithm for learning TN classifiers by using alternating least square (ALS) algorithm. Unlike conventional gradient-based methods, which suffer from vanishing gradients and inefficient training, our proposed approach can effectively minimize squared loss and logistic loss. To make ALS applicable to logistic regression, we introduce an auxiliary function derived from P ólya-Gamma augmentation, allowing logistic loss to be minimized as a weighted squared loss. We apply the proposed method to the MNIST classification task and discuss the effectiveness of the proposed method.

Keywords: Expectation-maximization (EM), Majorization-minimization (MM), Alternating Least Squares (ALS), tensor networks, Tensor train, Logistic regression, P ólya-Gamma (PG) Augmentation

Received: 14 Mar 2025; Accepted: 15 Apr 2025.

Copyright: © 2025 Yamauchi, Hontani and Yokota. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Tatsuya Yokota, Nagoya Institute of Technology, Nagoya, Japan

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.