ORIGINAL RESEARCH article
Front. Appl. Math. Stat.
Sec. Optimization
Volume 11 - 2025 | doi: 10.3389/fams.2025.1593680
This article is part of the Research TopicLarge Tensor Analysis and ApplicationsView all articles
Expectation-Maximization Alternating Least Squares for Tensor Network Logistic Regression
Provisionally accepted- 1Nagoya Institute of Technology, Nagoya, Japan
- 2RIKEN Center for Advanced Intelligence Project (AIP), Chuo-ku, Tokyo, Japan
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
In recent years, a learning method for classifiers using tensor networks (TNs) has attracted attention. When constructing a classification function for high-dimensional data using a basis function model, a huge number of basis functions and coefficients are generally required, but the TN model makes it possible to avoid the curse of dimensionality by representing the huge coefficients using TNs. However, there is a problem with TN learning, namely the gradient vanishing, and learning using the gradient method cannot be performed efficiently. In this study, we propose a novel optimization algorithm for learning TN classifiers by using alternating least square (ALS) algorithm. Unlike conventional gradient-based methods, which suffer from vanishing gradients and inefficient training, our proposed approach can effectively minimize squared loss and logistic loss. To make ALS applicable to logistic regression, we introduce an auxiliary function derived from P ólya-Gamma augmentation, allowing logistic loss to be minimized as a weighted squared loss. We apply the proposed method to the MNIST classification task and discuss the effectiveness of the proposed method.
Keywords: Expectation-maximization (EM), Majorization-minimization (MM), Alternating Least Squares (ALS), tensor networks, Tensor train, Logistic regression, P ólya-Gamma (PG) Augmentation
Received: 14 Mar 2025; Accepted: 15 Apr 2025.
Copyright: © 2025 Yamauchi, Hontani and Yokota. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Tatsuya Yokota, Nagoya Institute of Technology, Nagoya, Japan
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.