Research Topic

High-Performance Tensor Computations in Scientific Computing and Data Science

About this Research Topic

Matrix and tensor computations are at the core of a multitude of applications in diverse domains of scientific computing and data science. These computations present several challenges due to their complexity, high computational cost, and large memory footprint. While there is a widespread availability and use of general-purpose high-performance libraries for matrix computations on multicore CPUs and GPUs, the same is not true for tensor computations. In many different domains, application developers are largely developing domain-specific libraries and frameworks with little shared usability across disciplines resulting into significant replication of effort. In addition, researchers working on tensors computations lack a dedicated specialized publication outlet and tend to publish their results in journals across many different fields. As such, scientific advances on tensors are scattered and, despite there is a growing community working on tensors computations, researchers in this field have a rather limited visibility and tend to work in a compartmentalized fashion.

This Research Topic aims at bringing together leading experts from distinct domains such as Computational Chemistry, Condensed Matter Physics, Scientific Computing and Machine Learning, just to name a few, to uncover computational challenges, bottlenecks, and advances in high-performance tensor computations arising in those disciplines. The aim is to enhance understanding of the similarities and differences in the tensor operations and computational tasks across these fields, and to seek pathways to general purpose software libraries and frameworks for high-performance tensor computations. The ambition of this topic is to close the gap between the many different conventions used to represent tensor computations and create a common language both in terms of abstractions and granularity of computational kernels. This also takes into account the analytical and algebraic foundations of structured tensor decompositions and approximations. In the long term, the vision is to create a framework for tensor computations that is understandable by the many practitioners and can be utilized to carry out efficient, parallel, and high-performance tensor calculations.

In this call for contributions, we welcome original research manuscripts and review manuscripts that lie at the intersection of tensor computations, high-performance scientific computing and machine learning, focusing, but not limited to:

• High-performance tensor contractions in scientific computing
• Massively parallel tensor calculations
• Tensor networks in quantum physics and machine learning
• Tensor decompositions and approximations
• Tensor abstractions and representations
• High-performance algorithms and implementations of tensor operations
• Tensor libraries and compilers
• Optimization of algorithms for tensor computations
• Optimization methods in numerical multi-linear algebra
• Tensor methods in applied computational domains
• Tensor operations with application to Machine and Deep learning
• Emerging topics such as tensor benchmarking and tensor generation

Image Credit: Melinda Green and Andrey Astrelin


Keywords: Tensor contractions, Tensor decompositions, Tensor networks, Quantum chemistry and physics, Multi-linear algebra, N-way data.


Important Note: All contributions to this Research Topic must be within the scope of the section and journal to which they are submitted, as defined in their mission statements. Frontiers reserves the right to guide an out-of-scope manuscript to a more suitable section or journal at any stage of peer review.

Matrix and tensor computations are at the core of a multitude of applications in diverse domains of scientific computing and data science. These computations present several challenges due to their complexity, high computational cost, and large memory footprint. While there is a widespread availability and use of general-purpose high-performance libraries for matrix computations on multicore CPUs and GPUs, the same is not true for tensor computations. In many different domains, application developers are largely developing domain-specific libraries and frameworks with little shared usability across disciplines resulting into significant replication of effort. In addition, researchers working on tensors computations lack a dedicated specialized publication outlet and tend to publish their results in journals across many different fields. As such, scientific advances on tensors are scattered and, despite there is a growing community working on tensors computations, researchers in this field have a rather limited visibility and tend to work in a compartmentalized fashion.

This Research Topic aims at bringing together leading experts from distinct domains such as Computational Chemistry, Condensed Matter Physics, Scientific Computing and Machine Learning, just to name a few, to uncover computational challenges, bottlenecks, and advances in high-performance tensor computations arising in those disciplines. The aim is to enhance understanding of the similarities and differences in the tensor operations and computational tasks across these fields, and to seek pathways to general purpose software libraries and frameworks for high-performance tensor computations. The ambition of this topic is to close the gap between the many different conventions used to represent tensor computations and create a common language both in terms of abstractions and granularity of computational kernels. This also takes into account the analytical and algebraic foundations of structured tensor decompositions and approximations. In the long term, the vision is to create a framework for tensor computations that is understandable by the many practitioners and can be utilized to carry out efficient, parallel, and high-performance tensor calculations.

In this call for contributions, we welcome original research manuscripts and review manuscripts that lie at the intersection of tensor computations, high-performance scientific computing and machine learning, focusing, but not limited to:

• High-performance tensor contractions in scientific computing
• Massively parallel tensor calculations
• Tensor networks in quantum physics and machine learning
• Tensor decompositions and approximations
• Tensor abstractions and representations
• High-performance algorithms and implementations of tensor operations
• Tensor libraries and compilers
• Optimization of algorithms for tensor computations
• Optimization methods in numerical multi-linear algebra
• Tensor methods in applied computational domains
• Tensor operations with application to Machine and Deep learning
• Emerging topics such as tensor benchmarking and tensor generation

Image Credit: Melinda Green and Andrey Astrelin


Keywords: Tensor contractions, Tensor decompositions, Tensor networks, Quantum chemistry and physics, Multi-linear algebra, N-way data.


Important Note: All contributions to this Research Topic must be within the scope of the section and journal to which they are submitted, as defined in their mission statements. Frontiers reserves the right to guide an out-of-scope manuscript to a more suitable section or journal at any stage of peer review.

About Frontiers Research Topics

With their unique mixes of varied contributions from Original Research to Review Articles, Research Topics unify the most influential researchers, the latest key findings and historical advances in a hot research area! Find out more on how to host your own Frontiers Research Topic or contribute to one as an author.

Topic Editors

Loading..

Submission Deadlines

01 October 2021 Manuscript

Participating Journals

Manuscripts can be submitted to this Research Topic via the following journals:

Loading..

Topic Editors

Loading..

Submission Deadlines

01 October 2021 Manuscript

Participating Journals

Manuscripts can be submitted to this Research Topic via the following journals:

Loading..
Loading..

total views article views article downloads topic views

}
 
Top countries
Top referring sites
Loading..