Nahid Emad
Parallel and Distributed Eigenproblems for Large AI models
In this talk, we highlight the importance of high-performance eigenvalue computation for data analysis in large AI models, particularly in areas ranging from dimensionality reduction to understanding the internal dynamics of data and learning-models through the separation of target signals and noises in signal processing and analysis of and stability and convergence analysis of RNNs. We focus on dimensionality reduction problem which is a representative example of the strong interactions between machine learning and linear algebra, and which plays a prominent role in high-performance big data analysis. Using this problem, which is based on large sparse eigenproblem, we show how to take advantage of these interactions and commonalities to propose new approaches to problem solving in both domains. An innovative machine learning approach based on Unite and Conquer methods, used in linear algebra, will be presented. In addition to its efficiency from an accuracy point of view, the important characteristics of this inherently parallel and scalable technique make it well suited to multi-level and heterogeneous parallel and/or distributed architectures. Experimental results, partly on the #1 supercomputer of the HPCG list, Fugaku demonstrating the interest of the approach for efficient data analysis in the case of applications such as clustering, cybersecurity and health will be presented. We will finally emphasize that these methods open up very general perspectives showing their applicability in areas such as high-performance processing of GNNs and LLMs.
back to overview