27th Conference of the International Linear Algebra Society (ILAS 2026)
Virginia Tech
The 27th Conference of the International Linear Algebra Society will be held in Blacksburg, Virginia from May 18-22, 2026, on the campus of Virginia Tech. The conference's theme, "Linear Algebra on the Blue Ridge: Panoramas of Theory and Application," highlights the geographical setting for the conference in the mountains of southwest Virginia, but also serves as an invitation to researchers from across linear algebra, ranging from core areas through to numerical analysis, applications, and linear algebra education. Further details, such as registration information, lodging details, and program updates, will be available in due course.
We look forward to a vibrant ILAS meeting in May 2026, and hope to see you in Blacksburg!
-
-
7:30 AM
Registration McBryde Hall 100 Foyer
McBryde Hall 100 Foyer
Virginia Tech
-
1
Introductory Remarks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
Plenary Talks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
2
A new age of iterative Krylov methods for inverse problems: What to do with expensive inner-product computations?
Iterative Krylov projection methods have become widely used for solving large-scale linear inverse problems. However, methods based on orthogonality include computations of inner-products, which becomes costly when the number of iterations is high, are a bottleneck for parallelization, and can cause the algorithms to break down due to information loss in the projections.
In this talk, I will describe two new approaches to handle expensive inner-product computations in the context of solving large-scale inverse problems. First, we describe new iterative solvers based on the randomized Golub-Kahan approach, where sketched inner products are used to estimate inner products of high-dimensional vectors. Second, we describe new inner-product-free Krylov solvers that avoid inner-products completely. For both approaches, we also describe hybrid methods that combine iterative projection methods with Tikhonov regularization, where regularization parameters can be selected automatically during the iterative process. Numerical results from image reconstruction show the potential benefits of these approaches.
Speaker: Julianne Chung (Emory University) -
3
Optimal interpolation in Hardy, Bergman and $\ell^p_A$ spaces: a reproducing kernel Banach space approach
After a review of the reproducing kernel Banach space framework and semi-inner products, we apply the techniques to the settings of sequence spaces $\ell^p$ (including the finite dimensional case), the associated function space $\ell_A^p$, Hardy spaces $H^p$ and Bergman spaces $A^p$, $1<p<\infty$, on the unit ball in ${\mathbb C}^n$, as well as the Hardy space on the polydisk and half-space. In particular, we show how the framework leads to a procedure to find a minimal norm element $f$ satisfying interpolation conditions $f(z_j)=w_j$, $j=1,\ldots , n$. We also provide numerical examples.
This talk is based on joint works with Gilbert Groenewald, Sanne ter Horst and Eder Kikianty
Speaker: Hugo Woerdeman (Drexel University)
-
2
-
10:15 AM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Numerical Linear Algebra in Machine Learning: A McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
4
Understanding and Leveraging Adaptive Algorithms' Sensitivity to Change-of-Basis
Adaptive gradient optimization algorithms—including Adam, Adagrad, and their variants—have found widespread use in machine learning, signal processing, and many other settings. However many algorithms in this family are not rotationally equivariant: in this talk we examine how a simple change-of-basis in either parameter space or data space can drastically impact both the convergence rates and the generalization of these algorithms. We begin by studying reparameterizations in parameter space, and describe a data-driven method proposed in our recent work which produces a “favorable” basis for adaptive algorithms. Our method is an orthonormal transformation based on the expected gradient outer product (EGOP) matrix. We present theoretical results and empirical evidence that reparameterizations based on the EGOP eigenbasis can improve convergence of adaptive gradient methods, even when these leading eigenspaces are approximated using randomized numerical linear algebra methods. We show that for a broad class of functions, the sensitivity of adaptive algorithms to choice-of-basis is influenced by the decay of the EGOP matrix spectrum. We illustrate the potential impact of EGOP reparameterization by presenting empirical evidence and theoretical arguments that common machine learning tasks with ``natural'' data exhibit EGOP spectral decay.
Speaker: Adela DePavia (The University of Chicago) -
5
Fast and explainable clustering in the Manhattan and Tanimoto distance
The CLASSIX algorithm is a fast and explainable approach to data clustering. In its original form, this method utilizes the first principal component of the data matrix to truncate the search for nearby data points, using the Cauchy-Schwarz inequality, with proximity being defined in terms of the Euclidean distance. In this work, we demonstrate methods to extend CLASSIX to other distance measures by showcasing its effectiveness in the Manhattan distance and the Tanimoto distance. CLASSIX in these two distance metrics uses the 1-norm of the data vectors as the sorting criterion. The triangle inequality is used as a general search truncation criterion applicable to any p-norm, and the Baldi intersection inequality is used as a search truncation criterion for the Tanimoto distance.
Speaker: Kaustubh Roy (University of Manchester) -
6
Numerical linear algebra with neural operator preconditioning for solving some parametric PDEs
This work exclusively focuses on the mixed precision algorithms that integrate classical numerical linear algebra methods with nonlinear neural network-based preconditioners to accelerate the solution of some
parametric Partial Differential Equations (PDEs). Specifically, we consider Krylov subspace methods such as Flexible GMRES (FGMRES) and Flexible FOM (FFOM), combined with nonlinear or flexible preconditioners derived from the trained neural operator models. The parametric PDEs
addressed include the Helmholtz equations, Poisson equations, Darcy flow, and Diffusion-Advection equations, spanning both academic benchmarks and practical dataset. Compared to the classical
numerical preconditioners, the trained neural operator-based preconditioners exhibit significant generalization capabilities in addressing a wide range of numerical and parametric variations. These include the ability to address varying source term, speed of sound, diffusivity term, velocity field for advection, and varying boundary conditions for these parametric PDEs. In addition, we support both CPUs and GPUs implementations of the proposed mixed precision algorithms. Overall, this work demonstrates the efficiency and flexibility of combining modern neural network-based solvers with classical Krylov methods, leveraging the strengths of both to achieve higher attainable accuracy and broader applicability across diverse problem settings.Speaker: Yanfei Xiang (University of Strasbourg)
-
4
-
Matrix Nearness Problems: A Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
7
Nearest matrix with multiple eigenvalues by Riemannian optimization
Given a square complex matrix $A$, we tackle the problem of finding the nearest matrix with multiple eigenvalues or, equivalently when $A$ had distinct eigenvalues, the nearest defective matrix. To this goal, we extend the general framework described in [M. Gnazzo, V. Noferini, L. Nyman, F. Poloni, Riemann-Oracle: A general-purpose Riemannian optimizer to solve nearness problems in matrix theory, Found. Comput. Math. 2025] and based on variable projection and Riemannian optimization, allowing the ambient manifold to simultaneously track left and right eigenvectors. Our method also allows us to impose arbitrary complex-linear constraints on either the perturbation or the perturbed matrix; this can be useful to study structured eigenvalue condition numbers. We present numerical experiments, comparing with preexisting algorithms.
Speaker: Vanni Noferini (Aalto University) -
8
Robustly asymptotically stable dissipative Hamiltonian descriptor systems
We analyze the robust asymptotic stability under structure-preserving perturbations for the class of linear time-invariant dissipative-Hamiltonian differential-algebraic (dHDAE) systems. We show how to compute the distance to the nearest singular and high index system and determine stability radii for the finite spectrum under structure preserving perturbations
Speaker: Volker Mehrmann (TU Berlin) -
9
Bounds on the geodesic distances on the Stiefel manifold for a family of Riemannian metrics
We give bounds on geodesic distances on the Stiefel manifold, derived from new geometric insights. The considered geodesic distances are induced by the one-parameter family of Riemannian metrics introduced by Hüper et al. (2021), which contains the well-known Euclidean and canonical metrics. First, we give the best Lipschitz constants between the distances induced by any two members of the family of metrics. Then, we give a lower and an upper bound on the geodesic distance by the easily computable Frobenius distance. We give explicit families of pairs of matrices that depend on the parameter of the metric and the dimensions of the manifold, where the lower and the upper bound are attained. These bounds aim at improving the theoretical guarantees and performance of minimal geodesic computation algorithms by reducing the initial velocity search space. In addition, these findings contribute to advancing the understanding of geodesic distances on the Stiefel manifold and their applications.
Speaker: Simon Mataigne (UCLouvain)
-
7
-
Numerical Ranges and Numerical Radii: A Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
10
Matrix numerical ranges of Toeplitz operators with polynomial symbols
Using results from Brown-Halmos, Klein (1972) described the numerical range of a general Toeplitz operator on $H^2(\mathbb{D})$. In particular, the numerical range of a Toeplitz operator $T_p$ with polynomial symbol $p$ is the convex hull of the image of the unit disk under $p$. By analyzing $p(\mathbb{T})$ and its relationship to the Kippenhahn curve of a matrix, we provide conditions under which the closure of $W(T_p)$ is the numerical range of a finite matrix $M_p$. Results about the eigenvalues of $M_p$ will be discussed as well.
Speaker: Linda Patton (Cal Poly San Luis Obispo) -
11
Multiplicities and $k$-Numerical Range
Given $1\le k\le n$, the $k$-numerical range of $A\in \mathbb{C}_{n\times n}$ is defined by
$$ W_k(A): = \left\{ \sum_{i=1}^k x_i^*Ax_i: x_1, \dots, x_k\ \mbox {orthonormal vectors in } \mathbb{C}^n\right\}\subset \mathbb{C}. $$ Motivated by Davis' intuitive explanation of the Elliptical Range Theorem, we introduce two notions of multiplicity for points in $W_k(A)$, namely wedge multiplicity and projection multiplicity. The wedge multiplicity is related to the Grassmannian and the projection multiplicity is related to the set of rank $k$ orthogonal projectors. We present several results concerning each notion and provide examples illustrating these multiplicities. The corresponding real analogues $V_k(A)$ are also studied.Speaker: Nancy Menzelthe (University of Nevada, Reno) -
12
On the spatial numerical range
The spatial numerical range of an operator $T$ on a normed space $(\mathcal{X}, \| \cdot \|)$ is the set $$W(T) = \{f(Tx) : x \in \mathcal{X}, f \in \mathcal{X}^*, \|x\| = \|f\|^d = f(x) = 1\};$$ when $| \cdot |$ is induced by an inner product this coincides with the classical numerical range. We investigate some properties of the spatial numerical range.
Speaker: Edward Poon
-
10
-
Where Algebraic Coding Theory and Graph Theory Meet: A Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
13
An Eventown Result for Permutations
A family of permutations $\mathcal{F} \subseteq S_n$ is even-cycle-intersecting if $\sigma \pi^{-1}$ has an even cycle for all $\sigma,\pi \in \mathcal{F}$. We show that if $\mathcal{F} \subseteq S_n$ is an even-cycle-intersecting family of permutations, then $|\mathcal{F}| \leq 2^{n-1}$, and that equality holds when $n$ is a power of 2 and $\mathcal{F}$ is a double-translate of a Sylow 2-subgroup of $S_n$. This result can be seen as an analogue of the classical eventown problem for subsets and it confirms a conjecture of János Körner on maximum reversing families of the symmetric group. Along the way, we show that the canonically intersecting families of $S_n$ are also the extremal odd-cycle-intersecting families of $S_n$ for all even $n$. While the latter result has less combinatorial significance, its proof uses an interesting new character-theoretic identity that might be of independent interest in algebraic combinatorics.
Speaker: Nathan Lindzey (University of Memphis) -
14
A graph from the injection metric
The injection metric measures the distance between two subspaces and naturally arises in the context of subspace coding, when the codewords potentially have different dimensions. In this talk, we consider a graph associated with the injection metric and examine how its structure determines the existence of good codes for subspace coding.
Speaker: Matteo Bertuzzo (Eindhoven University of Technology) -
15
Coding theory via graph theory
We discuss the use of graph theory methods to attack coding theory problems.
This is joint work with Aida Abiad and Harper Reijnders.
Speaker: Michael Tait (Villanova University)
-
13
-
Krylov Iterative Methods for Linear Equations: A McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
16
Krylov Methods for Rank-one Updates of Eigenvalue Problems and Linear Equations
Krylov methods are given for rank-one updates of both eigenvalue and linear equations problems. For eigenvalues, an Arnoldi iteration for the original matrix can be continued on the rank-one changed matrix. We discuss how careful implementation allows the desired accuracy to be attained for the updated matrix. Next, methods are given for linear equations, one that uses the updated Arnoldi approach and the other has the Sherman-Morrison formula. We also give some theoretical results about how eigenvalues change with rank-one perturbation.
Speaker: Dona Ishara Saparamadu (Baylor University, Waco, Texas) -
17
Convergence Analysis for Infinite GMRES
Many applications require the solution of large-scale linear systems that have nonlinear parameter dependence. The Infinite GMRES algorithm, developed by Jarlebring and Correnty in 2022, converts these systems into infinite-dimensional systems with linear parameter dependence. This transformation involves a linearization process that results in a system with a special companion-like structure. Due to the shift invariance of the Krylov subspaces, all choices of parameter give the same Krylov subspaces. Thus performing the Arnoldi process for one parameter enables the efficient construction of the GMRES iterates for all parameter values. The special structure of the right-hand-side vector of the infinite-dimensional system allows this Arnoldi process to be executed with finite-dimensional vectors, where the dimension grows with every iteration. As such, Infinite GMRES uses finite-length Krylov basis vectors to solve an infinite-dimensional system. The special structure of the infinite-dimensional system invites further investigation into the convergence behavior of the Infinite GMRES residual norm. We aim to understand how various components of this system, including the size and nature of the block matrices, the number of blocks, and the choice of parameter value influence convergence behavior. In this presentation, we will provide rigorous convergence results using various tools of analysis, including the numerical range.
Speaker: Jordan Jackson (Virginia Tech) -
18
Twin CG for Linear Equations with Multiple Right-hand Sides
A very simple approach to solving multiple right-hand side systems is proposed. For symmetric problems, the conjugate gradient method is a very efficient way to solve linear equations. We will use the same parameters from solving the first system for other systems. This is called Twin CG. It corresponds to applying the same polynomial to the other systems as was used for the first system. No dot products are needed except for the first system, and the method is very parallelizable. Deflation of eigenvalues using seeding can be included. The remarkable natural stability control due to roundoff error in the symmetric Lanczos iteration will be discussed. Added stability control methods include deflating eigenvalues with linear factors and applying partial Global CG.
Speaker: Abigail Williams (Baylor University)
-
16
-
Algebraic Invariants of Graphs: A Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
19
Connected domination and the v-number of binomial edge ideals
A connected dominating set of a graph is a vertex set that induces a connected subgraph and such that every vertex outside the set is adjacent to at least one vertex in the set. The minimum cardinality of a connected dominating set is called the connected domination number. We present an algebraic expression for this combinatorial invariant using the theory of binomial edge ideals.
Speaker: Delio Jaramillo Velez (Virginia Tech) -
20
The characterization of graphs with two trivial distance ideals
The distance ideals of connected graphs are algebraic invariants extending the Smith normal form (SNF) and the spectrum of graph distance matrices.
In general, distance ideals are not monotone under taking induced subgraphs.
However, it was proved in 2017 that the set of graphs with one trivial distance ideal over $\mathbb{Z}[X]$ and over $\mathbb{Q}[X]$ was characterized in terms of induced subgraphs, where $X$ is a set of variables indexed by the vertices.
Here, we give a characterization of the $\{\mathcal{F},\textsf{odd-holes}_{7}\}$-free graphs, where $\textsf{odd-holes}_{7}$ consists of the odd cycles of length at least seven and $\cal F$ is a set of sixteen graphs. Moreover, we show that the $\{\mathcal{F},\textsf{odd-holes}_{7}\}$-free graphs are precisely the graphs with at most two trivial distance ideals over $\mathbb{Z}[X]$.
As byproduct, we also find that the determinant of the distance matrix of a connected bipartite graph is even, this suggests that it is possible to extend, to bipartite graphs, the Graham-Pollak-Lovász celebrated formula $\det(D(T_{n+1}))=(-1)^nn2^{n-1}$, and Hou-Woo's result stating that $\text{SNF}(D(T_{n+1}))=\sf{I}_2\oplus 2\sf{I}_{n-2}\oplus (2n)$, for any tree $T_{n+1}$ with $n+1$ vertices.
Furthermore, we also determine the graphs with at most two trivial distance ideals over $\mathbb{Q}[X]$, and the graphs with at most two trivial distance univariate ideals. We conclude by showing that the SNF of the distance matrix of a graph has exactly two invariant factors equal to 1 if and only if it is bipartite or complete tripartite.Speaker: Juan Pablo Serrano Perez (Cinvestav) -
21
Smith Normal Forms of Graphical Hermite Simplices
We introduce the family of graphical Hermite simplices and study the Smith normal forms of their matrices of vertex vectors, which is equivalent to studying the group structure of the cokernels for these matrices.
Our motivation is to study the behavior of lattice simplices subject to small lattice perturbations of their vertices.
In this case, a graphical Hermite simplex is a perturbation of a rectangular simplex, i.e., a simplex defined by a diagonal matrix and the origin, with the perturbation controlled by the structure of a directed graph.
We first establish sufficient conditions on the graphs and diagonal entries of these matrices that imply having a single non-unit invariant factor, i.e., a cyclic cokernel.
We then obtain bounds on the invariant factors of the defining matrices related to lengths of paths in the corresponding directed graph.Speaker: Antwon Park (University of Kentucky)
-
19
-
Low-Complexity Data-driven or Classical Algorithms and Applications: A Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
22
Convergence of Randomized and Greedy Block Gauss-Seidel methods, as well as Asynchronous Iterations
We extend results known for the randomized (point and block)
Gauss-Seidel and the Gauss-Southwell methods for the case of a Hermitian and positive definite matrix to certain classes of non-Hermitian matrices. We consider cases with overlapping variables (as in Domain Decomposition). We obtain convergence results for a whole range of parameters describing the probabilities in the randomized method or the greedy choice strategy in the Gauss-Southwell-type methods. We identify those choices which make our convergence bounds best possible. One result is that the best convergence bounds that we obtain for the expected values in the randomized algorithm are as good as the best for the deterministic, but more costly algorithms of Gauss-Southwell type. We use these new results to show a provable convergence rate for asynchronous iterations.Speaker: Daniel Szyld (Temple University) -
23
Randomized Sketching for Tucker Tensors: From Compressed Summation to GMRES
Tensor equations of the form $L(X)=B$, with $L$ a sum of Kronecker products, arise across scientific computing, from high-dimensional PDEs to large-scale inverse problems. When the tensor order is moderate but mode sizes are very large, the Tucker format is an attractive choice — yet using it inside iterative solvers is notoriously hard, as operator applications, orthogonalization, and linear combinations all trigger rank growth and force expensive truncations of large intermediate tensors.
In this talk, I will present a line of work that addresses these bottlenecks through randomized sketching applied directly to Tucker factors. The first part introduces sketching-based methods for Tucker tensor summation that exploit the structure of Khatri–Rao and Kronecker products to perform compressed arithmetic without ever forming dense intermediates. Building on this foundation, the second part develops a randomized Tucker-sketched GMRES solver in two variants: a robust method based on the randomized HOSVD with adaptive accuracy control, and a faster one built on the multilinear Nyström approximation, which — at the price of a prescribed maximal rank — accelerates orthogonalization, avoids storing the Krylov basis, and assembles the solution at negligible cost. Numerical experiments on image deblurring, a parameter-dependent elliptic problem, a transport equation, and a convection–diffusion problem illustrate the efficiency and robustness of the framework.Speaker: Mirjeta Pasha -
24
Novel Randomized Tensor-Train Sketch and Applications
The Tensor-Train (TT) or Matrix-Product States (MPS) format provides a compact, low-rank representation for high-dimensional tensors, widely used in many-body quantum physics and quantum chemistry. Its efficiency relies on rounding, which reduces tensor ranks to maintain feasible computational costs.
In this talk, we introduce a novel block-structured randomized sketch exploiting the TT format and provide explicit probabilistic guarantees. We discuss how this sketch enables in particular randomized rounding algorithms for TT rounding, extending low-rank matrix approximation techniques. These methods significantly accelerate computations, particularly for summing TT-tensors or performing matrix-vector products, key operations in Krylov iterative solvers. We present numerical results demonstrating the empirical accuracy and computational advantages of the randomized approach over deterministic methods.This research was supported in part by Simons Travel Support for Mathematicians Award No. MPS-TSM-00966604.
Speaker: Paul Cazeaux (Virginia Tech)
-
22
-
Approximate Computing in Numerical Linear Algebra: A Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
25
Restoring similarity in randomized Krylov methods with applications to eigenvalue problems and matrix functions
The randomized Arnoldi process has been used in large-scale scientific computing because it produces a well-conditioned basis for the Krylov subspace more quickly than the standard Arnoldi process. However, the resulting Hessenberg matrix is generally not similar to the one produced by the standard Arnoldi process, which can lead to delays or spike-like irregularities in convergence. In this talk we introduce a modification of the randomized Arnoldi process that restores similarity with the Hessenberg matrix generated by the standard Arnoldi process. This is accomplished by enforcing orthogonality between the last Arnoldi vector and the previously generated subspace, which requires solving only one additional least-squares problem. When applied to eigenvalue problems and matrix function evaluations, the modified randomized Arnoldi process produces approximations that are identical to those obtained with the standard Arnoldi process. Numerical experiments demonstrate that our approach is as fast as the randomized Arnoldi process and as robust as the standard Arnoldi process.
This talk is based on a joint work with Laura Grigori, Daniel Kressner and Nian Shao (https://arxiv.org/abs/2601.10248).Speaker: Igor Simunec (EPFL) -
26
A general framework for Krylov ODE residuals with applications to randomized Krylov methods
Randomized Krylov subspace methods that employ the sketch-and-solve paradigm to substantially reduce orthogonalization cost have recently shown great promise in speeding up computations for many core linear algebra tasks (e.g., solving linear systems, eigenvalue problems and matrix equations, as well as approximating the action of matrix functions on vectors) whenever a nonsymmetric matrix is involved. An important application that requires approximating the action of matrix functions on vectors is the implementation of exponential integration schemes for ordinary differential equations. In this paper, we specifically analyze randomized Krylov methods from this point of view. In particular, we use the residual of the underlying differential equation to derive a new, reliable a posteriori error estimate that can be used to monitor convergence and decide when to stop the iteration. To do so, we first develop a very general framework for Krylov ODE residuals that unifies existing results, simplifies their derivation and allows extending the concept to a wide variety of methods beyond randomized Arnoldi (e.g., rational Krylov methods, Krylov methods using a non-standard inner product, ...). In addition, we discuss certain aspects regarding the efficient implementation of sketched Krylov methods. Numerical experiments on large-scale ODE models from real-world applications illustrate the quality of the error estimate as well as the general competitiveness of sketched Krylov methods for ODEs in comparison to other Krylov-based methods.
Speaker: Emil Krieger (Bergische Universität Wuppertal) -
27
A mixed precision algorithm for matrix root functions
Mixed precision computation has attracted great attention in recent years partly due to the evolution of machine learning and hardware infrastructure. Recent development on mixed precision algorithms has largely enhanced the performance of various linear algebra solvers. In this talk, we propose a mixed precision algorithm for the computation of matrix root functions, primarily the matrix square root and other higher order roots. We introduce a new refinement framework that is capable of refining the lower precision solution to the working precision level. We show that carefully designed mixed precision algorithms compute the matrix root functions to full working precision accuracy and offer speedup in most scenarios compared to the fixed precision method.
Speaker: Mr Bowen Gao (Fudan University)
-
25
-
Linear Algebra Education: A Torgersen Hall 2150
Torgersen Hall 2150
Virginia Tech
-
28
Redesigning Assessment in Applied Linear Algebra in the Age of Generative AI
Generative AI has reshaped student behavior in linear algebra courses: attendance has declined, and traditional take-home assessments are increasingly vulnerable to “solution acquisition” rather than learning. We present a redesign for Applied Linear Algebra courses that treats generative AI not as a shortcut, but as a catalyst for disciplinary literacy.
We replace an exam-heavy structure with a high-frequency, project-based model. Each week, students complete a short, contextualized mini-project (paired with an in-class quiz) that requires implementing core matrix methods and machine learning workflows—e.g., least squares, QR, SVD/PCA—on realistic datasets. To generate diverse, fresh prompts at scale, the instructor team uses campus-approved tools (Google Gemini and/or NotebookLM) to create project contexts and data variations. A custom AI assistant (“Gemini Gem”) provides guardrailed conceptual explanations, debugging hints, and reflective prompts while avoiding complete solutions; students document AI use and connect it back to course theory.
This experiment targets four goals: re-igniting class engagement, expanding implementation mastery across many methods, strengthening theory through immediate application, and fostering professional habits around deadlines and deliverables. We will present sample tasks and an evaluation plan comparing engagement, learning outcomes, and equity indicators with prior cohorts.
Speaker: Prof. Maria Isabel Bueno Cachadina (University of California Santa Barbara) -
29
Blind Singal Separation as a model to introduce Linear Transformatios
The use of real or realistic problems to introduce university students to linear algebra concepts through modeling has proven effective in stimulating student learning. We present an experience based on mathematical modeling and APOS theory (Action, Process, Object, and Schema) to introduce the concepts of matrix transformation and inverse matrix transformation in the context of an introductory Linear Algebra course for engineering students.
The instructional sequence is inspired by Blind Source Separation (BSS), a complex engineering problem concerned with recovering original sources from observed mixtures. Building on a prior in-depth study of BSS models, we identify a linear, noise-free formulation in which observations are modeled as linear combinations of sources. From this formulation, we carried out a transposition of the structural elements of the problem to an introductory linear algebra setting, interpreting signals as vectors and mixing processes as linear transformations.
Based on this model, the inverse transformation emerges as a natural response to the associated inverse problem. Rather than introducing the inverse matrix as a formal object or technique, the proposed genetic decomposition describes how the need for an inverse transformation arises from interpreting linear transformations as models of mixing processes. The construction involves the coordination of various mathematical processes, including the matrix--vector product, systems of linear equations, and the interpretation of linear transformations as functions between vector spaces. Invertibility is conceptualized in terms of information loss and the impossibility of reversing certain transformations.
This work contributes to the literature through the design of a modeling-based instructional situation that foregrounds the emergence and construction of inverse transformations in an introductory linear algebra context.
Speaker: Maria Trigueros (Benemérita Universidad Autónoma de Puebla, México) -
30
The Inquiry-Oriented Linear Algebra Project
The Inquiry-Oriented Linear Algebra (IOLA) project is a research-based, student-centered approach to the teaching and learning of introductory linear algebra. The IOLA curricular materials build from a set of experientially real tasks that allow for active student engagement in the guided reinvention of key mathematical ideas through student and instructor inquiry. The online instructional support materials include various resources such as rationales for task design, implementation suggestions, and examples of student work. In this talk, we will highlight the IOLA least squares task sequence, which builds from a travel metaphor towards students’ creation of the normal equation and understanding of why it is relevant in solving the least squares problem.
Speaker: Megan Wawro (Virginia Tech)
-
28
-
Sparse Tensor Computations: Algorithms and Applications: A Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
31
Optimal tensor network structure search
Tensor networks provide a powerful framework for compressing multi-dimensional data. The optimal tensor network structure for a given data tensor depends on both data characteristics and specific optimality criteria, making tensor network structure search a difficult problem. Existing solutions typically rely on sampling and compressing numerous candidate structures; these procedures are computationally expensive and therefore limiting for practical applications. We address this challenge by viewing tensor network structure search as a program synthesis problem and introducing an efficient constraint-based assessment method that avoids costly tensor decomposition. Specifically, we establish a correspondence between transformation programs and network structures. We also design a novel operation named output-directed splits to reduce the search space without hindering expressiveness. We then propose a synthesis algorithm to identify promising network candidates through constraint solving, and avoid tensor decomposition for all but the most promising candidates. Finally, we extend this approach to work both with cross approximation and for arbitrary reshaping and re-orderings of nodes Experimental results show that our approach improves search speed by up to 10× and achieves compression ratios 1.5× to 3× better than state-of-the-art. Notably, our approach scales to larger tensors that are unattainable by prior work. Furthermore, the discovered topologies generalize well to similar data, yielding compression ratios up to 2.4× better than a generic structure while the runtime remains around 110 seconds.
Speaker: Alex Gorodetsky (University of Michigan) -
32
Decomposition of a tensor into multilinear rank-$(M_r,N_r,\cdot)$ terms
We present new generic and deterministic uniqueness results for block term decompositions (BTD). These uniqueness conditions hold under mild assumptions and apply to more general settings than previously known results. We also present an algebraic algorithm for the computation of BTDs. Our algorithm requires no knowledge of the block sizes appearing in the BTD: these block sizes are recovered from the algorithm. Through numerical simulations, we illustrate that, in contrast to competing optimization-based methods, even in noisy settings our algebraic algorithm can successfully recover an underlying BTD without knowledge of block sizes provided the signal-to-noise ratio is sufficiently high. We observe that the algorithm can significantly improve one's ability to successfully recover a BTD when it is used as an algebraic initialization for leading optimization routines. Moreover, only a few optimization iterations are required to successfully converge to the BTD from the algebraic solution.
Speaker: Nico Vervliet (KU Leuven) -
33
Synchronous and Asynchronous Parallelism Approaches for Generalized Canonical Polyadic Tensor Decomposition with GenTen
The Canonical Polyadic (CP) tensor decomposition is a well-known method for interpretable analysis of high-dimensional data. Recently, the Generalized CP (GCP) method was introduced by Hong, Kolda and Duersch (2020) to allow for flexible choice of the loss function in the optimization problem defining the CP model, enabling more interpretable decompositions of strongly non-Gaussian data such as count or binary data. Furthermore, Kolda and Hong (2020) introduced a version of GCP that leverages randomization and stochastic optimization to address scalability to large, sparse data sets. In this work, we take these ideas a step further and consider synchronous and asynchronous algorithms for parallel GCP tensor decomposition through the GenTen software package, exploiting both shared and distributed memory parallelism. We build on shared memory parallel CP decomposition algorithms utilizing Kokkos for portability across CPU and GPU architectures (Phipps and Kolda, 2019) to support the random sampling and stochastic optimization methods required by GCP. We then couple this approach to the well-known medium-grained distributed memory parallelism scheme (Smith and Karypis, 2016) developed for traditional CP decompositions through MPI, providing a synchronous, hybrid MPI+Kokkos, parallel GCP decomposition capability. Finally, we propose an asynchronous distributed parallelism approach building on related techniques for federated learning to achieve even better scalability to large data sets. We study the effectiveness of the proposed synchronous and asynchronous approaches vis-à-vis computational cost and accuracy on synthetic and publicly available real-world datasets of varying sizes, dimensions, and sparsity patterns using several loss functions.
Speaker: Eric Phipps (Sandia National Laboratories)
-
31
-
Quantum Numerical Linear Algebra: A Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
34
Computational linear algebra from time evolution and noisy inner products
The fundamental building blocks of iterative linear algebra algorithms in ordinary digital computation are matrix–vector multiplications and inner products. In quantum computing, we lose easy access to both primitives. But we also gain replacements. Instead of matrix–vector products, we can apply the unitary time evolution operator $e^{-itA}$, and we have access to noisy—but statistically unbiased—measurements of inner products. How can we solve linear algebra problems with these primitives? What is the greatest possible accuracy achievable? This talk will address these questions with a focus on eigenvalue problems and on the presenter's own research. No prior knowledge of quantum computing is expected.
Speaker: Dr Ethan Epperly (UC Berkeley) -
35
Quantum Linear Algebra: from Optimization to Differential Equation
Quantum computing relies heavily on the efficient manipulation of linear algebraic structures. This talk discusses the application of quantum linear algebra across two major domains: continuous optimization and differential equations. We demonstrate how quantum linear algebra can be utilized to solve these problems efficiently, discussing both the algorithmic construction and the theoretical analysis required to establish quantum advantage.
Speaker: Zeguan Wu (University of Pittsburgh) -
36
Quantum Linear Algebra for Optmization
Quantum linear algebra has emerged as a promising framework for accelerating the solution of fundamental computational problems, including systems of linear equations—a core subroutine in many scientific and engineering tasks. These problems arise prominently in optimization algorithms. In this talk, we discuss the opportunities and challenges associated with integrating quantum linear algebra techniques into modern optimization methods. We focus on two key applications: quantum interior point methods for conic optimization and a quantum adjoint method for PDE-constrained optimization. Together, these illustrate how quantum algorithms may reshape large-scale optimization by offering new pathways toward improved scalability and performance.
Speaker: Mohammadhossein Mohammadisiahroudi (Department of Mathematics and Statistics, University of Maryland Baltimore County)
-
34
-
Model- and Data-driven Reduced-order Models and Their Applications in Inverse Problems: A Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
37
Reduced order models for inverse scattering
We present a numerical method for solving an inverse boundary value problem of estimating the acoustic velocity in the Helmholtz equation from frequency domain measurements based on reduced order models (ROM). The ROM is the Galerkin projection of the Helmholtz operator onto a subspace spanned by its solution snapshots at certain wavenumbers. We show how to reconstruct the ROM in a data-driven way. Once the ROM is reconstructed, the acoustic velocity can be estimated using non-linear optimization that minimizes a misfit based on the ROM. Such an approach typically outperforms the conventional methods based on data misfit minimization.
Speaker: Andreas Tataris -
38
ROM-based Inverse Scattering for Monostatic Data
We consider an inverse scattering problem for monostatic synthetic aperture radar (SAR), where the goal is to estimate the wave speed in a heterogeneous, isotropic medium using measurements from a moving antenna. The forward map, derived from Maxwell’s equations, is inherently nonlinear, accounts for multiple scattering, and exhibits high-frequency oscillations that make traditional nonlinear least-squares formulations prone to local minima.
We present an alternative two-step approach centered on the construction of data-driven Reduced Order Models (ROMs). The first step computes a non-iterative map from the measurements to an approximation of the internal electric field. This internal wave is designed to fit the data by construction. The second step utilizes an iterative optimization to minimize the discrepancy between this internal field and the solution to the governing Maxwell's equations across all antenna locations.
While the ROM-based approach provides a superior estimate of the target support compared to standard SAR imaging, its performance is sensitive to the algebraic properties of the data-driven operators. In this talk, we focus on the linear algebraic challenges involved in stabilizing these ROMs. Specifically, we discuss the spectral properties of the data-driven mass and stiffness matrices and review regularization techniques to maintain robustness when the monostatic data is polluted with noise and preserve the structure of the ROM. Numerical simulations demonstrate that these stabilized ROMs allow for accurate quantitative estimation of the wave speed in regimes where standard inversion methods fail.
Speaker: Jörn Zimmerling (Uppsala University)
-
37
-
12:15 PM
Lunch Break
-
Contributed Talks: B1 Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
39
Graph Energy Change Under Edge Deletion
The energy $\mathcal{E}(G)$ of a graph $G$ is the sum of the absolute values of all the eigenvalues of its adjacency matrix. Gutman (2001) proposed a problem of characterizing the graph $G$ and the edge $e$ of $G$ such that $\mathcal{E}(G-e) \le \mathcal{E}(G)$. Later, Day and So (2008) gave a sufficient condition : $e$ is a cut-edge of $G$.Recently, Tang et al. (2023) gave another sufficient condition for $\mathcal{E}(G-e) \le \mathcal{E}(G)$, where $e$ is not a cut-edge.In this talk, we improve their result by giving a weaker sufficient condition and simplifying their proof. As a consequence, we prove that adding edges to a complete $k$-partite graph leads to higher energy than the energy of the original graph.
Speaker: Wasin So (San Jose State University) -
40
On some Graphs determined by their Signless Laplacian spectrum
Abstract: In this paper, we investigate when the disjoint union of complete graphs $K_a \cup K_b$ is determined by its signless Laplacian spectrum (Q-DS). We first prove that $K_a \cup K_b$ is Q-DS among disconnected graphs. We then show that no connected signless Laplacian co-spectral mate of $K_a \cup K_b$ exists on at most ten vertices. Further, we establish that $K_t \cup K_a$ is Q-DS for $t=1,2$, except in the case $t=1$ and $a=3$. We also prove that $K_a \cup K_{a+1}$ is Q-DS. Finally, we show that the complete bipartite graph $K_{a,a+1}$ is determined by its signless Laplacian spectrum.
Keywords: Complete graph; Co-spectral graph; Signless Laplacian
matrix.
AMS subject classifications: 05C50, 15A18.Speaker: Jitul Talukdar (IIT Kharagpur) -
41
Row and Column Equivalence Transversality Properties: Extensions of the Rank-Preserving Transversality Property
This 20-minute contributed talk focuses on two properties introduced in recent work on matrix manifolds: the Row Equivalence Transversality Property (RETP) and the Column Equivalence Transversality Property (CETP).
For an $ m \times n $ real matrix $ A $, RETP holds if the manifolds $ \{GA : G \in \mathrm{GL}(m, \mathbb{R})\} $ and $ Q(\mathrm{sgn}(A)) $ intersect transversally at $ A $, with tangent space $ \{YA : Y \in \mathbb{R}^{m \times m}\} $. Similarly, CETP involves $ \{AG : G \in \mathrm{GL}(n, \mathbb{R})\} $ with tangent space $ \{AX : X \in \mathbb{R}^{n \times n}\} $.
We establish that RETP and CETP each imply RPTP, but the converse does not hold (e.g., certain block matrices have RPTP without RETP/CETP). Key results include:Theorem 4.1: $ A $ has RETP iff for each row with zeros, the corresponding columns are linearly independent.
Theorem 4.2: Dual for CETP with rows.
Dimension bounds: If RETP, then $ m \cdot \mathrm{rank}(A) + \#(A) \geq mn $; similar for CETP.
Invariance under permutation and diagonal equivalence.
Theorem 4.4: RPTP with zero row/column implies full rank and RETP/CETP.
Corollaries for partitioning matrices to verify RPTP via RETP/CETP (e.g., Theorem 4.8, Corollary 4.9).
Examples: Matrices with RPTP but not RETP/CETP; conditions for both.These properties aid in studying minimum ranks of sign and zero-nonzero patterns, with connections to transverse intersections in smooth manifolds. No prior RPTP knowledge assumed; proofs and examples provided.
Speaker: Joonwon Seo (Georgia State University)
-
39
-
Contributed Talks: B2 Torgersen 1020, Virginia Tech
Torgersen 1020, Virginia Tech
-
42
Superfast 1-Norm Estimation
A matrix algorithm is said to be superfast (aka runs at sublinear cost) if it involves much fewer scalars and flops than an input matrix has entries. Such algorithms have been extensively studied and widely applied in modern computations for matrices with low displacement rank and more recently for low rank approximation of matrices, even though any superfast algorithm fails on worst case inputs for the latter problem. We extend this study to a new area by proposing three distinct superfast 1-norm estimators. In our extensive tests with synthetic and real word matrices all three have output 1-norm within relative error 1.5, while the outputs of two of them were consistently as accurate as LAPACK's. We point out some promising extensions of our surprisingly simple techniques. With further testing and refinement our algorithms should eventually win user's attention.
Speaker: Victor Pan (City University of New York) -
43
Superfast Low Rank Approximation
Low Rank Approximation (LRA) of a matrix are invaluable for Numerical Linear Algebra and Data Science. Some recent papers propose superfast algorithms that output LRAs with near-optimal accuracy for a large class of inputs but, as ANY superfast LRA algorithm, fail on a large class of inputs as well. To narrow the latter class we first superfast compute a crude initial LRA by applying one of these or another superfast algorithm (we specify a novel class of such algorithms). Then we recursively refine that LRA superfast by extending iterative refinement algorithms for the solution of the systems of linear and nonlinear equations and by applying the recent techniques of oversampling and compression of J. A. Tropp, A. Yurtsever, M. Udell, V. Cevher, Streaming Low-Rank Matrix Approximation with an Application to Scientific Simulation, SIAM J. on Scientific Computing, 41, pp. A2430–A2463, 2019; this enables us to control rank growth in the refinement. We analyze our algorithms and test them numerically.
Speaker: Victor Pan (City University of New York)
-
42
-
Numerical Linear Algebra in Machine Learning: B McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
44
Universal Kronecker Core Factorization of the NTK: Quantifying Implicit Bias of Gradient Descent
How are learned representations incrementally formed to solve tasks by Gradient Descent (GD)? In this talk, we will show that each step of GD is exactly given by the application of a massive tensor-valued linear operator, which we call the Configuration Space Neural Tangent Kernel (NTK). We prove that it can be decomposed into two operators: P and K, the former capturing state-to-state dynamical dependencies and the latter capturing immediate parameter-to-state dependencies. Based on this, we prove a universal result stating that any weight-based model in deep-learning or optimal control can be factored so that K is a simpler Kronecker product matrix (dubbed the Kronecker core factorization). Importantly, the key ingredients in this core are already computed during inference, hence they can be used to immediately determine which tasks are more easily learnable by GD for a particular model. We show that this structure implies a bottlenecking of GD dynamics, leading to low-rank dynamical corrections and implicit bias towards particular tasks. We specialize our analysis to recurrent models (e.g., neural ODEs or RNNs), showing how the factorization constrains the temporal modes of learning. In addition to these theoretical results, we develop a package (kpflow) using matrix-free numerical linear algebra fast analysis of the NTK gradient operator for any model. This package makes it easier to work with linear operators on tensor domains with fast, randomized analysis tools.
Speaker: James Hazelden (University of Washington) -
45
Principal Surjective Flows: Relaxing Bijection Assumption via the Smooth Co-Area Formula and Gram Determinants
Normalizing Flows (NFs) enable tractable density evaluation by modelling data through invertible neural transformations. However, this reliance on global bijectivity severely restricts their expressiveness when the target distribution lies on a low-dimensional manifold or exhibits complex topology. To overcome this limitation, we introduce Principal Surjective Flows (PSFs), a framework that replaces invertibility with carefully constructed surjective mappings. Leveraging the Smooth Co-area Formula, we derive a principled likelihood expression in which the density transformation is governed by the normal Jacobian, which is equivalently the square root of a Gram determinant formed from the Jacobian’s nonzero singular values. This reformulation enables stable and geometrically meaningful density computation for dimension-reducing generative maps. Through Gaussian examples, we show that integrating densities over fibres reduces to evaluating Gaussians on affine subspaces determined by the row space and nullspace of the map, revealing a direct connection between surjective flows and core numerical linear algebra concepts including low-rank operators, projection matrices, and conditioning. Overall, PSFs provide a mathematically grounded and computationally efficient extension of normalizing flows, broadening their applicability while preserving exact and tractable density estimation.
Speaker: Haoran Ni (University of Warwick) -
46
Reduced- and Mixed-Precision Algorithms for QR Decomposition
As machine learning and AI continue to shape modern hardware design, reduced-precision arithmetic has become essential in high-performance computing. Recent advances in hardware architectures—such as AI accelerators, GPUs, and tensor-core technologies—many of which are driven by machine-learning workloads, are optimized for low-precision operations to improve performance and reduce energy consumption in computationally intensive scientific applications. However, the primary drawback of uniformly low-precision methods remains their potential numerical instability, particularly when solving linear systems. As a result, mixed-precision strategies have become widely used in numerical linear and multilinear algebra to balance performance, energy efficiency, and accuracy while still exploiting ML-optimized hardware.
QR factorizations are among the core algorithms in numerical linear algebra and are also playing an increasing role in machine-learning applications such as randomized embeddings, dimensionality reduction, and least-squares problems. However, their computational cost remains significant for large-scale matrices. In this talk, we first present a uniform half-precision Householder QR decomposition based on the WY representation and examine both its potential performance advantages on modern AI hardware and its significant limitations. We then show how increasing the precision of key computations can substantially improve accuracy and how these choices interact with ML-optimized hardware. Finally, we discuss how reduced- and mixed-precision strategies can be incorporated into other QR decomposition algorithms, including shiftedCholeskyQR3 and TSQR, to further leverage modern AI hardware and software ecosystems.
Speaker: Eda Oktay (Max Planck Institute for Dynamics of Complex Technical Systems Magdeburg)
-
44
-
Matrix Nearness Problems: B Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
47
Reliably Computing the Worst-case H-infinity Norm of a Parametric System Using an Interpolation-based Algorithm.
In this talk, we consider computing the worst-case (highest) $\mathcal{H}_\infty$ norm of a either a continuous-time or discrete-time linear time-invariant parametric system, where the state-space matrices all depend on a single real-valued scalar parameter $\mathsf{p}$ on a domain $\mathcal{P}$ consisting of a finite number of intervals. On each interval in $\mathcal{P}$, we assume that $\mathsf{p}$ may vary nonlinearly and that each state-space matrix is differentiable with respect to $\mathsf{p}$ almost everywhere, but we do permit that the parameter may be discontinuous at a finite number of points inside each interval of $\mathcal{P}$. Using the framework of interpolation-based globality certificates, which were first devised as a fast way to compute Kreiss constants and the distance to uncontrollability, we present a new optimization-with-restarts algorithm that, under reasonable assumptions, solves the underlying nonconvex optimization with global convergence and thus computes the worst-case $\mathcal{H}_\infty$ norm to arbitrary accuracy. Our algorithm does $\mathcal{O}(kn^3)$ work, where $n$ is the dimension of the state vector and $k$ is the total number of function evaluations needed to globally approximate our certificate functions over the parameter domain $\mathcal{P}$ using, say, Chebfun. In practice, $k$ is not strongly correlated with $n$ and typically only a handful of optimization restarts are needed before a global optimizer attaining the value of the worst-case $\mathcal{H}_\infty$ norm is found. Moreover, the overall cost of our algorithm is almost entirely due to the cost of approximating the final certificate function, which asserts that the final maximizer found is in indeed a global maximizer. Experiments show that our new algorithm is both significantly faster and more reliable than other reasonable approaches for computing the worst-case $\mathcal{H}_\infty$ norm.
Speaker: Tim Mitchell (CUNY Queens College & The Graduate Center) -
48
Distance to prescribed rank matrix polynomials via generic factorizations and alternating least squares
We propose an algorithm that approximates a given matrix polynomial of any degree $d$ by another matrix polynomial of a prescribed rank and degree at most $d$. The algorithm combines recent advances in the theory of generic factorizations for matrix polynomials of bounded rank and degree with an alternating least squares strategy. For $d=1$, the algorithm includes the important case of matrix pencils. The algorithm solves, as a particular case, the well-known problem of computing the distance to singularity of a regular matrix polynomial. We present numerical experiments for testing the proposed algorithm and for comparison to the previously known ones.
Speaker: Froilán M. Dopico (Universidad Carlos III de Madrid (Spain)) -
49
Structured stability radii of dissipative Hamiltonian systems
We study linear time-invariant Dissipative Hamiltonian (DH) systems arising in energy-based modeling of dynamical systems. An advantage of DH systems is that they are always stable due to the structure of their coefficient matrices, and, under further weak conditions, even asymptotically stable. Here, we consider the computation of the stability radii for a given asymptotically stable DH system; i.e., the smallest structured perturbation that puts a DH system on the boundary of the region of asymptotic stability, so that it has purely imaginary eigenvalues. We obtain explicit computable formulas for various structured stability radii. For this, the problem of computing stability radii is reformulated in terms of minimizing the Rayleigh quotient of a Hermitian matrix or the sum of two generalized Rayleigh quotients of Hermitian semidefinite matrices. This reformulation results in the problem of minimizing the largest eigenvalue of an eigenvector-dependent Hermitian matrix or minimizing the smallest eigenvalue of a Hermitian matrix which depends on the eigenvector. We then, demonstrate (via numerical experiments) that, under structure-preserving perturbations, the asymptotic stability of a DH system is much more robust than under general perturbations, since the distance to instability is typically much larger when structure-preserving perturbations are considered.
Speaker: Anshul Prajapati (Max Planck Institute for Dynamics of Complex Technical Systems)
-
47
-
Matrix Inequalities, Matrix Equations, and Their Applications: B Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
50
Random Matrix Ensembles with Split Limiting Behavior
Random matrix theory has successfully modeled a variety of systems, from energy levels of heavy nuclei to zeros of the Riemann zeta function. One of the central results is Wigner's semi-circle law: the distribution of normalized eigenvalues for ensembles of real symmetric matrices converge to the semi-circle density (in some sense) as the matrix size tends to infinity. We introduce a new family of $N\times N$ random real symmetric matrix ensembles, the $k$-checkerboard matrices, whose limiting spectral measure has two components. All but $k$ eigenvalues are in the bulk, and their behavior, appropriately normalized, converges to the semi-circle as $N\to\infty$; the remaining $k$ are tightly constrained near $N/k$ and their distribution converges to the $k \times k$ hollow GOE ensemble (this is the density arising by modifying the GOE ensemble by forcing all entries on the main diagonal to be zero). Similar results hold for complex and quaternionic analogues. We are able to isolate each regime separately through appropriate choices of weight functions for theeigenvalues and then an analysis of the resulting combinatorics. This is joint work with Paula Burkhardt, Peter Cohen, Jonathan Dewitt, Max Hlavacek, Carsten Sprunger, Yen Nhi Truong Vu, Roger Van Peski, and Kevin Yang.
Speaker: Steven Miller (Williams College) -
51
On a characterization of the equality case in the generalized Böttcher-Wenzel inequality
The main purpose of this presentation is to illustrate an equivalent condition for the equality case in the generalized Böttcher-Wenzel (BW) inequality for three matrices.
Let $A, B,$ and $C$ be square matrices with complex elements. The BW inequality is an upper bound estimate on the Frobenius norm of the commutator of $A$ and $B$, defined as $AB-BA$. After the BW inequality was proved in 2008, various subsequent problems related to it have been considered until now. An necessary and sufficient condition for the equality case in the inequality is one of them. Also, there are some papers which investigate relationships between the BW inequality and the uncertainty principle in quantum physics. Recently, the presenter estimated the Frobenius norm of the generalized commutator of $A, B$, and $C$, defined as $ABC-CBA$, and obtained an inequality, which is a generalization of the BW inequality [M. Nobori, A generalization of the Böttcher-Wenzel inequality for three rectangular matrices, Linear Algebra Appl. 725 (2025) 135–144]. In this talk, we shall deduce an equivalent condition for the equality case in the generalized BW inequality. The core idea of the derivation is a necessary and sufficient condition for the Weyl inequality.Speaker: Motoyuki Nobori (Graduate School of Science and Engineering, Ehime University) -
52
Normal Matrices
Normal matrices form a central class in matrix analysis, including Hermitian, skew-Hermitian, and unitary, positive semidefinite, permutation matrices and so on. This presentation surveys fundamental properties of normal matrices, including spectral characterization, unitary diagonalization, and trace (in)equality through majorization. It highlights equivalent conditions for normality, with discussions extending to matrix exponentials and polynomials. Examples and counterexamples are provided to clarify certain subtle points about matrix normality.
Speaker: Fuzhen Zhang (Nova Southeastern University)
-
50
-
Where Algebraic Coding Theory and Graph Theory Meet: B Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
53
Graph-based error correction code constructions made practical by modern decoder developments
The study of error correcting codes has two important facets: code construction and decoder design. Many graph-based code constructions have been established to have desirable theoretical properties, but, heretofore, have not been practically decodable. In this talk, we explain recent developments in code-agnostic decoders, including soft-input soft-output variants, that offer a way forward in the provision of a practical decoder for such codes.
Speaker: Prof. Ken Duffy (Northeastern University) -
54
On the Hamming Weight Distribution of Cyclic Codes with Arbitrarily Many Nonzeroes
Determining the minimum Hamming distance of an error-correcting code $\mathcal{C}$ has long stood as a fundamental challenge in coding theory. In this talk, we turn our attention to an even more ambitious problem: determining the full Hamming weight distribution of $\mathcal{C}$. This means counting the number of codewords in $\mathcal{C}$ of each possible Hamming weight—an essential but notoriously difficult task, especially for general codes.
While computing the weight distribution for an arbitrary code remains largely out of reach, the problem becomes more tractable under additional structural assumptions, especially when the code $\mathcal{C}$ is cyclic. We will discuss the calculation of three families of cyclic codes with arbitrarily many nonzeroes, highlighting its connections to linear algebra, graph theory, and association schemes.
The talk is based on joint work with Gennian Ge (Capital Normal University), Sihuang Hu (Shandong University), Tao Feng (Zhejiang University); as well as Maosheng Xiong (Hong Kong University of Science and Technology) and Haode Yan (Harbin Institute of Technology, Shenzhen).
Speaker: Shuxing Li (University of Delaware) -
55
Hierarchical quasi-cyclic codes from polynomial evaluation codes
In this talk, we present a family of algebraically constructed hierarchical quasi-cyclic codes. These codes are built from Reed-Solomon and polynomial evaluation codes using a construction of superimposed codes by Kautz and Singleton. Using a novel ordering of the codewords and evaluation points, we show both the number of levels in the hierarchy and the index of these $q$-ary-derived codes are determined by the field size. We compute explicit code parameters and properties as well as some additional bounds on parameters such as rank and distance. In particular, starting with Reed-Solomon codes of dimension two yields hierarchical quasi-cyclic codes with Tanner graphs of girth 6. Finally, we present a table of small code parameters and note that some of these codes meet the best known minimum distance for binary codes, with the additional hierarchical quasi-cyclic structure.
Speaker: Dr Kathryn Haymaker (Villanova University)
-
53
-
Krylov Iterative Methods for Linear Equations: B McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
56
Polynomial Preconditioning for Indefinite Matrices
Preconditioning plays a central role in accelerating the convergence of iterative methods for solving large linear systems. Among the various approaches, polynomial preconditioning offers a flexible approach. Krylov subspace methods provide a natural setting for constructing polynomials that can be used as preconditioners. In this work, we investigate the use of polynomial preconditioning using the GMRES polynomial on indefinite linear systems. Several challenges can arise in polynomial preconditioning for indefinite matrices. These are mentioned along with some algorithmic strategies designed to mitigate their effects and enhance robustness.
Speaker: Hayden Henson (Baylor University) -
57
Polynomial Preconditioned Golub-Kahan for Finding Singular Values and Solving Least Squares
Golub-Kahan bidiagonalization procedure is well known for its role in computing singular values and solving least squares problems. We present a polynomial-preconditioned variant of this framework aimed at reducing the need for restarting and extensive orthogonalization. This talk outlines the formulation and motivation for the method and examines its potential to improve performance in practical settings.
Speaker: Kingsley Michael (Baylor University) -
58
Preconditioner Updating with Lasso-based Sparse Approximate Maps
Design problems such as topology optimization and PDE-based inverse problems require the solution of a sequence of linear-systems derived from finite element or finite difference discretization. Preconditioning is essential for the fast solution of these systems by iterative methods. However, computing an accurate preconditioner for every system in the sequence may be a computational bottleneck, warranting a cheap preconditioner-updating scheme. The sparse approximate map (SAM), introduced in Preconditioning Parametrized Linear Systems by Carr et al. (SISC 2021), computes a sparse approximation of the exact map between matrices in the sequence. Each column of the SAM is computed by solving a small least-squares problem. The resulting map can be leveraged to update a preconditioner computed earlier in the sequence. This method is cheap as long as the underlying sparsity pattern has few non-zeros. However, as revealed in prior research, often there is no intuitive best choice for such a pattern. Often the sparsity pattern of the matrix (or powers of it) can be used, but for topology optimization matrices or for high-order finite element stiffness matrices with a large number of non-zeros per column this is a poor choice. In this talk, we show how generating patterns using an L-1 constrained least squares solve (lasso) can be very effective. We solve the lasso problem to modest accuracy using the alternating direction method of multipliers (ADMM). We demonstrate that the parameters of ADMM and the regularization parameter can be heuristically selected to compute matrix maps that are significantly sparse and sufficiently accurate. This is joint work with Johann Rudi and Eric de Sturler.
Speaker: Achintya Sunil
-
56
-
Algebraic Invariants of Graphs: B Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
59
Generalized chip firing and critical groups of arithmetical structures on trees
Chip firing provides a way to study the sandpile group (also known as the Jacobian) of a graph. We use a generalized version of chip firing to bound the number of invariant factors of the critical group of an arithmetical structure on a graph. We also show that, under suitable hypotheses, critical groups are additive under wedge sums of graphs with arithmetical structures. These results allow us to relate the number of invariant factors of critical groups associated to any given tree to decompositions of the tree into simpler trees. We use this to classify those trees for which every arithmetical structure has cyclic critical group. Finally, we show how to construct arithmetical structures on trees with prescribed critical groups. In particular, every finite abelian group is realized as the critical group of some arithmetical structure on a tree.
Speaker: Joel Louwsma (Niagara University) -
60
The Smith normal form of distance matrices of high dimensional trees
Graham, Lovász and Pollak obtained a well known formula for the determinant of distance matrices of trees. This formula depends only on the number of vertices of the tree and not on its topological structure. Later, Hou and Woo computed the explicit expression of the Smith form of the distance matrix of a tree, which again, it depends only on the number of vertices. In this talk, we will show results that extend Graham-Lovász-Pollak and Hou-Woo formulas to higher dimensions by computing the Smith normal form and the determinant of a distance matrix associated to high dimensional trees.
Speaker: Dr Carlos Alfaro (Banxico) -
61
From Chip Firing to Determinantal Ideals and Back
Chip-firing is a discrete dynamical process on a graph that exhibits striking phenomena, including fractal-like symmetries and self-organized criticality. From an algebraic perspective, the states of this process (combinatorially) define a group, the Sandpile group, whose algebraic structure is given by the Smith normal form of the Laplacian matrix. We will discuss problems related to Sandpile Groups and introduce some determinantal ideals for graphs. We will discuss how these algebraic invariants provide a framework for solving said problems in graph theory and, conversely, how the structure of Sandpile groups offers new insights into the properties of these ideals.
Speaker: Ralihe Raul Villagran Olivas (Worcester Polytechnic Institute)
-
59
-
Low-Complexity Data-driven or Classical Algorithms and Applications: B Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
62
Optimal Experimental Design for Gaussian Processes via Column Subset Selection
Gaussian process regression uses data measured at sensor locations to reconstruct a spatially dependent function with quantified uncertainty. However, if only a limited number of sensors can be deployed, it is important to determine how to optimally place the sensors to minimize a measure of the uncertainty in the reconstruction. We consider the Bayesian D-optimal criterion to determine the optimal sensor locations by choosing $k$ sensors from a candidate set of $n$ sensors. Since this is an NP-hard problem, our approach models sensor placement as a column subset selection problem (CSSP) on the covariance matrix, computed using the kernel function on the candidate sensor points. We propose an algorithm that uses the Golub-Klema-Stewart framework (GKS) to select sensors and provide an analysis of lower bounds on the D-optimality of these sensor placements. To reduce the computational cost in the GKS step, we propose and analyze algorithms for the D-optimal sensor placements using Nyström approximations on the covariance matrix. Moreover, we propose several algorithms that select sensors via Nyström approximation of the covariance matrix, utilizing the randomized Nyström approximation, random pivoted Cholesky and greedy pivoted Cholesky. We demonstrate the performance of our method on two applications: thin liquid film dynamics and sea surface temperature.
Speaker: Jessie Chen (North Carolina State University) -
63
Accelerating Spectral Clustering of Time Series by approximating the Similarity Matrix using Randomly Pivoted Cholesky
A major problem for time series clustering is that computing the similarity matrix for the most used similarity measures becomes infeasible if number amount or length of time series becomes too large. However, since this similarity matrix typically has low-rank structure, it can be approximated using a low-rank approximation. In this work, we show that existing numerical linear algebra methods, more specifically Randomly Pivoted Cholesky can be used in the context of time series clustering to drastically reduce the computational cost of calculating the similarity matrix, while maintaining the clustering quality. This shows that low-rank approximation algorithms are an effective and scalable technique that can be used in time series clustering.
Speaker: Raf Vandebril (Dept. Computer Science, KU Leuven) -
64
Spotlight inversion by orthogonal projections
Many computational problems involve solving a linear system of equations, although only a subset of the entries of the solution are needed. In inverse problems, where the goal is to estimate unknown parameters from indirect noisy observations, it is not uncommon that the forward model linking the observed variables to the unknowns depends on variables that are not of primary interest, often referred to as nuisance parameters. In this talk, we consider linear problems, and propose a novel projection technique to eliminate, or at least mitigate, the contribution of the nuisance parameters in the model. We refer to this approach as spotlight inversion, as it allows to focus on only the portion of primary interest of the unknown parameter vector, leaving the uninteresting part in the shadow. The viability of the approach is illustrated with two computed examples.
Speaker: Daniela Calvetti (Case Western Reserve University)
-
62
-
Approximate Computing in Numerical Linear Algebra: B Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
65
Adaptive Mixed Precision Hierarchical Matrices
Hierarchical matrices or $\mathcal{H}$-matrices are the block low-rank representation of the original matrices and are widely used in fast matrix computations. In this talk, we show that the low-rank blocks of $\mathcal{H}$-matrices can be represented in low precision (precision lower than the working precision) without degrading the overall approximation quality. We provide an explicit rule to decide which precision should be chosen for a particular low-rank block. We propose an adaptive mixed precision algorithm for constructing and storing $\mathcal{H}$-matrices. We also show that use of mixed precision does not compromise the numerical stability and accuracy of the $\mathcal{H}$-matrix-vector product. We perform a wide range of numerical experiments to validate our theoretical results. Our numerical results illustrate that adaptive mixed precision $\mathcal{H}$-matrices can achieve significant storage reductions compared to uniform precision $\mathcal{H}$-matrices, without compromising accuracy.
Speaker: Dr Ritesh Khan -
66
Automated Precision Tuning for Numerical Algorithms
Achieving optimal performance in numerical computations often hinges on aggressively reducing precision or performing rigorous rounding-error analysis to retain numerical accuracy. The precision tuning methods provide a unified, task-specific validation platform for automated precision tuning, enabling a balance between computational efficiency and numerical fidelity. In this talk, we present novel precision-tuning methods, including expanded precision options and low-precision math functions, which streamline the exploration of low-precision configurations. This talk reveals the potential for significant performance gains by leveraging low-precision settings. This work underscores the transformative potential of automated precision tuning while maintaining accuracy and performance objectives.
Speaker: Xinye Chen (LIP6, CNRS, Sorbonne University) -
67
Iterative Refinement for a Subset of Eigenpairs of a Real Symmetric Matrix and Its Convergence Analysis
We consider the eigenvalue problem $Ax^{(i)} = \lambda_i x^{(i)}$ for a real symmetric matrix $A \in \mathbb{R}^{n \times n}$, where $\lambda_i \in \mathbb{R}$ is an eigenvalue of $A$ and $x^{(i)} \in \mathbb{R}^n$ is the corresponding eigenvector. This work investigates iterative refinement methods to improve the accuracy of eigenvectors $x^{(i)}$.
Efficient methods are known for improving the accuracy of either a single approximate eigenvector or all approximate eigenvectors simultaneously. In practical applications, however, it is often necessary to improve only a subset of eigenvectors: $X = \left( x^{(p_1)}, x^{(p_2)}, \dots, x^{(p_k)} \right) \in \mathbb{R}^{n \times k}, \{p_i\} = \{1, \dots, n\}, 1 \le k \le n$.
The proposed method improves the accuracy of an approximation $\widehat{X}$ to $X$. Given the matrix $A$ and an approximate eigenvector matrix $\widehat{X}$, the goal is to compute $\widetilde{X}$ such that
$\|X - \widetilde{X}\| < \|X - \widehat{X}\|$. A necessary condition for improving accuracy is $\min_{1 \le i \le k} |\lambda_{p_i}| > \max_{k+1 \le j \le n} |\lambda_{p_j}|$.The proposed method has two main advantages: its computational cost is dominated by matrix multiplications, and it can handle large sparse matrices, making it applicable to mixed-precision and high-precision numerical computations. We report both analytical results and numerical experiments.
Speaker: Takeshi Terao (Waseda University)
-
65
-
Sparse Tensor Computations: Algorithms and Applications: B Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
68
A Tree Tensor Network Integrator for the Chemical Master Equation
The development of new drugs and therapies increasingly relies on the numerical simulation of the reaction networks inside biological cells. However, the most accurate description of such reaction networks with the chemical master equation (CME) suffers from the curse of dimensionality, meaning that memory and computational cost grow exponentially with the number of dimensions. This renders the simulation of such networks infeasible on currently available computer hardware and it will remain a fundamental limitation in the future.
In this talk, we present a novel numerical solution method for the stochastic description of chemical reaction networks. It is based on the dynamical low-rank approximation with tree tensor networks, which effectively approximates the high-dimensional solution of the CME as a linear combination of low-rank factors.
This low-rank approximation corresponds to a hierarchical separation of the reaction network into smaller partitions. Only reactions that occur between chemical species in different partitions are approximated. The number of low-rank factors, the so-called rank, determines the accuracy of the approximation; typically, 10 to 20 low-rank functions are sufficient for a good solution.
We derive an efficient integration scheme for the low-rank factors that exploits the low-rank tensor structure of the CME. This scheme is based on the projector-splitting integrator for tree tensor networks [Ceruti, Lubich, and Walach, SIAM J. Numer. Anal., vol. 59, 2021]. We apply our method to models from biochemistry and show that the method significantly reduces the memory consumption, while achieving improved computational performance and better runtimes compared to a Monte Carlo method.
Finally, we investigate how the choice of the partitioning affects the required rank. To this end, a heuristic partitioning algorithm is presented that minimizes information entropy and the number of reactions between partitions.
Speaker: Julian Mangott (Universität Innsbruck) -
69
Low-Rank CP Tensor Compression and Its Application to High-Dimensional PDEs
The Canonical Polyadic (CP) decomposition is widely used to represent high-dimensional data in many applications, for example, solving high-dimensional PDEs like kinetic equations. A key challenge in these problems is the efficient estimation and reduction of the CP rank. The CP rank reduction task can be formulated as approximating the Khatri–Rao product of the CP factor matrices with a lower rank. We propose an approach based on the pivoted Cholesky decomposition to construct interpolative decompositions of the Khatri–Rao product. This method can serve as a standalone CP rank reduction technique or be integrated into classical optimization schemes such as CP-ALS. Moreover, the residuals produced at each step of the pivoted Cholesky naturally provide error indicators, enabling effective rank estimation. Preliminary numerical results demonstrate that this method achieves a balance between computational cost and approximation accuracy. Its effectiveness is further validated through applications to the Vlasov–Poisson equation.
Speaker: Zhanrui Zhang (University of Illinois Urbana-Champaign) -
70
Iterative low-rank time integration of the time-dependent Schrödinger equation
Standard numerical methods for solving PDEs typically suffer from the curse of dimensionality: their computational cost scales exponentially with the dimension of the underlying domain, making them impractical even at low resolution. In many cases of interest, however, such limitations can be overcome by appropriately compressed representations of approximate solutions, for example, by low-rank tensor representations.
For time-dependent PDEs, several approaches to low-rank approximation exist that control ranks in different ways. These range from methods that keep ranks fixed, such as dynamical low-rank approximations, which may lead to uncontrolled errors, to methods that approximate standard time-stepping schemes to any desired accuracy but can produce unnecessarily large ranks.
We develop time integration methods in low‑rank tensor representations that adaptively adjust the approximation ranks to meet a prescribed accuracy while simultaneously maintaining control over the ranks of the computed approximations and all intermediate quantities. These ranks are a main determining factor in the computational costs of such methods. Our approach combines an iterative time-stepping scheme with soft thresholding of the iterates. In the matrix case, the proposed strategy yields iterates whose ranks remain comparable to natural benchmark quantities — namely, the best approximation ranks of the sought solutions at the achieved accuracy. In the higher-dimensional tensor case the algorithm can be adapted appropriately, leading to global error and rank bounds that depend only polynomially on the dimension. Numerical experiments illustrate the theory for linear time-dependent Schrödinger equations.
Speaker: Polina Sachsenmaier (RWTH Aachen University)
-
68
-
Quantum Numerical Linear Algebra: B Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
71
Near-term quantum subspace diagonalization
Quantum subspace diagonalization methods have emerged as a promising application of quantum computers. For matrices $A$ of suitable structure, the unitary operation $U = e^{\imath A t}$ can be efficiently approximated on current quantum devices. This has motivated the use of Krylov subspaces
$$ \begin{equation} \mathcal{K}_m=\text{span} \left\{\mathbf{x},U\mathbf{x}, ..., U^{m-1} \mathbf{x}\right\}, \end{equation} $$ whose resulting Rayleigh-Ritz approximation admits rigorous convergence guarantees. In practice, however, one must contend with the noise inherent to quantum computation. In particular, the quantum subroutine for estimating the inner products $\langle U^j \mathbf{x}, U^k \mathbf{x}\rangle$ is both prohibitively noisy and computationally expensive, hindering the ability to obtain accurate Krylov overlap matrices. We present a quantum-subspace protocol that circumvents the need to compute these inner products by exploiting symmetries of the matrix $A$. This approach uses only operations that can be efficiently performed on a quantum computer and that incur relatively small errors. The resulting algorithm retains the convergence guarantees while substantially reducing computational overhead. The method is demonstrated using large-scale tensor network simulations.Speaker: Adam Byrne (Trinity College Dublin & IBM Research) -
72
Quantum Krylov methods with Hamiltonian powers
Quantum Krylov methods are strong candidates for computing ground states on NISQ and MegaQuop computers. While typically implemented with powers of the time evolution unitary $e^{-iH t}$ for a given Hamiltonian $H$, convergence can be markedly faster with powers of the Hamiltonian $H$ itself as in classical methods. We discuss these convergence rates and present several ways to implement Hamiltonian powers on quantum computers, including qubitization and randomized methods, for use in quantum Krylov. We compare the approaches by providing resource estimates for common applications on current and future quantum computers.
Speaker: Ryan LaRose -
73
Quantum Matrix Encodings
The emergence of Quantum Numerical Linear Algebra (qNLA) offers a paradigm shift in solving large-scale linear systems and matrix functions. However, the practical utility of these algorithms, such as the seminal HHL, is fundamentally bottlenecked by the "input problem", namely the efficient representation of classical matrices as quantum circuits.
In this talk, we explore two distinct methodologies for addressing this challenge. First, we present a systematic framework for constructing block encodings from state preparation circuits. This is achieved by introducing efficient basis-transformation protocols and low-overhead, constant-depth Pauli multiplexers. We then establish a bidirectional operational equivalence by demonstrating the converse: converting block encodings back into state preparation circuits using quantum linear algebra operations.
Furthermore, we introduce a structurally-driven approach based on the factorization of matrix state preparation. We prove that when the state preparation algorithm is applied to vectorized classical data, this algorithm naturally decomposes into a standard Norm Loader and a Direction Multiplexer. This observation reveals that a well-known approach for creating block encodings was effectively "hidden" within the classical logic of state preparation. By leveraging this hidden structure, we present a new method for block encoding that offers unique advantages in qubit utilization and circuit depth.
Speaker: Liron Mor Yosef (Tel Aviv University)
-
71
-
3:15 PM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Contributed Talks: C Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
74
Convergence of flexible GMRES with and without randomized sketching
In this talk, we consider the theoretical convergence of flexible GMRES. While convergence of standard GMRES is well studied, there exist few results of similar nature for flexible GMRES. The aim of this talk is to discuss and fill in this gap. In addition, we report on experiences of using these ideas in the context of randomized sketching.
Speaker: Lauri Nyman (University of Manchester) -
75
New Insights into the Equivalence of Thick and Implicit Restarting Lanczos
Given a symmetric tridiagonal matrix, it has been well--established, that in exact arithmetic, applying some of its eigenvalues as shifts via the $QR$ strategy produces a particular structured matrix where the leading tridiagonal block contains the non--shifted eigenvalues and a trailing diagonal submatrix of the shifted eigenvalues. We will show that the leading tridiagonal block can be created, up to sign differences, by an orthogonal similarity reduction of a certain symmetric arrowhead matrix constructed directly from the non--shifted eigenvalues and associated eigenvectors. This also provides leverage to show a similar result with an upper bidiagonal matrix and a certain triangular arrowhead matrix. Both of the results set the foundation to show, in a unique way, that the well--known implicitly restarted Lanczos method is mathematically equivalent to thick-restarted Lanczos method with a similar extension to singular value problems using the Golub-Kahan Bidiagonalization process.
Speaker: Kyle Monette (University of Rhode Island) -
76
Domain decomposition method for eigenproblems
I discuss solving large-scale symmetric and positive definite eigenproblems in distributed computing environments where communication between tasks is expensive, such as a cluster of networked workstations running the HTCondor batch system. As a model problem, I consider computing a few smallest eigenvalues of several eigenproblems related to FE-discretization of elliptic PDEs. The matrices are exported from COMSOL.
In particular, I develop a Ritz method using method subspace constructed from local spaces associated with non-overlapping partitions of the connectivity graph of the matrix. I discuss a technique for constructing these local subspaces from boundary-to-interior mapping related to the subdomain. I give an outline of error analysis and report numerical tests conducted on a cluster at Aalto University.
Speaker: Antti Hannukainen (Aalto University) -
77
Householder Sets and Relative Pseudospectra
We study eigenvalue inclusion regions first defined by Householder and explore applications to large, sparse matrices. We also study pseudospectra and relative pseudospectra, which are similarly defined eigenvalue inclusion regions. Then we explore connections between these sets.
Speaker: Paul Zachlin (Lakeland Community College)
-
74
-
Numerical Linear Algebra in Machine Learning: C McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
78
The Fréchet derivative of the tensor t-function
The tensor t-function, a formalism that generalizes the well-known concept of matrix functions to third-order tensors, is introduced in Lund (Numer Linear Algebra Appl 27(3):e2288). In this work, we investigate properties of the Fréchet derivative of the tensor t-function and derive algorithms for its efficient numerical computation. Applications in condition number estimation and nuclear norm minimization are explored. Numerical experiments implemented by the t-Frechet toolbox hosted at https:// gitlab.com/katlund/t-frechet illustrate properties of the t-function Fréchet derivative, as well as the efficiency and accuracy of the proposed algorithms.
Speaker: Kathryn Lund (STFC Scientific Computing) -
79
Accurate Models of NVIDIA Tensor Cores
Matrix multiplication is a fundamental operation in both training of neural networks and inference. To accelerate matrix multiplication, Graphical Processing Units (GPUs) provide it implemented in hardware. Due to the increased throughput over the software-based matrix multiplication, the multipliers are increasingly used outside of AI, to accelerate various applications in scientific computing. However, matrix multipliers targeted at AI are at present not compliant with IEEE 754 floating-point arithmetic behaviour, with different vendors offering different numerical features. This leads to non-reproducible results across different generations of GPU architectures, at the matrix multiply-accumulate instruction level. To study numerical characteristics of matrix multipliers - such as rounding behaviour, accumulator width, normalization points, extra carry bits, and others - test vectors are typically constructed. Yet, these vectors may or may not distinguish between different hardware models, and due to limited hardware availability, their reliability across many different platforms remains largely untested.
We present software models for emulating the inner product behavior of low- and mixed-precision matrix multipliers in the V100, A100, H100 and B200 data center GPUs in most supported input formats of interest to mixed-precision algorithm developers: 8-, 16-, and 19-bit floating point. These models enable verification of test vectors before applying them to real hardware and also support computational scientists and mixed-precision algorithm developers with easy-to-use accurate models available in MATLAB - we demonstrate their use in multi-word emulation algorithms for matrix multiplication.
Speaker: Mantas Mikaitis (University of Leeds) -
80
Reduced Rank Extrapolation for Matrix Equations
Reduced rank extrapolation (RRE) is a classic acceleration method for vector-valued fixed-point process, commonly airsing from iterative solution of algebraic equations. In this talk, we discuss the generalization of this extrapolation framework to sequences of low-rank matrices generated by iterative methods for large-scale matrix equations, such as low-rank alternating directions implicit methods for Lyapunov and Riccati equations, as well as generalized Sylvester equation. We also briefly discuss extrapolation strategies for fixed-point iterations arising in machine learning algorithms.
Speaker: Xiaobo Liu (Max Planck Institute for Dynamics of Complex Technical Systems) -
81
Look-ahead mixed-precision inference of LLMs
We address the floating-point computation of compositionally-rich functions, concentrating on LLM inference. Based on the rounding error analysis of a composition, we provide an adaptive strategy to select components of the inner function that need to be recomputed more accurately to improve the numerical stability. We explain how this strategy can be applied to different compositions within a transformer neural network and illustrate its overall effect on LLM inference.
Speaker: Stanislav Budzinskiy (University of Vienna)
-
78
-
New Directions and Challenges in Linear Algebra: C Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
82
Column subset selection: A new perspective
It is well known that the problem of selecting k columns with maximal volume from a real matrix is NP-hard, and does not admit a polynomial time approximation scheme. However, one can show that the logarithm of the volume is a "submodular" function. Intuitively, this means: Adding a new column to a larger submatrix tends to be less beneficial than adding it to a smaller submatrix.
If subset selection for maximizing the logarithmic volume is implemented via QR with column pivoting, we show that: The Businger Golub algorithm is a greedy algorithm with relative error at most .37, while the Gu-Eisenstat algorithm is a 1-interchange algorithm with relative error at most .5
Speaker: Ilse Ipsen (North Carolina State University) -
83
Structured Matrix Approximation via Matrix-Vector Products
In this talk, I will give an overview of recent progress on the problem of structured matrix approximation from matrix-vector products. Given a target matrix A that can only be accessed through a limited number of (possibly adaptively chosen) matrix-vector products, we seek to find a near-optimal approximation to A from some structured matrix class – e.g., a low-rank approximation, a hierarchical low-rank approximation, a sparse or diagonal approximation, etc. This general problem arises across the computational sciences and data science, both in algorithmic applications and, more recently, in scientific machine learning, where it is closely related to the problem of linear operator learning from input/output samples.
I will overview recent work, where we give 1) optimal algorithms for approximating A with a matrix with a fixed sparsity pattern (e.g., a diagonal or banded matrix), 2) the first algorithms with strong relative error bounds for hierarchical low-rank approximation, and 3) the first bounds for generic structured families with sample complexity depending on the parametric complexity of the family. I will highlight several open questions related to these results.
Speaker: Cameron Musco (University of Massachusetts Amherst) -
84
How ill conditioned can sub-matrices of the Fourier matrix be?
The discrete Fourier transform matrix, and sub-matrices of it, appear in a wide variety of applications. While the Fourier matrix itself is a scaled unitary matrix, its sub-matrices can be exponentially ill-conditioned. In this talk, we discuss applications, prior work, and we provide tight estimates for just how ill-conditioned such matrices can be.
Speaker: John Urschel (MIT) -
85
Inner product-free approximation of matrix functions
Randomized sketching is a promising tool to reduce the number of inner products computed in Krylov methods for solving large systems of linear equations $Ax = b$, or more generally, when approximating the action of a matrix function on a vector, $f(A)b$. For the case of linear systems, it has recently been observed that one can often get away with completely inner product-free Krylov-based preconditioners when a stabilizing outer iteration (like FGMRES) is used. In this talk we discuss whether similar (almost) inner product-free methods can be devised for the $f(A)b$ problem.
Speaker: Stefan Güttel (The University of Manchester)
-
82
-
Advanced Acceleration and Convergence Techniques for Solving Linear and Nonlinear Systems: C McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
86
Constant Memory and Synchronization Costs for Nonsymmetric Krylov Methods
For nonsymmetric operators, GMRES convergence is governed not only by the eigenvalues but also by the field of values and the resolvent norm. Non-normality amplifies the resolvent $||(zI - A)^{-1}||$ far from the spectrum, and the spectral geometry of the numerical range $W(A)$ determines how rapidly GMRES residual polynomials can decrease. The intrinsic information dimension $K_\infty$, a problem-dependent quantity measuring the spectral information a Krylov process can resolve above the floating-point noise floor, provides an effective subspace dimension for non-normal operators.
This talk presents an $s$-step Newton-Leja GMRES method that addresses non-normality directly. The Newton-Leja polynomial basis generates the Krylov subspace with conditioning controlled by the spectral geometry of $W(A)$, achieving constant memory through the s-step recurrence. NL--GMRES minimises the residual with one global reduction per MGS orthogonalisation step. A polar preconditioner based on the Newton-Schulz iteration applied per block to the basis vectors restores orthogonality without forming the Gram matrix, counteracting the basis degradation caused by non-normality. Together, these components achieve constant memory and bounded synchronisation cost, with convergence governed by the field of values and bounded by $K_\infty$.
Numerical experiments on non-normal convection-diffusion operators illustrate the role of spectral geometry in determining both convergence and basis conditioning.
Speaker: Stephen Thomas (Lehigh University) -
87
A Generalized Alternating Anderson Acceleration Method
In this talk, we propose a generalized alternating Anderson acceleration method, a periodic scheme composed of $t$ fixed-point iteration steps, interleaved with $s$ steps of Anderson acceleration with window size $m$, to solve linear and nonlinear problems. This allows flexibility to use different combinations of fixed-point iteration and Anderson iteration. We present a convergence analysis of the proposed scheme for accelerating the Richardson iteration in the linear case, with a focus on specific parameter choices of interest. Specifically, we prove convergence of the proposed method under contractive fixed-point iteration and provide a sufficient condition for convergence when the Richardson iteration matrix is diagonalizable and noncontractive. To demonstrate the broader applicability of our proposed method, we use it to accelerate Picard iteration, gradient descent, and the alternating direction method of multipliers in solving partial differential equations and nonlinear, nonsmooth optimization problems. The numerical results illustrate that the proposed scheme is more efficient than the existing windowed Anderson acceleration and alternating Anderson ($s=1$) in terms of iteration number and CPU time for careful choice of parameters $m, s, t$.
Speaker: Yunhui He (University of Houston) -
88
nlKrylov: A unified framework for nonlinear GCR-type Krylov subspace methods
In this talk, we introduce a unified framework for nonlinear Krylov subspace methods (nlKrylov ) to solve systems of nonlinear equations. Building on the recent development of nlTGCR as well as earlier work on classical GCR-like linear Krylov solvers such as GMRESR, we generalize these approaches to non-linear problems via nested algorithmic structures. We establish connections of nlKrylov methods to other existing nonlinear methods such as quasi-Newton and subspace projection methods. Our theory is completed by rigorous convergence results for problems with both nonsingular and singular Jacobian. The framework is further extended to matrix-valued rootfinding problems using global nonlinear Krylov approaches. Extensive numerical experiments validate the theoretical insights and demonstrate the robustness and efficiency of our proposed algorithms.
Speaker: Tom Werner (TU Braunschweig) -
89
Accelerated local-global coupling for non-isothermal multiphase reactive flow in hydrate-bearing sediments
We present a dynamic local–global coupling strategy for non-isothermal multiphase reactive flow in hydrate-bearing sediments, where hydrate phase change strongly couples transport, pressure, and temperature through variations in porosity and permeability. The nonlinear system is solved by sequential Picard iterations with physics-based splitting into transport, flow, and heat processes. To reduce the cost of global iterations, we introduce dynamic nonlinearity localization. Reaction-based markers identify active subdomains, which are enlarged by oversampling (buffer cells). A locally accelerated nonlinear solve is performed in these regions before a global correction step.
Speaker: Maria Vasilyeva
-
86
-
Matrix Inequalities, Matrix Equations, and Their Applications: C Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
90
Permutations induced by GEPP
Gaussian elimination with partial pivoting (GEPP) remains the most widely used dense linear solver. GEPP produces the factorization $PA = LU$, where $L$ and $U$ are lower and upper triangular matrices and $P$ is a permutation matrix; together, these encode the pivoting strategy, directly influencing stability through classical growth-factor bounds and matrix norm inequalities. When $A$ is random, the permutation arising from the $P$ factor is itself random. When is this permutation uniform? How many cycles appear in its disjoint cycle decomposition (equivalently, how many pivot movements occur during GEPP — providing a new link between random matrix theory and permutation statistics through numerical linear algebra)? What is the length of the longest increasing subsequence of this permutation? We provide statistical answers to these questions for selected random matrix ensembles and transformations. For structured cases involving butterfly permutations, we present full distributional descriptions of these statistics. Moreover, we introduce a random butterfly matrix ensemble whose GEPP-induced permutations are distributed according to Haar measure on a full 2-Sylow subgroup of the symmetric group on a set of size $2^n$.
Speaker: John Peca-Medlin (UCSD) -
91
Stabilizer Fields and Dimension Growth in Product-Spans
We survey classical results in additive combinatorics and develop linear analogues over field extensions, with an emphasis on Kneser-type phenomena. In addition to recalling Kneser's theorem and stabilizer methods (including Cauchy--Davenport and DeVos's refinement), we present a rigidity theorem showing that if $|A+B|=|A|+|B|-1$ with $A+B\neq G$, then $A+B$ is a subgroup and $A$ is a coset; an $n$-fold Kneser bound expressed in terms of stabilizers; a coset-sparsity condition that yields near Cauchy--Davenport bounds; and a density estimate guaranteeing large sumsets.
We then establish linear analogues for finite-dimensional $K$-subspaces $A,B\subseteq L$: the $K$-span $\langle AB\rangle$ admits stabilizers that are intermediate fields, leading to a linear rigidity theorem and sharpened Hou--Leung--Xiang bounds under separability.
We conclude with open problems on weakening separability assumptions and on bridging the group and linear settings via stabilizer geometry and coset projections.
Speaker: Mohsen Aliabadi (Clayton State University) -
92
From binary symmetric matrices to Coxeter-like graphs and self-dual codes
The matrix equation $rank(A-B)=1$ is well studied in linear algebra and combinatorics within preserver problems and the theory of distance-regular graphs/association schemes. In the talk I will present how this equality is related to coding theory, namely to binary self-dual codes.
Let $\widehat{\Gamma}_{n}$ be the graph with the vertex set formed by all $n\times n$ symmetric matrices with coefficients in the binary field $\mathbb{F}_2=\{0,1\}$ where two matrices $A$ and $B$ form an edge if and only if $rank(A-B)=1$. Graph $\widehat{\Gamma}_n$ was studied in numerous papers and the distance function $d_{\widehat{\Gamma}_{n}}(A,B)$ between arbitrary vertices is well known and easy to compute. On the other hand, its subgraph $\Gamma_n$, which is induced by invertible matrices, has not attained much attention till recently. In fact, graph $\Gamma_n$ was introduced in 2015 [1]. It generalizes the well-known Coxeter graph (obtained if $n=3$).
Recently, we computed the distance function $d_{\Gamma_n}(A,B)$ [2], which turns out to be related to coding theory. Namely, for odd $n\geq 3$, each linear self-dual code $C$ in $\mathbb{F}_2^{n+1}$ can be identified with a certain subset $\mathcal{F}_C$ of the set
$\mathcal{SD}_n=\left\{A\in V(\Gamma_n): d_{\Gamma_n}(A,I)=\frac{n+5}{2}, rank(A-I)=\frac{n+1}{2} \right\}$ where $I$ is the identity matrix. In particular, the matrices $A\in V(\Gamma_n)$, which represent self-dual codes, are fully determined by the values of two graph parameters: $d_{\Gamma_n}(A,I)$ and $d_{\widehat{\Gamma}_n}(A,I)$. In the talk, I will describe the identification $C\leftrightarrow \mathcal{F}_C$ in more details.References
[1] M. Orel, On generalizations of the Petersen and the Coxeter graph. Electron. J. Combin. 22(4) (2015), Paper #P.4.27.
[2] M. Orel, D. Višnjić, The distance function on Coxeter-like graphs and self-dual codes. Finite Fields Appl. 103 (2025), paper 102580, 51 pp., https://doi.org/10.1016/j.ffa.2025.102580Speaker: Marko Orel (University of Primorska) -
93
Inequalities for different means of positive definite matrices
We discuss majorisation inequalities for different means of positive definite matrices focusing on the geometric mean, the Wasserstein mean, the log Euclidean mean and the power mean.
Speaker: Tanvi Jain (Indian Statistical Institute)
-
90
-
Spectral Graph Theory: C Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
94
The inverse symplectic eigenvalue problem of a graph
The graph ${\mathcal G}(A)$ of a real symmetric matrix $n\times n$ matrix $A=[a_{ij}]$ has vertices $V=\{1,\dots,n\}$ and edges $E=\{ \{i,j\}: a_{ij}\ne 0\mbox{ and } i\ne j\}$. The set of matrices described by a graph $G$ is ${\mathcal S}(G)=\{A\in{\mathbb R}^{n\times n}:{\mathcal G}(A)=G \mbox{ and } A^\top=A\}$ and the
Inverse Eigenvalue Problem of $G$ (IEP-$G$) is to determine all possible spectra (i.e., multisets of eigenvalues) of matrices $A\in {\mathcal S}(G)$. Other inverse eigenvalue problems for $G$ have also been studied based on matrix properties, e.g., the PSD Inverse Eigenvalue Problem of $G$ where matrices must be positive semidefinite.A symplectic matrix is a $(2n) \times (2n)$ real matrix $S$ such that $S^\top \Omega S=\Omega$ where
$\Omega=\left[\begin{array}{r|r} O&I\\\hline -I&O \end{array} \right].$ It is known that for any $(2n) \times (2n)$ symmetric positive definite matrix $A$, there exists a symplectic matrix $S$ and an $n\times n$ diagonal matrix $D$ such that $S^\top A S= \left[\begin{array}{c|c}D&O\\ \hline O&D\end{array} \right] .$ The diagonal entries of $D$ are unique (up to re-ordering) and are called the symplectic eigenvalues of $A$; the symplectic spectrum of $A$ is the multiset of symplectic eigenvalues of $A$. The Inverse Symplectic Eigenvalue Problem of $G$ (ISEP-$G$) is to determine all possible symplectic spectra of a labeled graph $G$ with vertex set $\{1,\dots,2n\}$ (unlike the IEP-$G$, the labeling affects the symplectic spectrum). There are many differences between the IEP-$G$ and the ISEP-$G$, e.g., that there are many labelled graphs that allow all possible symplectic spectra including all symplectic eigenvalues equal. Whereas, for the IEP-$G$, all eigenvalues equal is allowed only by a graph with no edges.This talk will provide an introduction to the ISEP-$G$, including methods to construct matrices with any given symplectic spectrum and solutions to the ISEP-$G$ for all labeled graphs of order four.
Speaker: Leslie Hogben (American Institute of Mathematics, Iowa State University, Purdue University) -
95
Zero-Nonzero Patterns of Symmetric Orthogonal Matrices
Given a graph $G$ on $n$ vertices, $\mathcal{S}(G)$ is the set of symmetric $n\times n$ matrices with the same off-diagonal zero pattern as the adjacency matrix $A(G)$. We say that a graph $G$ has $q(G)=2$ if there is a matrix $M\in \mathcal{S}(G)$ with exactly $2$ distinct eigenvalues. This is equivalent to the existence of an orthogonal matrix $M\in\mathcal{S}(G)$. We are interested in understanding the graphs $G$ for which $q(G)=2$ (or equivalently, the zero-nonzero patterns of symmetric orthogonal matrices).
In this talk, we focus on the following conjecture: if the complement of $G$ has at most $n-3$ edges, then $q(G)=2$. This conjecture has been resolved for graphs with bipartite complement, and verified in some cases where the complement of $G$ is not bipartite. We report on this ongoing project.
Speaker: Brendan Rooney (Rochester Institute of Technology) -
96
$q(G)$ for Threshold Graphs
We study the minimum number of distinct eigenvalues $q(G)$ of threshold graphs. It is known that every threshold graph satisfies $q(G)\leq 4$, which suggests that a complete combinatorial characterization according to the values $q(G)=2,3,4$ is possible.
In this talk, we present ongoing results toward such a characterization.Speaker: Luiz Emilio Allem (Universidade Federal do Rio Grande do Sul) -
97
How many eigenvalues of a tree can attain the maximum multiplicity?
For a tree $T$ we consider the set ${\mathcal S}(T)$ of real symmetric matrices whose off-diagonal zero-nonzero pattern is equal to the pattern of the adjacency matrix of $T$. It is well known that the maximum multiplicity of an eigenvalue over matrices in ${\mathcal S}(T)$ is equal to the path cover number $P(T)$ of the tree $T$.
We present a novel decomposition of the tree into a set of vertices and paths, which serves as a tool for analysing matrices achieving maximum multiplicity $P(T)$. We define a parameter ${\rm MM}(T)$ to be the maximal positive integer $k$ for which there exists a matrix $A\in S(T)$ with $k$ eigenvalues achieving the multiplicity $P(T)$. We derive a combinatorial upper bound for ${\rm MM}(T)$ and identify some families of trees that achieve this bound.
Speaker: Polona Oblak (University of Ljubljana)
-
94
-
Numerical Linear Algebra Tools for Model Order Reduction: C Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
98
Inverse problems for history-enriched linear model reduction
Standard projection-based model reduction for dynamical systems incurs closure error because it only accounts for instantaneous dependence on the resolved state. From the Mori–Zwanzig (MZ) perspective, projecting the full dynamics onto a low-dimensional resolved subspace induces additional noise and memory terms arising from the dynamics of the unresolved component in the orthogonal complement. The memory term makes the resolved dynamics explicitly history dependent. In this work, based on the MZ identity, we derive exact, history-enriched models for the resolved dynamics of linear driven dynamical systems and formulate inverse problems to learn model operators from discrete snapshot data via least-squares regression. We propose a greedy time-marching scheme to solve the inverse problems efficiently and analyze operator identifiability under full and partial observation data availability. For full observation data, we show that, under mild assumptions, the operators are identifiable even when the full-state dynamics are governed by a general time-varying linear operator, whereas with partial observation data the inverse problem has a unique solution only when the full-state operator is time-invariant. To address the resulting non-uniqueness in the time-varying case, we introduce a time-smoothing Tikhonov regularization. Numerical results demonstrate that the operators can be faithfully reconstructed from both full and partial observation data and that the learned history-enriched MZ models yield accurate trajectories of the resolved state.
Speaker: Arjun Vijaywargiya (University of Texas at Austin) -
99
Efficient Streaming Operator Learning for Large-Scale Dynamical Systems
Modeling and simulation of real-world applications often involve dynamical systems with large degrees of freedom, requiring substantial computational time and resources. Projection-based model reduction enables efficient simulation of such dynamical systems by constructing low-dimensional surrogate models from high-dimensional data. Specifically, Operator Inference (OpInf) learns such reduced surrogate models through a two-step process: constructing a low-dimensional basis via Singular Value Decomposition (SVD) to compress the data, then solving a linear least-squares (LS) problem to infer reduced operators that govern the dynamics in this compressed space, all without access to the underlying code or full model operators, i.e., non-intrusively. Traditional OpInf operates as a batch learning method, where both the SVD and LS steps process all data simultaneously, which limits scalability to large-scale applications generating terabytes to petabytes of data and prevents real-time model updates in online scenarios. To address these limitations, we propose Streaming OpInf, which learns reduced models incrementally as snapshot data arrives. Our method employs prominent streaming algorithms from numerical linear algebra: incremental SVD for adaptive basis construction and recursive LS for streaming operator updates, eliminating the need to store complete datasets while enabling online model adaptation. We systematically compare multiple streaming algorithm variants to identify effective combinations for accurate reduced model learning. A Numerical experiment on a large-scale turbulent channel flow with a friction Reynolds number of $Re_\tau = 5200$ demonstrate that Streaming OpInf achieves accuracy comparable to batch OpInf while reducing memory requirements significantly and enabling dimension reductions exceeding 31,000x, resulting in orders-of-magnitude faster predictions.
Speaker: Tomoki Koike (Georgia Institute of Technology) -
100
Solving Generalized Lyapunov Equations with guarantees: application to the Model Reduction of Switched Linear Systems
In this talk, we present an efficient strategy to approximate the solutions of large-scale generalized Lyapunov equations (GLEs) while providing rigorous error guarantees. The motivation for this study stems from the use of GLEs in model order reduction (MOR) of switched linear systems (SLS) in control form. Specifically, we analyze how inaccuracies in the computed GLE solution influence the performance and reliability of the resulting MOR procedure. Furthermore, since the classical balanced-truncation error estimate for MOR of SLS is neither theoretically nor practically viable—because it relies on restrictive assumptions that require several linear matrix inequalities (LMI) to be satisfied by numerically computed solutions of the GLEs—we propose a new MOR framework for SLS. Our method is based on solving multiple GLEs and constructing projection matrices that are piecewise constant in time to appropriately balance and subsequently reduce the SLS; we therefore refer to it as piecewise balancing reduction (PBR) for SLS. We extend the standard balance-truncation error bounds to incorporate the effects of inexact LMI. We show how the PBR formulation allows us to control the error arising from the inexact LMI. In addition, our new error bound accounts for the influence of the piecewise constant time-varying projection matrices. Altogether, this renders the PBR approach for SLS applicable to a broad and flexible class of SLS. We conclude by showing numerical experiments to corroborate our theoretical results.
Speaker: Dr Mattia Manucci (Karlsruhe Institute of Technology) -
101
$\mathcal{H}_2$ model order reduction for Bilinear Quadratic Output Systems
Today, mathematical modeling is dominated by increasingly high-dimensional and complex dynamical systems. One special type of structure is the bilinear state equation, which either naturally appears in various applications or results from the Carleman bilinearization of the underlying nonlinear dynamics. Recently, dynamical systems with quadratic outputs have also gained significant attention as they appear, e.g., in the modeling of the variance of a quantity of interest of a stochastic model. As a combination, we study the so-called bilinear quadratic output (BQO) systems described as
$\dot{x}(t)=Ax(t)+Nx(t)u(t) + Bu(t), y(t)=Cx(t)+x(t)^TMx(t)$
with state $x(t)\in\mathbb{R}^n$, scalar input $u(t)$ and output $y(t)$ and matrices $A, B, C, N$, and $M$ of suitable dimensions (SISO case).BQO systems generalize bilinear systems ($M=0$) and linear quadratic output systems ($N=0$). For large dimensional systems, the task of structure-preserving model order reduction arises. Thus, we seek a BQO system of reduced dimension that approximates the original input-output behavior well. Whereas a balancing approach has been already discussed in [2], here we focus on an $\mathcal{H}_2$ optimal approach, similar to $\mathcal{H}_2$ optimal reduction for bilinear systems; see, e.g., [3,1]. We will establish the framework for the $\mathcal{H}_2$ model order reduction of BQO systems, develop iterative algorithms for finding a locally optimal reduced system, and test these algorithms on several numerical examples.
References:
[1] P. Benner and T. Breiten. Interpolation-based $\mathcal{H}_2$-model reduction of bilinear control systems. SIAM J. Matrix Anal. Appl., 33(3):859-885, 2012.
[2] H. Faßbender, S. Gugercin, and T. Peters. Bilinear Quadratic Output Systems and Balanced Truncation, July 2025. arXiv:2507.03684 [math].
[3] L. Zhang and J. Lam. On $\mathcal{H}_2$ model reduction of bilinear systems. Automatica, 38(2):205-216, Feb. 2002.Speaker: Till Peters (TU Braunschweig, Institute for Numerical Analysis)
-
98
-
Computational Advances in Discrete Inverse Problems: C Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
102
A Regularization Method for Compact Linear Operator Equations Based on the Arnoldi Process
We consider the numerical solution of linear operator equations involving compact operators. Since compact operators do not admit bounded inverses, the associated equations are ill-posed and require regularization. The Arnoldi process provides a natural framework for approximating a compact operator by a nearby operator of finite rank, thereby reducing the infinite-dimensional problem to a sequence of small, structured subproblems. Regularization is incorporated by applying Tikhonov’s method to the projected equations.
This work investigates theoretical properties of the resulting Arnoldi–Tikhonov approach, including convergence behavior and the influence of Krylov subspace dimension on the regularized solutions. Numerical experiments are presented to illustrate the theoretical results and to demonstrate the effectiveness of the method for representative compact operator equations.
Speaker: Mykhailo Kuian (Case Western Reserve University) -
103
Recycling and Streaming for Large Scale Nonlinear Inverse Problems
Many imaging inverse problems assume a known linear forward operator, yet practical systems often suffer from uncertainty in acquisition geometry, such as projection angles in computed tomography, sensor positions in photoacoustic tomography. These uncertainties introduce nonlinearity and require joint estimation of both the image and the forward model parameters.
We propose a nonlinear recycled majorization–minimization generalized Krylov subspace (NL-RMM-GKS) framework for large-scale inverse problems with uncertain forward operators. The method extends MM-GKS to nonlinear settings by combining majorization–minimization for nonsmooth regularization with Krylov subspace projection and subspace recycling, ensuring bounded memory usage.
Two complementary formulations are developed: an alternating minimization approach that alternates between image updates and Gauss–Newton parameter estimation, and a variable projection approach that eliminates the image variable and optimizes directly over the parameters using inexact inner solves. We further introduce streaming variants that process data sequentially, enabling reconstruction from large or dynamically acquired datasets without storing the full operator.
We carry out rigorous numerical experiments in fan-beam computed tomography and photoacoustic tomography to demonstrate that our proposed framework achieves high-quality reconstructions with bounded memory requirements, making it suitable for large-scale and dynamic imaging problems with uncertain geometry.
Speaker: Toluwani Okunola (Tufts University) -
104
Latent Twin Operator
Learning solution operators in a manner that is independent of discretization and resolution remains a central challenge in data-driven modeling. The latent twins framework addresses this problem by constructing operators in a task-adaptive latent space for inverse problems and differential equations. However, in its classical form, latent twins relies on autoencoder architectures that are tied to fixed discretizations, coupling representation learning to a particular grid or resolution.
We address this limitation by separating state information from coordinate information. The latent variables encode the global system state, while a coordinate-conditioned decoder acts as an evaluation operator that reconstructs the state at arbitrary spatial or temporal locations through a family of linear or nonlinear maps. This viewpoint naturally supports sparse, irregular, and multi-resolution data, and connects latent twins to operator learning and reduced-order modeling perspectives. The resulting framework is well suited for applications in imaging, inverse problems, and the reduced-order modeling of time-dependent systems.
Speaker: Riley Yizhou Chen (Emory University) -
105
New flexible and inexact Krylov solvers for inverse problems
This talk presents a new family of algorithms for large-scale linear inverse problems built on flexible and inexact variants of the Golub–Kahan factorization. The proposed approach constructs regularized solutions through a sequence of projected (re)weighted least-squares problems, where the projection spaces are adaptively generated and endowed with iteration-dependent preconditioning and controlled inexactness. This framework enables a unified and flexible treatment of challenging problem settings, including general data fidelity models such as those involving p-norms. Numerical experiments in imaging applications, such as deblurring and computed tomography, highlight the effectiveness and competitiveness of the proposed methods with respect other popular methods.
Speaker: Dr Malena Sabate Landman (University of Bath)
-
102
-
Advances and Challenges in Eigensolvers: C Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
106
The accuracy of the QZ algorithm and some tricks to improve it
For small and/or dense generalized eigenvalue problems, the QZ method remains the method of choice because of its robustness. In this talk, we will discuss how that robustness may occasionally fail, especially if the matrices are badly scaled or singular. We will discuss how new deflation criteria and some changes in the preprocessing can improve the robustness.
Speaker: Dr Thijs Steel (KU Leuven) -
107
From eigenvector nonlinearities to eigenvalue nonlinearities
We are interested in an eigenvector-nonlinear eigenvalue problem (NEPv), that is, a problem of the form $A(x)x=\lambda x$, where $A:\mathbb{R}^{n}\rightarrow\mathbb{R}^{n\times n}$ is symmetric, and the eigenvector has a prescribed norm, for instance $\Vert x \Vert = 1$. In this sense, this class of problems generalize the linear eigenvalue problem. Motivated by applications such as the Gross-Pitaevskii equation (GPE) [E. Jarlebring and P. Henning. SIAM Review, 67:256-317, 2025], we further specialize this problem by considering a specific structure of $A(x)$. More precisely, we consider the case where the nonlinearity appears as a sum of products of scalar functions of the eigenvector, and rank-one matrices. We present a method based on transforming this class of problems into equivalent problems with eigenvalue nonlinearities (NEP), i.e., problems of the form $M(\lambda)x=0$. In our application, $M:\mathbb{R}\rightarrow\mathbb{R}^{n\times n}$ is an algebraic function of the eigenvalue. This transformation enables us to use efficient methods for NEPs as a means of obtaining solutions to the NEPv. In particular, these methods can efficiently compute several eigenvalues, something that is rare among methods for the NEPv. We show how the transformation is constructed theoretically, and how it can be handled in practice. A numerical illustration on a large-scale problem related to a modification of the GPE shows the effectiveness of our approach in computing several eigenvalues of the original problem. This presentation is based on the preprint [E. Jarlebring and V. P. Lithell. arXiv:2506.16182, 2025].
Speaker: Vilhelm P. Lithell (KTH Royal Institute of Technology) -
108
Fast Algorithms for Optimal Damping in Mechanical Systems
Optimal damping aims at determining a vector of damping coefficients $\nu$ that maximizes the decay rate of a mechanical system's response. This problem can be formulated as the minimization of the trace of the solution of a Lyapunov equation whose coefficient matrix depends on $\nu$. For physical relevance, the damping coefficients must be nonnegative and the resulting system must be asymptotically stable. We derive explicit expressions for the gradient and Hessian of the objective function and show that the Karush--Kuhn--Tucker conditions are equivalent to the vanishing of a nonlinear residual function at an optimal solution. To compute such a solution, we propose a Barzilai--Borwein residual minimization algorithm (BBRMA), which is simple and efficient but not globally convergent, and a spectral projected gradient (SPG) method, which is globally convergent. By exploiting the structure of the problem, we show how to efficiently compute the objective function and its gradient, with eigenvalue decompositions constituting the dominant cost in terms of execution time. Numerical experiments show that both methods require fewer eigenvalue decompositions than the fast optimal damping algorithm (FODA), and that, although SPG may incur additional decompositions due to line search, it often converges faster than BBRMA, leading to a lower overall computational effort.
Speaker: Francoise Tisseur (The University of Manchester) -
109
From SCF to LOBPCG: Accelerated Solvers for Nonlinear Eigenvector Problems
This talk surveys modern iterative algorithms for solving large-scale nonlinear eigenvector problems, with a focus on both foundational methods and recent advances in nonlinear acceleration techniques. We examine SCF, FEAST, BPSD, and the LOBPCG eigenvector problems solvers, highlighting their theoretical foundations, practical performance, and applications. Through comparisons and illustrative examples, the talk provides a unified perspective and guidance for selecting and tuning solvers in practice.
Speaker: Agnieszka Miedlar (Virginia Tech)
-
106
-
Linear Algebra Education: C Torgersen Hall 2150
Torgersen Hall 2150
Virginia Tech
-
110
Immersive Exercises for Linear Algebra
Modern technology is most often employed in two ways in math education: to enhance explanations (e.g. animated videos or widgets), or to find solutions and calculate answers. The fact that students sometimes do poorly even when taught by excellent teachers shows the limitations of the first strategy. The second strategy, which has been made incredibly easy by generative AI, is increasingly used by students, to the detriment of their learning (and grades). Learning mathematics requires both immersion in problem-solving, and practice. I will describe immersive exercises (IEs) that I have designed for sophomore-level linear algebra, in which students make problem-solving decisions in real time while the software handles routine calculations and provides immediate feedback. Unencumbered by low-level calculations, students may solve problems rapidly, and practice far more (and more complex) problems. External "black box" computational software is not needed. Students quickly realize that entering individual problem-solving steps in real time makes using AI solutions less efficient than simply learning to solve the problems and practicing. I have designed 75 IEs that have been coded by software engineers from The Math Atlas, which will be licensed to students to be used with the book Interactive Linear Algebra published this year by Taylor and Francis.
Speaker: Prof. Conrad Plaut (University of Tennessee--Mathematics) -
111
Finding the Right Basis (or Bases)
Traditionally, half or more of a first course in linear algebra has been devoted to the theory and machinery required to solve linear systems and to find bases for the four fundamental subspaces associated with a matrix. At the heart of this is the Gauss-Jordan algorithm, used to construct the reduced row echelon form (RREF) of a matrix. Far too many students complete their only linear algebra by desperately clutching the manual computation of the RREF as if it were a life preserver in a frightening sea of theory, an approach fostered by many texts.
Increasingly, students arrive in linear algebra less well-prepared to handle definitions and the reasoning of proofs. Simultaneously client disciplines are increasingly using concepts rarely or only lightly covered in a first course: the spectral theorem, the QR algorithm, orthogonal projections, the SVD (or its applied cousin, PCA). Data science and machine learning, heavy users of these techniques, did not even exist as programs of study ten years ago.
Over the last decade, talks in linear algebra education sessions have focused on what needs to be in a "new" version of linear algebra that better addresses the changing needs of our client disciplines. Frank Uhlig has been a notable proponent of making significant changes. Others have suggested ways to cleverly fold advanced topics into a traditional course. How do we reconcile the weaknesses of incoming students and the deepening topical needs of our clients?
My original title for this talk was Matrices Just Want to Be Diagonal, a title which emphasizes algorithms. Then I realized that those algorithms are really consequences of a different talk, Finding The Right Basis (or Bases). Naturally, I will discuss what constitutes a "right" basis, and what content changes to an introductory course might enable us to develop such bases and the resulting algorithms.
Speaker: Jeffrey Stuart (Pacific Lutheran University) -
112
Cards on the Table: Playing with Linear Algebra Concepts
Linear Algebra is a core mathematics course in numerous undergraduate programs, yet students often struggle with its abstract nature and with recognizing key connections between topics. This lack of conceptual coherence motivates the need for innovative teaching approaches that make linear algebra more accessible, engaging, and meaningful.
We developed a Linear Algebra card game grounded in ludodidactical principles and aimed at fostering conceptual understanding through collaborative learning. The game is not intended to trivialize mathematical content, but to elicit meaningful interaction and explanation among students. It is played in groups of 2 to 6 students and invites players to strategically place and connect concept cards representing key topics in linear algebra (such as systems of linear equations, vector spaces, eigenvalues) by explaining the mathematical relationships between them. Players earn points for correct and meaningful connections, shifting the emphasis from computation to explanation and reasoning.
The game was implemented in a variety of educational settings and evaluated using end-of-course student surveys focusing on engagement and learning outcomes. Preliminary results indicate increased student engagement, improved understanding of relationships between linear algebra concepts, and positive perceptions of learning through peer interaction. In addition, multiple playtesting sessions with colleagues in mathematics education were organized to iteratively refine both the game mechanics and its pedagogical focus.
Our findings suggest that well-designed ludodidactical tools can play a valuable role in supporting conceptual learning in abstract mathematics. The Linear Algebra Card Game is adaptable across different course levels, and all materials are shared as open educational resources.
Speaker: Martijn Boussé (Maastricht University)
-
110
-
Low-rank Matrix and Tensor Decompositions: Theory, Algorithms and Applications: C Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
113
Robustness of Minimum-Volume Nonnegative Matrix Factorization under an Expanded Sufficiently Scattered Condition
Minimum-volume nonnegative matrix factorization (min-vol NMF) has been used successfully in many applications, such as hyperspectral imaging, chemical kinetics, spectroscopy, topic modeling, and audio source separation. However, its robustness to noise has been a long-standing open problem. In this paper, we prove that min-vol NMF identifies the groundtruth factors in the presence of noise under a condition referred to as the expanded sufficiently scattered condition which requires the data points to be sufficiently well scattered in the latent simplex generated by the basis vectors.
Speaker: Subhayan Saha (Universite de Mons) -
114
Manifold-based Algorithms for the Hadamard Decomposition
Given a matrix $X$, and two ranks $r_1$ and $r_2$, the Hadamard decomposition (HD) consists in looking for two low-rank matrices, $X_1$ of rank $r_1$ and $X_2$ of rank $r_2$, both of the same size as $X$, such that $X\approx X_1\circ X_2$, where $\circ$ is the Hadamard (element-wise) product. HD is more expressive than standard low-rank approximations, such as the truncated singular value decomposition (TSVD), as it can represent higher-rank matrices with the same number of parameters; this is because the rank of $X_1 \circ X_2$ is generically equal to $r_1 r_2$. In this paper, we first present some new theoretical results for HD. These allow us to develop two new manifold-based gradient descent algorithms for computing HDs: the first one uses the representation $X\approx X_1\circ X_2$, while the second one uses $X\approx WH^\top$, where $W$ and $H$ have $r_1 r_2$ columns. We also propose new initializations that allow us to obtain better HDs. We compare our algorithms and initialization strategies with the TSVD and with the state of the art. Numerical results show that the new methods are efficient and competitive, both on synthetic and real data.
Speaker: Stefano Sicilia (University of Mons) -
115
Symmetric nonnegative trifactorization rank of matrices with a given pattern without a four cycle
We consider a symmetric nonnegative matrix $A$ of order $n \times n$. A factorization of the form $A = BCB^T$, where $B$ is a nonnegative matrix of order $n \times k$ and $C$ is a symmetric nonnegative matrix of order $k \times k$, is called symmetric nonnegative trifactorization (SNT for short) of $A$. Minimal possible $k$ in such factorization is called the SNT-rank of $A$.
The zero-nonzero pattern of a matrix can be described by a simple graph that allows loops. The SNT-rank of a graph $G$ is the minimal SNT-rank of all symmetric matrices with pattern determined by $G$, and it can be characterized combinatorially using set-join covers of $G$. In the talk we will consider a family of graphs that do not contain four cycles. We will present an algorithm on the graph for computing SNT-rank of such graphs.
Speaker: Damjana Kokol Bukovšek (University of Ljubljana) -
116
Randomized algorithms for streaming low-rank approximation in tree tensor network format
In this work, we present the tree tensor network Nyström (TTNN), an algorithm that extends recent research on streamable tensor approximation, such as for Tucker or tensor-train formats, to the more general tree tensor network format, enabling a unified treatment of various existing methods. Our method retains the key features of the generalized Nyström approximation for matrices, i.e. it is randomized, single-pass, streamable, and cost-effective. Additionally, the structure of the sketching allows for parallel implementation. We provide a deterministic error bound for the algorithm and, in the specific case of Gaussian dimension reduction maps, also a probabilistic one. We also introduce a sequential variant of the algorithm, referred to as sequential tree tensor network Nyström (STTNN), which offers better performance for dense tensors. Furthermore, both algorithms are well-suited for the recompression or rounding of tensors in the tree tensor network format. Numerical experiments highlight the efficiency and effectiveness of the proposed methods.
Speaker: Gianfranco Verzella (University of Geneva)
-
113
-
Combinatorial Matrix Theory: C Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
117
The sum of a topological index and its reciprocal index
Let $G$ be a simple connected graph. A vertex-degree-based topological index is defined as $$TI_f(G) = \sum_{uv \in E(G)} f(d_u, d_v),$$ where $f(x, y)$ is a symmetric real function. In theoretical chemistry, these indices serve as essential numerical molecular descriptors in QSAR/QSPR models. In this work, we investigate the extremal properties of $TI_f + RTI_f$, defined as the sum of a topological index and its reciprocal. Focusing on the first Zagreb index ($f(x, y) = x + y$), the second Zagreb index ($f(x, y) = xy$), and the forgotten index ($f(x, y) = x^2 + y^2$), we characterize the graphs that achieve the maximum and minimum values of $TI_f + RTI_f$ among all trees. Furthermore, we extend our analysis to the extremal problem for $TI_f + RTI_f$ of $k$-uniform hypergraphs.
Speaker: Wei Gao -
118
Combinatorial characterization of matrix algebras over finite fields
In this talk we demonstrate how a matrix algebra over a finite field can be completely described using combinatorial properties. The main tool that allows one to do this is the compressed zero-divisor graph of a ring, which describes pairs of matrices $A$ and $B$ such that $AB=0$. We list a set of $5$ combinatorial axioms that uniquely determine the compressed zero-divisor graph $\Theta(M_n(\mathbb{F}))$ of the matrix algebra $M_n(\mathbb{F})$, where $|\mathbb{F}|=p^m$. Furthermore, the structure of this graph uniquely determines the ring $M_n(\mathbb{F})$ itself up to isomorphism. We also discuss some properties of the compressed zero-divisor graphs that may be useful for investigating subalgebras of $M_n(\mathbb{F})$ and individual matrices.
Speaker: Nik Stopar (University of Ljubljana) -
119
The Rank-Preserving Transversality Property
Let $A$ be an $m \times n$ real matrix. If the manifolds ${\widetilde{\cal M}_A}= \{ H^{-1} A G : G, H \text{ are nonsingular} \}$ and $Q(\text{sgn}(A))$ intersect transversally at $A,$ that is, the tangent spaces of ${\widetilde{\cal M}_A}$ and $Q(\text{sgn}(A))$ at $A$ sum to $ \mathbb R ^{m\times n},$ we say that $A$ has the rank-preserving transversality property (RPTP) and that $A$ is an RPTP matrix. We establish many important properties of RPTP matrices as well as several sign pattern classes that require the RPTP. For example, RPTP matrices are closed under permutation equivalence, diagonal equivalence, and transpose. Further, a block upper triangular matrix with all diagonal blocks square has the RPTP if and only if each diagonal block has the RPTP and at most one diagonal block is singular. Just as the Strong Spectral Property is useful in studying the spectra of symmetric matrices associated with a graph, the notion of RPTP is a useful tool for studying the minimum ranks of sign patterns and zero-nonzero patterns.
Speaker: Chris Hart (Georgia State University) -
120
On the Length of a Multiplicity List of a Graph
Given a simple graph, $G$, on $n$ vertices, let $S(G)$ be the set of $n \times n$ real symmetric matrices, $A = [a_{ij}]$, associated to $G$ where, when $i\neq j$, $a_{ij} \neq 0$ if and only if $ij$ is an edge in $G$ and the main diagonal is free to be chosen. For any square matrix, $A$, let $q(A)$ equal the number of distinct eigenvalues of $A$. The minimum number of distinct eigenvalues of $G$ is defined as $q(G) = {\rm min} \{ q(A) : A \in S(G)\}$. A natural question that arises from this association is: Does there exist an $A \in S(G)$ with $q(A) = k$ for every integer $k$ in the interval $[q(G), n]$? Since the inverse eigenvalue problem for graphs has been studied from many angles for years, a general solution seems difficult, so any progress made on subproblems, such as $q(G)$, contributes to an understanding of the whole problem. In this talk, we discuss some recent progress on this question for certain classes of graphs.
Speaker: Johnna Parenteau
-
117
-
Application-Driven Family of Matrix Computations: Factorization, Inverse, Linear Solve: C Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
121
Scalable Approximate Selected Inversion via ILU and Spectral Corrections for Sparse Systems
We analyze two parallel numerical strategies for computing selected entries of the matrix inverse of large, sparse, symmetric systems: The selected inverse method and a factorized approximate inverse method. Both techniques are aimed at computations via LU factorizations or incomplete LU (ILU) factorizations. The selected inverse approach exploits the LU/ILU factorization to recover the inverse within the pattern of the (incomplete) LU factorization. In contrast, the factorized approximate inverse method applies a truncated series expansion (Neumann-type), providing an alternative approach at the cost of reduced accuracy. To improve accuracy while keeping sparsity, we introduce low-rank corrections and an adaptive eigenvector deflation strategy. This hybrid approach sets tightening drop tolerances off when possible and helps balance accuracy, fill-in, and computation time. Numerical illustrations for both parallel and sequential computations demonstrate the effectiveness and robustness of our approach.
Speaker: Tahamina Akter (TU Braunschweig) -
122
Matrix-Free Geometric Multigrid Preconditioning Of Combined Newton-GMRES For Solving Phase-Field Fracture With Local Mesh Refinement
In this talk, the matrix-free solution of quasi-static phase-field fracture problems is further investigated. More specifically, we consider a quasi-monolithic formulation in which the irreversibility constraint is imposed with a primal-dual active set method. The resulting nonlinear problem is solved with a line-search assisted Newton method. Therein, the arising linear equation systems are solved with a generalized minimal residual method (GMRES), which is preconditioned with a matrix-free geometric multigrid method including geometric local mesh refinement. Our solver is substantiated with a numerical test on locally refined meshes. Afterwards, we discuss applications of this solver framework in fluid-filled fractures in porous media.
Speaker: Prof. Thomas Wick (Leibniz University Hannover) -
123
Loracx: Low Rank Approximations with Constraints at Exascale
Analyzing large-scale scientific data—such as molecular dynamics simulations of $MoS_2$ recrystallization—poses significant challenges for traditional methods like Nonnegative Matrix Factorization (NMF), particularly on exascale systems. In this talk, we introduce Low-Rank Approximations with Constraints at Exascale (LORACX), a scalable framework that employs distributed, GPU-accelerated NMF integrated into a modern, Python-based HPC stack.
Key innovations include communication-efficient designs using blocked and overlapped algorithms to mitigate latency and memory constraints, as well as GPU-optimized Nonnegative Least Squares (NNLS) solvers. Performance evaluations on up to 8,192 Frontier nodes demonstrate strong scalability, processing a 16.3 × 16.3 million matrix in 3 seconds and achieving 0.67 exaflops in double precision. We present detailed weak-scaling results, including computational versus communication cost analyses, and show that baseline comparisons consistently confirm the superior performance of LORACX-GPU.
Applied to $MoS_2$ molecular dynamics data, LORACX successfully identifies structural motifs and captures phase transition dynamics, highlighting its potential as a powerful tool for large-scale materials science discovery.
Speaker: Ramakrishnan Kannan -
124
Exploiting Kronecker Structure and Krylov Subspaces for Scalable Second-Order Optimization in Federated Physics-Informed Neural Networks
Solving partial differential equations (PDEs) using distributed Physics-Informed Neural Networks (PINNs) introduces major computational challenges associated with high-dimensional curvature estimation, ill-conditioned optimization landscapes, and communication overhead in federated environments. In this work, we exploit Kronecker structure and Krylov subspace methods to develop a scalable second-order optimization framework for federated PINNs applied to the 2D Poisson equation. The proposed method combines Kronecker-Factored Approximate Curvature (K-FAC) with Conjugate Gradient (CG) iterations within a matrix-free Gauss–Newton formulation, avoiding explicit Hessian or Jacobian construction through automatic differentiation-based Jacobian-vector products. The symmetric positive definite structure induced by discretized Poisson operators supports efficient CG convergence while enabling scalable curvature-aware optimization. Within the federated learning framework, each client independently solves a local Gauss–Newton system using K-FAC approximations and Krylov iterations, transmitting only local parameter corrections to the central server. Experimental results across grid resolutions from $32\times32$ to $128\times128$ demonstrate up to $57.8\%$ relative improvement in test accuracy and $44.4\%$ reduction in required communication rounds compared with first-order Adam-based optimization. These findings illustrate how structured linear algebra techniques, particularly Kronecker factorization and Krylov subspace solvers, provide an efficient foundation for communication-efficient scientific machine learning under distributed computing constraints.
Speaker: Amit Upadhyay (Indian Institute Of Technology (Indian School Of Mines) Dhanbad)
-
121
-
5:30 PM
Registration Latham Ballroom Foyer (The Inn at Virginia Tech)
Latham Ballroom Foyer
The Inn at Virginia Tech
901 Prices Fork Rd, Blacksburg, VA 24061 -
6:00 PM
Welcome Reception Latham Ballroom (The Inn at Virginia Tech)
Latham Ballroom
The Inn at Virginia Tech
901 Prices Fork Rd, Blacksburg, VA 24061
-
7:30 AM
-
-
125
Announcements McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
Plenary Talks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
126
Stochastic trace estimation for parameter-dependent matrices
Estimating the trace of a matrix, that is only accessible by matrix-vector products, is a fundamental task in scientific computing and data science. This has many applications including network analysis, estimation of matrix norms and spectral densities, estimation of log-determinants, etc. Monte Carlo methods is one of the prevalent approaches for estimating the trace of the matrix. We consider Monte Carlo methods for minimizing the trace of a symmetric matrix $A(\theta)$, parameterized (perhaps, nonlinearly) by a parameter $\theta$ from a compact space $\Theta \subset \mathbb{R}^d$. We derive bounds for the number of samples required to estimate the backward error of the Monte Carlo to a given accuracy with high probability. The bounds are based on epsilon nets and chaining and use ideas from statistical learning theory and high dimensional probability. We illustrate this problem using applications to optimal experimental design and hyperparameter estimation in Bayesian inverse problems. Joint work with Ilse Ipsen.
Speaker: Arvind Krishna Saibaba (North Carolina State University) -
127
Inverse problems on a graph: strong matrices and graph minors
Inverse problems on a graph investigate how spectral behaviors interact with the matrices associated with the given graph. Such problems not only uncover structural information about the graph from its spectral data, but also identify fundamental properties shared by all matrices defined on the graph. A classic example is the Colin de Verdière parameter, which characterizes planarity via the maximum nullity over matrices defined on the graph. Central to these studies are strong properties, which either preserve the matrix pattern while perturbing spectral data or preserve the spectral data while adjusting the pattern. A recurring theme is that if a spectral behavior is realizable by a strong matrix on a graph, then it often remains realizable for any graph containing it as a minor. In this talk, we will survey these connections and present new results on strong properties for discrete Schrödinger operators.
Speaker: Jephian C.-H. Lin (National Yang Ming Chiao Tung University)
-
126
-
10:15 AM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Contributed Talks: D1 McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
128
Nearest correlation matrices with structure: a dynamical systems approach
The nearest correlation matrix problem consists in finding the closest valid correlation matrix to a given symmetric matrix that may fail to be positive semi-definite. In other words, given a symmetric unit-diagonal matrix that is not a proper correlation matrix, one seeks the nearest positive semi-definite matrix with unit diagonal entries.
We address the problem of finding the nearest correlation matrix to a given symmetric unit-diagonal matrix under additional structural constraints such as sparsity, block, or band patterns. This task arises in applications where positive semi-definiteness must be restored without losing essential structure.
Our method combines a two-level iteration: a structured gradient flow computes feasible perturbations within the prescribed structure, while an outer Newton scheme adjusts their magnitude to meet accuracy requirements. To handle high-dimensional settings efficiently, we replace full eigenvalue decompositions with a Rayleigh quotient approximation, focusing only on the critical invariant subspace needed to restore positive semi-definiteness.
The resulting algorithm systematically incorporates structural constraints into the nearest correlation matrix problem. Numerical experiments highlight its robustness across diverse structured scenarios, with promising applications in finance, statistics, and network analysis.
References
- N. J. Higham, Computing the nearest correlation matrix — a problem from finance, IMA J. Numer. Anal. 22 (2002), 329–343.
- N. Guglielmi, C. Lubich, The matrix nearness problem and eigenvalue optimization, preprint, 2025.
- N. J. Higham, N. Strabić, Anderson acceleration of the alternating projections method for computing the nearest correlation matrix, Numer. Algorithms 72 (2016), no. 4, 1021–1042.
Speaker: Eugenio Turchet (GSSI) -
129
Linear Geometry Insights in the Expectation Maximization Algorithm Convergence
The expectation maximization (EM) algorithm, widely used in medical image tomographic reconstruction, iteratively estimates the values of a discrete distribution of emission sources $\rho_{t=1,2,..,T}$ from a set of detector measurements $A_{k=1,2,..,K}$. The two steps of an iteration are:
$A_k^{(n)} = \sum_{t} P_{k \leftarrow t} \cdot \rho_t^{(n)}$ (E) and $\rho_t^{(n+1)} = \rho_t^{(n)} \sum_k \left[ \frac{A_k}{A_k^{(n)}} P_{k \leftarrow t} \right]$ (M)
where superscript indicates estimated values at step n, and $P_{k \leftarrow t}$ is the probability that an emission from source t is detected by detector k.Using analytical tools, it has been shown that the algorithm progresses to a fixed point, but further description of the ideal algorithm behavior is incomplete.
We propose to view measurements $A_k$ as defining hyperplanes in the finite-dimensional linear space defined by the emission source values ($\rho_{t=1,2,..,T}$) (equation E). In imaging the system is overdetermined (more measurements than emission sources), thus, when the attached linear system is consistent, these hyperplanes have a single common point (the solution).
Because all values appearing in equations (E) and (M) are positive, we argue adjustments to $\rho_t$ in step (M) induced by measurements k with $\frac{A_k}{A_k^{(n)}} > 1$ cannot cancel adjustments induced by measurements k with $\frac{A_k}{A_k^{(n)}}<1$. We show the two convex cones defined by the direction vectors to the $\frac{A_k}{A_k^{(n)}} > 1$ hyperplanes and the direction vectors to the $\frac{A_k}{A_k^{(n)}}<1$ hyperplanes, respectively, with the vertex in the solution point, have only that point as intersection. Thus, the corresponding adjustments cannot cancel each other and, for a consistent overdetermined problem, the only fixed point for the EM algorithm is the solution.
The linear geometry view of the EM algorithm promises a path for better understanding of this important algorithm.
Speaker: Dacian Bonta (Emory University) -
130
Tensor Train-Compressed FDTD Solvers for Electromagnetic Simulations
This work presents a quantized tensor train (QTT)-accelerated finite-difference time-domain (FDTD) algorithm for solving Maxwell equations in 3D open domains.
Upon a QTT representation of both field variables and differential operators on uniform grids, the proposed algorithm can achieve up to logarithmic scaling in memory and per-step computational cost with respect to the number of spatial grid points. Our algorithm constitutes two major algorithmic contributions: (1) The integration of auxiliary differential equation perfectly matched layers (ADE–PML) directly in compressed form, enabling fully tensorized leapfrog time integration without decompression. (2) For complex excitations and media, QTT ranks can significantly grow due to both complex wave fields and numerical noise. To address this, we introduce spatial smoothing and filtering strategies that stabilize QTT ranks while preserving physical accuracy.
In summary, we demonstrate that classical explicit FDTD schemes can be fundamentally reinterpreted through QTT decompositions, providing new opportunities for scalable Maxwell solvers for large-scale and complex wave propagation simulations.
Speaker: Jonathan Tabares (Florida International University)
-
128
-
Contributed Talks: D2 Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
131
On the properties of solution sets of absolute value equations
We investigate the structural properties of the solution set of absolute value equations (AVE) of the form $Ax - |x| = b$. Extending the seminal work of Hladík (SIAM J. Matrix Anal. Appl., 2023), we address his open questions originally posed for $Ax + |x| = b$ and establish analogous results for the alternative form considered here. Using the equivalence between AVE and the linear complementarity problem (LCP), we derive new results on the convexity and solvability of the AVE under various matrix classes, including positive semidefinite, Metzler, and $Z$-matrices. Further, we present sufficient conditions guaranteeing unique and nonnegative solutions for $b \ge 0$ or $b \le 0$, and provide counterexamples showing that several properties valid for $Ax + |x| = b$ fail to extend to $Ax - |x| = b$. These findings contribute to the theoretical understanding of AVEs and enrich the connection between matrix theory and complementarity problems.
Speaker: Shweta Yadav (Shiv Nadar University) -
132
A Classification of Seeds via the Minimum Number of Distinct Eigenvalues
The underlying graph $G$ of a symmetric matrix $M=(m_{ij})\in \mathbb{R}^{n\times n}$ is the graph with vertex set $\{v_1,\ldots,v_n\}$ such that a pair $\{v_i,v_j\}$ with $i\neq j$ is an edge if and only if $m_{ij}\neq 0$. For a graph $G$, let $q(G)$ denote the minimum number of distinct eigenvalues among all symmetric matrices whose underlying graph is $G$. A symmetric matrix $M$ is a realization of $q(G)$ if it has underlying graph $G$ and exactly $q(G)$ distinct eigenvalues. For any tree $T$ with diameter $d$, it is well known that $q(T)\ge d+1$. We call $T$ diminimal if equality holds, that is, if $q(T)=d+1$.
Johnson and Saiago (2013) introduced a decomposition of the set of trees of any fixed diameter $d$ based on a finite family of trees, called seeds, and on an operation known as combinatorial branch duplication (CBD, for short). A tree $T'$ obtained from a tree $T$ through a sequence of CBDs is called an unfolding of $T$. In particular, every tree with diameter $d$ can be obtained as the unfolding of a unique seed of diameter $d$.
Relating the concept of seed with the minimum number of distinct eigenvalues problem, a seed $S$ is called diminimal if every unfolding of $S$ is diminimal, and defective otherwise. In this talk, we present ongoing results showing that most seeds are defective. In particular, the proportion of defective seeds tends to 1 as the diameter grows to infinity.
Speaker: Lucas Siviero Sibemberg (UFRGS)
-
131
-
Advanced Acceleration and Convergence Techniques for Solving Linear and Nonlinear Systems: D McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
133
A BFBt Preconditioner for Double Saddle-Point Systems
We consider nonsymmetric double saddle-point systems. Given the 3-by-3 block structure of the matrix, the associated block LU decomposition features two Schur complements. A theoretical question we explore is what happens when one of the Schur complements is inverted exactly and the second, nested one, is approximated. Eigenvalue analysis sheds some light on the effect of this type of inexactness on the speed of convergence of minimum residual iterative solvers. Specific analytical and computational observations are made for the case where the nested Schur complement is approximated using the BFBt method.
Speaker: Chen Greif (The Department of Computer Science) -
134
Robust Numerical Differentiation for Entropy-regularized Optimal Transport (EOT) with application to Shuffled Regression
In this presentation, I will begin by introducing shuffled regression and the entropic optimal transport (EOT) as one possible tool for solving shuffled regression. A common approach for this optimization is to use a first-order optimizer, which requires the gradient of the OT distance. For faster convergence, one might also resort to a second-order optimizer, which additionally requires the Hessian. I will present the analytical derivatives of EOT, provide a brief overview of numerical condition numbers, and explain how to compute a crucial linear system robustly. Through analytical derivation and spectral analysis, we identify the numerical instability caused by the singularity and ill-posedness of a key linear system, prove the asymptotic limits of its condition number when both sample size $N$ goes to infinity and regularization strength $\varepsilon$ goes to 0, and improve the efficiency and robustness of computation. Finally, I would like to discuss future work as well as extensions.
Speaker: Xingjie Li (UNC Charlotte) -
135
GA-NGMRES: An Alternating NGMRES Method for Accelerating First-Order Optimization
We propose a generalized alternating nonlinear generalized minimal residual method (GA-NGMRES) for accelerating first-order optimization algorithms. The method is applied to preconditioned first-order schemes by interpreting their update rules as fixed-point iterations. GA-NGMRES introduces a periodic mixing strategy that alternates between NGMRES extrapolation and fixed-point updates, resulting in improved robustness and efficiency.
We demonstrate that the proposed approach reduces both iteration counts and overall runtime compared to state-of-the-art methods. Numerical comparisons are provided against preconditioned gradient descent and preconditioned, inexact Gauss–Newton–Krylov methods. Since GA-NGMRES relies only on first-order derivative information, it is straightforward to implement. Performance is evaluated with respect to algorithmic hyperparameters, mesh resolution, and regularization parameters. For the problems considered, GA-NGMRES consistently outperforms Anderson acceleration.Speaker: Andreas Mang (Department of Mathematics, University of Houston)
-
133
-
Matrix Nearness Problems: D Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
136
Constrained Rayleigh quotient optimization and its applications in polynomial eigenvalue problem
Let $G, H_1,\ldots,H_s \in \mathbb C^{n,n}$ be Hermitian and $S_1,\ldots,S_k \in \mathbb C^{n,n}$ be symmetric matrices. In this talk, we maximize the Rayleigh quotient of a Hermitian matrix $H$ under certain constraints involving Hermitian matrices $H_1,\ldots, H_s$ and symmetric matrices $S_1,\ldots, S_k$. More specifically, we compute
$$ \begin{eqnarray} & \mu(G,H_1,\ldots,H_s,S_1,\ldots,S_k):=\sup\Big\{\frac{v^*Gv}{v^*v} :~v\in \mathbb C^{n} \setminus \{0\},\,v^*H_iv=0,\, \nonumber \\ &\hspace{6cm} ~i=0,\ldots,s ,\, v^TS_jv=0,~j=1,\ldots,k \Big\}, \end{eqnarray} $$ where $T$ and $*$ denote the transpose and the conjugate transpose of a matrix or a vector, respectively. Such problems occur in stability analysis of uncertain systems and in the eigenvalue perturbation theory of matrices and matrix polynomials. The problem without symmetric conditions has been studied by [Bora et al., SIMAX, 35(2), 2014] and a computable formula for $\mu(G,H_1,\ldots,H_s)$ was obtained in terms of the largest eigenvalue of a parameter-dependent Hermitian matrix. Similarly, this problem without Hermitian conditions was considered by [Prajapati & Sharma, LAA, 645, 2022] and a computable formula $\mu(G,S_1,\ldots,S_k)$ was obtained in terms of the second largest eigenvalue of some parameter-dependent Hermitian matrix. In this talk, we extend the ideas from [Bora et al., SIMAX, 35(2), 2014] and [Prajapati & Sharma, LAA, 645, 2022] and derive a computable estimation for $\mu(G,H_1,\ldots,H_s,S_1,\ldots,S_k)$ with both Hermitian and symmetric constraints.
This estimation is exact when the eigenvalue at the optimal is simple. We then apply these results in computing the structured eigenvalue backward errors of real T-even and real T-odd matrix polynomials. These results are also shown to be useful in obtaining the eigenvalue backward errors of matrix functions involving both Hermitian and skew-symmetric matrices.Speaker: Punit Sharma (Indian Institute of Technology Delhi - Abu Dhabi, UAE) -
137
Singular Value Characterizations for a Nearest Rectangular Polynomial Matrix with an Eigenvalue
A rectangular polynomial matrix with more columns than rows generically has no eigenvalues. We aim to find a smallest perturbation (with respect to the 2-norm of the concatenated coefficients of the polynomial matrix) so that the perturbed polynomial matrix has an eigenvalue, that is prescribed. This problem is motivated by locating a nearest uncontrollable system for a first-order, as well as a higher-order continuous-time control system. We consider the cases when the polynomial matrix is subject to complex and real perturbations. For both, we derive singular value optimization characterizations that facilitate locating a nearest polynomial matrix with the prescribed eigenvalue. The real perturbation case, inspired by the work of Qiu et al. [Automatica, Vol. 31, pp. 879-890] for the distance to instability of a linear continuous-time system under real perturbations, is much more involved, and leads to a more complicated singular value formula.
We exploit the derived singular value optimization characterizations to compute a nearest rectangular polynomial matrix with an eigenvalue under complex perturbations and under real perturbations. The approaches that we devise are based on level-set methods that date back to Byers, Boyd-Balakrishnan, Bruinsma-Steinbuch for the distance to instability, as well as Lipschitz-continuity based global optimization techniques. They seem to work extremely effectively. For instance, we are able to compute a nearest uncontrollable system for a first-order or higher-order system of medium size in a few seconds.
Speaker: Emre Mengi (Koc University, Istanbul) -
138
A Newton–bisection method with monotone convergence for matrix nearness problems.
In this talk, I will address eigenvalue-optimization-based matrix nearness problems such as the stability radius of a matrix or a time invariant system, $\mathcal{H}^\infty$ norm computation, the structured distance to singularity.
These are formulated here as 2-variable optimization problems of functionals depending either on a single or on several target eigenvalues of the matrix.
It is well-known that the classical Newton-bisection method (proposed, e.g., in [1]) may fail to converge.For this, in 2016, Mitchell and Overton [3] proposed a very effective convergent alternating iteration method, the Hybrid Expansion Contraction (HEC), which was then extended to passivity optimization in [4].
The method we propose here uses a two-level nested iteration, where, at the inner level, a constrained gradient system is integrated to optimize the functional over the set of matrices of specified norm and, at the outer level, to compute the optimal norm.
However, an important modification of the method exploits the free gradient system, allowing for guaranteed convergence (see [2]). A discussion and a comparison wrt the HEC methodology will conclude the talk.References.
[1] N. Guglielmi, M. Gürbüzbalaban, and M.L. Overton.
Fast approximation of the $\mathcal{H}^{\infty}$ norm via optimization over spectral value sets.
SIAM J. Matrix Anal. Appl., 34(2): 709--737, 2013.[2] N. Guglielmi and C. Lubich.
Matrix nearness problems and eigenvalue optimization.
https://arxiv.org/abs/2503.14750, 2025.[3] T. Mitchell and M.~L. Overton.
Hybrid expansion–contraction: a robust scalable method for approximating the $\mathcal{H}^\infty$ norm.
IMA J. Numer. Anal., 36(3): 985--1014, 2016.[4] T. Mitchell and P. Van Dooren.
Root-max problems, hybrid expansion-contraction, and quadratically convergent optimization of passive systems.
SIAM J. Matrix Anal. Appl., 44(2): 753--780, 2023.Speaker: Nicola Guglielmi (Gran Sasso Science Institute)
-
136
-
The Inverse Eigenvalue Problem of a Graph and Zero Forcing: D Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
139
A unified framework for the Expander Mixing Lemma for graphs and its applications
A unified framework of the Expander Mixing Lemma for irregular graphs using adjacency eigenvalues will be presented, as well as several new versions of it. We will also show some of its applications in graph theory, which include spectral bounds on the zero forcing number of a graph. To derive our results we use a new application of weight partitions of graphs, where the Perron eigenvector entries are used as vertex weights as a way to regularise the graph.
Speaker: Aida Abiad (Eindhoven University of Techonolgy) -
140
Inverse Eigenvalue Problems for Graphs: The Weighted Laplacian Case
An inverse eigenvalue problem for a graph (IEP-G) asks a fundamental question: What are the possible spectra for (symmetric) real matrices fitting a given graph? Many have worked on several aspects of the IEP-G with exciting advances and variations appearing over the past forty years. Here, the focus will be on weighted Laplacian matrices associated with a graph. Such matrices are permanently intertwined in combinatorial matrix theory and numerical analysis and form a natural matrix class to study for an inverse eigenvalue problem.
For a given graph G, we aim to determine the possible realizable spectra for a generalized (or weighted) Laplacian matrix associated with G. We present such results for certain families of graphs and graphs on a small number of vertices, including possible ordered multiplicity lists.
Strong matrix properties have been associated with several adaptations of inverse eigenvalue problems and sign pattern eigenvalue problems. We introduce a new strong property, the strong spectral property for weighted Laplacian matrices (SSPWL), and establish its Supergraph and Bifurcation lemmas. We develop a Jacobian method which can be used to verify such a strong property. These tools are then applied to derive potential spectral regions of weighted Laplacian matrices of a graph on four vertices.Speaker: Shaun Fallat (University of Regina)
-
139
-
Numerical Linear Algebra Tools for Model Order Reduction: D Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
141
Least-squares Rational Approximation Using Riemannian Optimization
Vector fitting is a widely used method for least-squares rational approximation that approaches the nonlinear least-squares problem with a sequence of linear least-squares problems. By contrast, the iterative rational Krylov algorithm (IRKA) is a method for a continuous least-squares approximation, originally formulated as a fixed-point iteration and with a recent interpretation as a Riemannian optimization method. In this talk, we discuss the application of Riemannian optimization techniques to the (discrete) least-squares rational approximation problem and compare them with vector fitting.
Speaker: Petar Mlinarić (University of Zagreb) -
142
Extensions of data-driven balancing to LQO and QB systems
We develop the theoretical framework for extending the quadrature based balanced truncation (QuadBT) method to linear systems with quadratic outputs (LQO). QuadBT which was originally designed for data-driven balanced truncation of standard linear systems with linear outputs only. We show that by sampling the extended impulse responses (kernels) and their derivatives (in the time domain) or the corresponding transfer functions (in the frequency domain), we can construct a reduced order model that mimics the approximation quality of intrusive balanced truncation. At its core, the method can be interpreted as an implicit sampling of the Gramians of the system using input-output observations. We demonstrate a proof of concept for the proposed framework on an example using numerically evaluated data. We also briefly discuss the extension of the QuadBT method for the more involved quadratic-bilinear (QB) system case.
Speaker: Reetish Padhi (Virginia Tech) -
143
PIRKA: The Iterative Rational Krylov Algorithm for Linear Time-Periodic Systems
We consider single-input, single-output systems with time-varying, periodic parameters:
ẋ(t) = A(t)x(t) + b(t)u(t),
y(t) = c(t) x(t),
where A(t) ∈ ℝⁿˣⁿ and b(t), c(t) ∈ ℝⁿ all have period T.
Such systems arise when modeling phenomena in fluid dynamics, structural mechanics, and electronic circuits. In particular, linearization around known periodic orbits of a nonlinear model produces a periodic system of partial differential equations, and subsequent spatial semi-discretization yields large-scale linear time-periodic (LTP) dynamical systems. The need to simulate system responses to a variety of inputs motivates the development of effective model reduction tools for these systems.
While research on model reduction for LTP systems is relatively limited, there is a substantial body of literature devoted to control, spectral analysis, and harmonic response of LTP systems. From this literature arise the Harmonic Transfer Function and the H₂ norm for LTP systems. These concepts lead to necessary conditions for an H₂-optimal reduced-order model. In the linear time-invariant case, the Iterative Rational Krylov Algorithm (IRKA) is a standard approach for the H₂ model reduction problem. The work presented here extends IRKA to the LTP setting.Speaker: Sam Bender
-
141
-
Where Algebraic Coding Theory and Graph Theory Meet: D Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
144
Secure Distributed Matrix Multiplication
Matrix multiplication is, oftentimes, the most expensive computational task in an algorithm. It is the computational bottleneck for training many of the now well-celebrated learning algorithms, for example. To speed up the algorithm, the data can be distributed on many machines to perform the computations in parallel. This sharing of the data, however, raises security concerns when the data is sensitive and has to remain private, such as financial or medical data. Secure distributed matrix multiplication (SDMM) studies how to parallelize matrix multiplication while keeping the data secure.
Speaker: Rafael D'Oliveira (Clemson University) -
145
Graph Properties of Codes from Dyadic and Quasi-Dyadic Matrices
Dyadic matrices are a subclass of matrices known as reproducible matrices, where the entries of the matrix are completely determined by their first row. Quasi-dyadic matrices are block matrices with dyadic matrices in the blocks.
There has been extensive work analyzing quasi-cyclic codes, codes defined by quasi-cyclic parity check matrices, but less is known about codes arising from dyadic or quasi-dyadic matrices. In this work, we present results on the properties of the Tanner graphs of dyadic and quasi-dyadic codes. In particular, we present results on their isomorphism classes, absorbing sets, and conditions for constructing quasi-dyadic matrices with Tanner graphs of a certain girth. We also compare the performance of these code constructions under Belief Propagation decoding.
Speaker: Dr Kirsten Morris (Virginia Tech) -
146
External codes for multiple unicast networks via interference alignment
We introduce a formal framework to study the multiple unicast problem for a coded network in which the network code is linear over a finite field and fixed. We show that the problem corresponds to an interference alignment problem over a finite field. In this context, we establish an outer bound for the achievable rate region and provide examples of networks where the bound is sharp. We finally give evidence of the crucial role played by the field characteristic in the problem.
Speaker: Felice Manganiello (Clemson University)
-
144
-
Algebraic Invariants of Graphs: D Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
147
Spectrum of trees of diameter 4 for the distance Laplacian matrix
Graphs can be encoded into a matrix according to some rule. The eigenvalues of the matrix are used to understand the structural properties of graphs. If two graphs share a set of eigenvalues, they are called cospectral. A tree is a graph with no cycles, and for most matrix representations, almost all trees have a cospectral mate. The distance Laplacian matrix is found by subtracting the distance matrix from the diagonal transmission matrix. There is an open conjecture that trees, for their distance Laplacian matrices, do not have a cospectral mate and are therefore spectrally determined. We explicitly find the spectrum of a family of trees of diameter 4 and how it is uniquely defined by its parameters.
Speaker: Kate Lorenzen (Linfield University) -
148
Advances on Brouwer's Conjecture
Within Spectral Graph Theory, Brouwer’s Conjecture (BC) is a fundamental problem concerning Laplacian eigenvalues and graph invariants. It proposes a relationship between the sum of the largest Laplacian eigenvalues of a graph and its number of edges, with direct implications for the study of Laplacian energy. More precisely, for a graph ( G = (V, E) ) with ( n = |V| ) vertices and ( m = |E| ) edges, the conjecture states that, for ( k = 1, \ldots, n ), the sum ( S(k) ) of the ( k ) largest eigenvalues of the Laplacian matrix of ( G ) satisfies
S(k) ≤ m + k(k − 1)/2.
In this talk, we discuss the role of Brouwer’s Conjecture in the area of Spectral Graph Theory and describe recent progress on the problem, highlighting results and contributions obtained in this framework.
Speaker: Prof. Vilmar Trevisan (UFRGS - Brazil) -
149
A representation theoretical approach to the p-rank of subset incidence matrices
Let $W^i_{k,n}(m)$ denote a matrix with rows and columns indexed
by the $k$-subsets and $n$-subsets, respectively, of an $m$-element set. The row $S$, column $T$ entry of $W^i_{k,n}(m)$ is 1 if $|S \cap T|= i$, and is 0 otherwise. When $i=k$ the matrix $W^k_{k,n}(m)$ is the subset inclusion matrix for which Wilson found a diagonal form, solving the $p$-rank problem for any prime $p$. This diagonal form was used to calculate the Smith group of the hypercube graph.We compute the rank of the matrix $W^1_{2,n}(m)$ over any field by making use of the representation theory of the symmetric group. We also give a simple condition under which $W^i_{k,n}(m)$ has large $p$-rank.
Speaker: Colby Sherwood (University of Delaware)
-
147
-
Low-Complexity Data-driven or Classical Algorithms and Applications: D Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
150
Conditional Denoising Diffusion Model-Based Robust MR Image Reconstruction from Highly Undersampled Data
Magnetic Resonance Imaging (MRI) is a critical tool in modern medical diagnostics, yet its prolonged acquisition time remains a critical limitation, especially in time-sensitive clinical scenarios. While undersampling strategies can accelerate image acquisition, they often result in image artifacts and degraded quality. Recent diffusion models have shown promise for reconstructing high-fidelity images from undersampled data by learning powerful image priors; however, most existing approaches either (i) rely on unsupervised score functions without paired supervision or (ii) apply data consistency only as a post-processing step. In this work, we introduce a conditional denoising diffusion framework with iterative data consistency correction, which differs from prior methods by embedding the measurement model directly into every reverse diffusion step and training the model on paired undersampled–ground truth data. This hybrid design bridges generative flexibility with explicit enforcement of MRI physics. Experiments on the fastMRI dataset demonstrate that our framework consistently outperforms recent state-of-the-art deep learning and diffusion-based methods in SSIM, PSNR, and LPIPS, with LPIPS capturing perceptual improvements more faithfully. Specifically, under an acceleration factor of 8 and Gaussian 1D sampling, the proposed model achieves SSIM = 0.834 ± 0.063, PSNR = 32.52 ± 2.63 dB, and LPIPS = 0.063 ± 0.029. These results demonstrate that integrating conditional supervision with iterative consistency updates yields substantial improvements in both pixel-level fidelity and perceptual realism, establishing a principled and practical advance toward robust, accelerated MRI reconstruction.
Speaker: Dr Xianqi Li (Florida Institute of Technology) -
151
A low-complexity locally recoverable matrix-influenced algorithm to globally recover codes
Code Recovery using algebraic-geometric approaches becomes computationally expensive with the cardinality of the field and the complexity of the code structures. In response, we present a low-complexity algorithm that utilizes structures in algebraic-geometric codes over finite fields. The low-complexity algorithm recovers algebraic codes over finite fields locally, which we name as lrc algorithm. The lrc algorithm is derived based on a sparse matrix factorization of the inverse of an $(r+1) \times r$ locally recoverable matrix, whose elements are defined by the points on the surface in $\mathbb{P}^3$ over the finite field $\mathbb{F}_q$ having locality $r$. We will show that the lrc algorithm reduces the complexity from $\mathcal{O}(n^3)$ to $\mathcal{O}(n \log n)$ for the $n=2^s(s \geq 1) > r$ length codeword. Next, we utilize an extended lrc algorithm followed by a structured neural network (StNN) to globally recover codes over the surface $\mathbb{P}^3$ over the finite field $\mathbb{F}_q$. We discuss two StNNs, i.e., DFT-StNN and DCT-StNN, based on the factorization of the locally recoverable matrix to globally recover codes. Numerical simulations will be shown to compare the performance of the extended lrc algorithm in global recovery codes, with brute-force calculation, DFT-StNN, DCT-StNN, and a feedforward neural network for codewords with lengths from $n=6, 12, 27, 48, 96$ and $210$, having locality $2$ for points on the surface $\mathbb{P}^3$ over the finite field $\mathbb{F}_q$. Our empirical results demonstrate that the extended lrc algorithm achieves the lowest flops, the highest accuracy with an error order $10^{-16}$ compared to brute-force calculations, DFT-StNN, DCT-StNN, and the feedforward neural network, showing the existence of local to global codeword setting.
Speaker: Oshani Jayawardane (Embry-Riddle Aeronautical University) -
152
Computing Singular Values Above a Certain Threshold
In this talk we discuss various computational aspects of determining all singular triplets $\{\sigma_j, u_j, v_j\}$ corresponding to singular values of $A$ above a user-specified threshold parameter $sigma$, or in other words, determining a $k$-PSVD of $A$ such that $\sigma_k \geq sigma$ and $\sigma_{k+1} < sigma$. While various numerical schemes with publicly available software have been developed for finding an approximate $k$–PSVD of a large sparse matrix, e.g., PROPACK, ${\tt irlba}$, ${\tt primme\_svds}$, ${\tt rsvd}$, and MATLAB’s ${\tt svds}$, their use is somewhat limited in this context as they all require knowing the number of desired singular triplets, $k$, in advance.
We propose a hybrid algorithm that works directly on $A$ and is based on the explicit deflation procedure applied to Golub-Kahan-Lanczos Bidiagonalization (GKLB)-based methods. An important feature of GKLB-methods that is relevant here is that they can be thought of as a one–sided PSVD approximations of $A$ -- this allows for the explicit deflation technique to be applied only to either the left or right singular vectors of $A$, depending on its dimensions, and thus significantly reducing the overall computational cost. Due to iterative deflation process, a modest non–zero error is added to the side in the GKLB method that had theoretically zero error. Although this error growth is theoretically manageable, our numerical experiments showed that at times one has to restore the one–sided GKLB structure which we have achieved by combining the GKLB-based method with a variant of the block SVD power method. Implementations of this algorithm are provided in MATLAB and a widely used statistical software ${\sf R}$ along with alternative thresholding criteria based on the notions of energy and the normalized root mean squared error.
Speaker: Vasilije Perovic (University of Rhode Island)
-
150
-
Approximate Computing in Numerical Linear Algebra: D Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
153
Forward and backward error bounds for a mixed precision preconditioned conjugate gradient algorithm
The preconditioned conjugate gradient (PCG) algorithm is one of the most popular algorithms for solving large-scale linear systems $Ax = b$, where $A$ is a symmetric positive definite matrix. Rather than computing residuals directly, it updates the residual vectors recursively. Current analyses of the conjugate gradient (CG) algorithm in finite precision typically assume that the norm of the recursively updated residual goes orders of magnitude below the machine precision, focusing mainly on bounding the residual gap thereafter. This work introduces a framework for the PCG algorithm and provides rigorous proofs that the relative backward and forward errors of the computed results of PCG can reach the levels $O(u)$ and $O(u)\kappa(A)^{1/2}$, respectively, after a sufficient number of iterations without relying on an assumption concerning the norm of the recursively updated residual, where $u$ represents the unit roundoff and $\kappa(A)$ is the condition number of $A$. Our PCG framework further shows that applying preconditioners in low precision does not compromise the accuracy of the final results, provided that reasonable conditions are satisfied. Our theoretical results are illustrated through a set of numerical experiments.
Speaker: Yuxin Ma (Charles University) -
154
Mixed Precision General Alternating-Direction Implicit Method for Solving Large Sparse Linear Systems
Our work presents a novel mixed precision formulation of the General Alternating-Direction Implicit (GADI) method, designed to accelerate the solution of large-scale sparse linear systems $Ax=b$. By solving the computationally intensive subsystems in low precision (e.g., Bfloat16 or FP32) and performing residual and solution updates in high precision, the proposed method significantly reduces execution time without compromising the final solution accuracy. We provide a comprehensive rounding error analysis that establishes convergence rates and limiting accuracies under mixed precision arithmetic. Additionally, we introduce a robust strategy based on Gaussian Process Regression (GPR) for selecting the optimal regularization parameter $\alpha$. Performance benchmarks on an NVIDIA A100 GPU demonstrate that mixed precision GADI achieves speedups of up to 3.1× compared to standard double precision implementations on large-scale convection-diffusion and reaction-diffusion problems.
Speaker: Juan Zhang -
155
Analysis of Floating-Point Matrix Multiplication Computed via Integer Arithmetic
Ootomo, Ozaki, and Yokota [Int. J. High Perform. Comput. Appl., 38 (2024), p. 297–313] have proposed a strategy to recast a floating-point matrix multiplication in terms of integer matrix products. The factors $A$ and $B$ are split into integer slices, the product of these slices is computed exactly, and $AB$ is approximated by accumulating these integer products in floating-point arithmetic. This technique is particularly well suited to mixed-precision matrix multiply–accumulate units with integer support, such as the NVIDIA tensor cores or the AMD matrix cores The number of slices allows for performance-accuracy tradeoffs: more slices yield better accuracy but require more multiplications, which in turn reduce performance. We propose an inexpensive way to estimate the minimum number of multiplications needed to achieve a prescribed level of accuracy. Our error analysis shows that the algorithm may become inaccurate (or inefficient) if rows of $A$ or columns of $B$ are badly scaled. We perform a range of numerical experiments, both in simulation and on the latest NVIDIA GPUs, that confirm the analysis and illustrate strengths and weaknesses of the algorithm.
Speaker: Mantas Mikaitis (University of Leeds)
-
153
-
Linear Algebra Education: D Torgersen Hall 2150
Torgersen Hall 2150
Virginia Tech
-
156
A few issues in teaching linear algebra
Linear algebra is fundamental to mathematics and has wide-ranging applications in science, engineering, and data analysis. As a core mathematical subject, it is taught to students at both the high school and college levels, not only in mathematics but also across the sciences. Is linear algebra easy? It may not be as easy as it sounds. Is teaching linear algebra easy? It may not be as easy as one thinks. Seemingly simple questions often reveal subtle conceptual issues: for example, what is the degree of the zero polynomial, and why? If every subset of a vector space spans a subspace, what does the empty set span? Are the eigenvalues of a matrix continuous functions of its entries?
This talk addresses several issues in linear algebra arising in both teaching and research. Such issues are frequently overlooked or treated carelessly in textbooks and classroom instruction.
Speaker: Fuzhen Zhang (Nova Southeastern University) -
157
Linear algebra applications in students’ post linear algebra course work
Linear algebra enrollments are increasingly comprised of students from applied STEM majors. Little research has systematically examined the ways in which the curricular resources commonly used in first linear algebra courses relate to the linear algebra applications encountered by students from applied fields in their subsequent coursework and career fields. In this talk, we will summarize findings from interviews with instructors of STEM classes that list a first course in linear algebra as a pre-requisite course. Major areas of focus include computer engineering, electrical engineering, industrial engineering, economics, statistics, computational science, and biomathematics. Interviews focus applied STEM fields instructors’ perceptions of how linear algebra is applied in their fields, topics they need to review or teach to cover their own course content, which concepts are most important more broadly in their field, and the kinds of notation and computational tools needed and used in their disciplines. We also ask instructors to reflect on example applications problems taken from linear algebra textbooks and reflect on those from the lens of their own discipline and students’ learning in that discipline. We will provide a preliminary summary of findings from our interviews, relate these to current research on the teaching and learning of linear algebra, and discuss how our findings can inform curriculum development and enactment in the context of linear algebra.
Speaker: Christine Andrews-Larson (Florida State University) -
158
Promoting Linear Algebraic Reasoning Among Students: Affordances and Challenges
As students transition from the mathematics they learn in school years, including their first-year calculus courses, to the first course in linear algebra, they experience discontinuities in their perspective of what mathematics is. Their propensity to continue applying the same habits of learning in the face of this change leads to failure and frustration. The failure manifests itself in the quality of understanding basic concepts as well as in the lack of linear-algebraic reasoning. Instructional treatments applied in my teaching experiments to foster students’ ability to reason linear algebraically resulted in mixed success—some of the treatments were successful, others less so. The latter are accounted for by the structural complexity of the subject matter and students’ background knowledge. The pedagogical approaches offered in this talk are oriented within a particular theoretical framework for the learning and teaching of mathematics, called DNR-based instruction in mathematics. Reflections and broader implications are addressed through the lenses of the DNR framework.
Speaker: Guershon Harel (University of California, San Diego)
-
156
-
Sparse Tensor Computations: Algorithms and Applications: D Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
159
Efficient oversampled Tensor-Train approximations
Recently, there have been many advances in the area of randomized and sampling-based methods for data approximation. This has led to significant progress towards the efficient treatment of large data in both compression and utilization in computation. In this talk, I will discuss a current work that uses random oversampling on a Tensor Train Cross (TT-Cross) approximation in order to reduce the observed error of a tensor approximation. This work includes two separate formulations utilizing projection techniques to construct tensor cores with results demonstrating a reduction in error. Along with the observed reduction in error, we will show that in practice, the oversampling procedure does
not substantially increase computation time compared to a standard TT-Cross construction.Speaker: Daniel Hayes (University of Delaware) -
160
Adaptive Randomized Tensor Train Rounding using Khatri-Rao Products
The Tensor Train (TT) format provides a compact and scalable way to represent high-dimensional tensors, making it essential for solving certain parametrized partial differential equations and other large-scale problems. A critical operation in TT-based computations is rounding, which reduces the ranks of a tensor in TT format to maintain efficiency. While recent randomized rounding algorithms have improved performance over traditional deterministic approaches, they typically require manual rank specification, limiting their practicality in many applications.
In this talk, I will present a new adaptive randomized TT-rounding algorithm based on Khatri-Rao random projections. Unlike previous methods, our algorithm automatically determines the TT ranks needed to meet a user-specified approximation tolerance, eliminating the need for manual tuning. It achieves up to 45× speedup over deterministic rounding and a 2× improvement over the fastest existing randomized approach (which requires specifying target ranks), particularly in the context of rounding sums of TT tensors—a key bottleneck in adaptation of GMRES to vectors in TT format. I will discuss the algorithm's design, performance benefits, and its implications for high-dimensional numerical computing.
Speaker: Bhisham Dev Verma (Wake Forest University)
-
159
-
Quantum Numerical Linear Algebra: D Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
161
Efficient Encoding of Semiseparable Matrices in Quantum Circuits
Quantum block encoding (QBE) is a crucial step in the development of most quantum algorithms, providing an embedding of a given matrix into a suitable larger unitary matrix. Efficient techniques for QBE have primarily focused on sparse matrices, and less effort has been devoted to data-sparse matrices, such as rank-structured matrices.
In this talk, we examine a specific case of rank structure: one-pair semiseparable matrices. We present a new block encoding approach that relies on an efficient state preparation technique for the representation of the two generators and on the reordering of entries obtained through the tensor product of the generators. This process takes polylogarithmic time if the two generators can be prepared efficiently, with a scaling factor of $2\sqrt{2}^{n}$, where $n=log_2(N)$ is the number of qubits to represent the matrix.
Speaker: Giacomo Antonioli -
162
Efficient LCU block encodings through Dicke states preparation
With the Quantum Singular Value Transformation (QSVT) emerging as a unifying framework for diverse quantum speedups, the efficient construction of block encodings—their fundamental input model—has become increasingly crucial. However, devising explicit block encoding circuits remains a significant challenge. A widely adopted strategy is the Linear Combination of Unitaries (LCU) method. While general, its practical utility is often limited by substantial gate overhead. To address this, we introduce the Fast One-Qubit-Controlled Select LCU (FOQCS-LCU), a compact LCU formulation that requires only a linear number of ancilla qubits and is explicitly decomposed into one- and two-qubit gates. By exploiting the underlying Hamiltonian structure, we design a parametrized family of efficient Dicke state preparation routines, enabling systematic realization of the state preparation oracle at substantially reduced gate cost. The check matrix formalism further yields a constant-depth SELECT oracle, implemented as two fully parallelizable layers of singly controlled Pauli gates. We construct explicit block encoding circuits for representative spin models such as the Heisenberg and spin glass Hamiltonians and provide detailed, non-asymptotic gate counts. Our numerical benchmarks confirm the efficiency of the FOQCS-LCU approach, illustrating over an order-of-magnitude reduction in CNOT count compared to conventional LCU. This framework opens a pathway toward practical, low-depth block encodings for a broad class of structured matrices beyond those considered here.
Speaker: Roel Van Beeumen (Lawrence Berkeley National Laboratory) -
163
Practical block encodings of matrix polynomials that can also be trivially controlled
Quantum circuits naturally implement unitary operations on input quantum states. However, non-unitary operations can also be implemented through “block encodings”, where additional ancilla qubits are introduced and later measured. While block encoding has a number of well-established theoretical applications, its practical implementation has been prohibitively expensive for current quantum hardware. In this paper, we present practical and explicit block encoding circuits implementing matrix polynomial transformations of a target matrix. With standard approaches, blockencoding a degree-$d$ matrix polynomial requires a circuit depth scaling as $d$ times the depth for block-encoding the original matrix alone. By leveraging the recently introduced Fast One-Qubit Controlled Select LCU (FOQCS-LCU) framework, we show that the additional circuit-depth overhead required for encoding matrix polynomials can be reduced to scale linearly in d with no dependence on system size or the cost of block encoding the original matrix. Moreover, we demonstrate that the FOQCS-LCU circuits and their associated matrix polynomial transformations can be controlled with negligible overhead, enabling efficient applications such as Hadamard tests. Finally, we provide explicit circuits for representative spin models, together with detailed non-asymptotic gate counts and circuit depths.
Speaker: Filippo Della Chiara (KU Leuven)
-
161
-
12:15 PM
Lunch Break
-
Contributed Talks: E Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
164
Order automorphisms of effect algebras
The general form of order automorphisms of effect algebras has been known in the complex case. We present a much simpler proof based on projective geometry which works also in the real case.
Speaker: Peter Semrl (Institute of Mathematics, Physics, and Mechanics, Ljubljana, Slovenia) -
165
Polynomial Factorization via Matrix Representations
Polynomial factorization is classically studied within commutative polynomial rings, where irreducibility is an intrinsic algebraic property. In this talk, we present a linear-algebraic approach to factorization via entangled polynomial rings, in which polynomials are represented by structured matrices and analyzed using tools from matrix theory.
By embedding a polynomial into a family of $m \times m$ matrices (through $m$-nomials and their finite representations as Szab\'o rings), questions of reducibility are translated into questions about matrix factorizations. In this setting, reducibility is detected using the $m$-terminant, defined as the determinant of the associated matrix. Circulant matrices and their eigenvalues play a central role in determining when such factorizations occur.
We focus on the polynomials
$$ f_p(x) = 1 + x + x^2 + \cdots + x^{p-1}, $$ which are irreducible in $\mathbb{Q}[x]$ for prime $p$. Using determinant and eigenvalue computations, we show that while these polynomials remain irreducible for all $m < p$, their matrix representations become reducible at $m = p$. This leads to the notion of \emph{valence}, the minimal matrix size at which reducibility occurs, and establishes that the valence of $f_p(x)$ is exactly $p$ for all primes $p \ge 3$.This perspective reframes polynomial factorization as a problem in linear algebra over rings, illustrating how matrix representations, determinants, and eigenvalues uncover hidden algebraic structure beyond the classical setting.
Speaker: Dr Azam Mozaffarikhah (Virginia Tech, Department of Mathematics) -
166
Pretty good state transfer in Grover walks on abelian Cayley graphs
In this work, we study pretty good state transfer (PGST) in Grover walks on graphs. We consider the transfer of quantum states localized at the vertices of a graph and use Chebyshev polynomials to analyze PGST between such states. In general, we find a necessary and sufficient condition for the occurrence of PGST on graphs. We then focus our analysis on abelian Cayley graphs and derive a necessary and sufficient condition for the occurrence of PGST on such graphs. Consequently, we obtain a complete characterization of PGST on unitary Cayley graphs. Our results yield infinite families of graphs that exhibit PGST yet do not exhibit perfect state transfer.
Speaker: Mr Koushik Bhakta (Indian Institute of Technology Guwahati)
-
164
-
Matrix Inequalities, Matrix Equations, and Their Applications: E Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
167
On the stability criteria via finite Hankel matrices for regular matrix polynomials
This talk focuses on several stability criteria via Markov parameters for regular matrix polynomials, which generalize the corresponding criteria constrained by the monic assumption. The testing framework employs two finite Hankel matrices, whose rectangular blocks are the submatrices of the Markov parameters redefined through a column-wise splitting and column reduction for matrix polynomials. Specifically, a Hurwitz stability criterion is characterized by the Hermitian positive definiteness of two finite Hankel matrices. Further, our stability criterion is derived in terms of the Hermitian nonnegative definiteness and nullity of these finite Hankel matrices.
Speaker: Dr Xuzhou Zhan (Beijing Normal University at Zhuhai) -
168
Sharp lower bounds for generalized operator products
We consider general bilinear products parameterized by positive semidefinite matrices. Typically non-commutative, non-associative, and non-unital, these products preserve positivity and include the classical Hadamard, Kronecker, and convolutional products as special cases. We prove that every such product satisfies a sharp nonzero lower bound in the Loewner order, generalizing previous results of Vybíral and Khare that were obtained in the special case of the Hadamard product. Our results naturally extend to Hilbert spaces for a family of products parametrized by positive trace-class operators, providing a lower bound in the Loewner order for such general products, including for the Hilbert tensor product.
(Joint work with Javad Mashreghi and Prateek Kumar Vishwakarma.)
Speaker: Dominique Guillot (University of Delaware) -
169
Convolution of matrices: Cayley–Hamilton theory, matrix transforms, and positivity preservers, with connections to the Bruhat order
Classical approaches to matrix function theory — i.e., extending scalar functions to matrices — are largely organized around two frameworks: entrywise calculus via the Schur (Hadamard) product, and functional calculus via the spectral theorem. In this talk, I present a third, fundamentally different framework based on matrix convolution, in which convolution itself is viewed as a matrix multiplication and gives rise to a new class of matrix transforms.
This perspective leads to a systematic theory of convolution-based transforms and their positivity preserving properties. I will describe convolutional analogues of the classical results of Pólya–Szegő, Schoenberg, Rudin, Loewner, and Horn, including a Schoenberg-type classification theorem showing that absolute monotonicity is again the correct notion for positivity preservers across all dimensions, together with refined results in fixed matrix dimensions.
Beyond positivity, I will highlight two structural features that distinguish convolution from both standard and entrywise products: a Cayley–Hamilton-type theorem for convolution with optimal degree, and a novel polynomial–matrix identity underlying the associated transforms. If time permits, I will also briefly describe an unexpected connection with the Bruhat order on the symmetric group, and an application to sums of discrete random variables, illustrating new links between matrix analysis, algebraic combinatorics, and probability.
(Based on joint work with Javad Mashreghi and Mostafa Nasri.)
Speaker: Prateek Kumar Vishwakarma (Universite Laval, Quebec, Canada)
-
167
-
The Inverse Eigenvalue Problem of a Graph and Zero Forcing: E Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
170
Sign patterns that require or allow the non-symmetric strong spectral property
A matrix $A$ has the non-symmetric strong spectral property (nSSP) if $X=O$ is the only matrix which satisfies $A\circ X=O$ and $AX^T=X^TA$. This property comes with implications for eigenvalue properties of sign patterns, including a bifurcation lemma and superpattern lemma. We describe some classes of sign patterns for which every matrix with the sign pattern will have the nSSP. We also provide methods for recognizing when a sign pattern restricts matrices from having the nSSP. In particular, we provide a characterization of the sign patterns that allow a matrix with the nSSP. This presentation is based on joint work with Michael Cavers and Zhongshan Li.
Speaker: Kevin Vander Meulen (Redeemer University) -
171
Jacobian method and strong properties
A sign pattern is a matrix whose entries are in $\{+, -, 0\}$, while its quantitative class is the set of real matrices whose entries match the corresponding signs. A sign pattern is said to be spectrally arbitrary if its quantitative class contains matrices demonstrating all possible monic real polynomials as the characteristic polynomials. Historically, there are the Jacobian method, the nilpotent-centralizer method, and the nonsymmetric strong spectral property (SSP, which is equivalent to the similarity-transversality property) to witness a spectrally arbitrary sign pattern. However, their relations are unclear. In this talk, we survey these methods and show that they are equivalent when the minimal polynomial has degree $n$.
Speaker: Jephian C.-H. Lin (National Yang Ming Chiao Tung University) -
172
A general strong property for IEP-G
The Inverse Eigenvalue Problem for a Graph (IEP-G) asks for the possible spectra of a real symmetric matrix knowing only which off-diagonal entries are non-zero, as described by a graph $G$. Three matrix properties, collectively called the “strong properties”, have become prominent in the study of this problem, due in part to their good behavior with respect to edge deletion and contraction within $G$.
The three properties differ in what eigenvalue data is preserved:- For the Strong Arnold Property (SAP), a single multiplicity.
- For the Strong Multiplicity Property (SMP), all multiplicities.
- For the Strong Spectral Property (SSP), all multiplicities and locations.
In most cases, a question that calls for a strong property will be served well by one of these three perspectives. There does, however, exist a more general perspective that contains each of these as a special case, here introduced as the General Strong Property (GSP). For the GSP, any chosen subset of eigenvalue multiplicities can be preserved, and eigenvalue locations can be allowed to vary locally subject to any set of linear constraints. The usual strong property consequences follow, which include: an algebraic definition, a verification matrix, superpatterns, edge decontraction, matrix liberation, and bifurcation.
In addition, the null space characterization of SAP given by H. van der Holst is extended to an eigenspace characterization of GSP, including the special cases of SMP and SSP, whose eigenspace characterization has not appeared in the literature.
Speaker: H Tracy Hall (none)
-
170
-
Theoretical Advances in Operator Learning: E McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
173
Learning operators with continuous spectrum from data
Linear operators with a continuous spectrum often lurk behind complex physical phenomena in nature, from wave attractors in geometrically confined fluids to topological bifurcations in dynamical systems. However, they are notoriously tricky to learn from data. For example, finite-dimensional approximations of the operator must “discretize” the continuous spectrum into finitely many points and cannot converge uniformly.
In this talk, we explore two principled approaches to data-driven approximation of unitary operators with continuous spectrum. The first approach synthesizes recent advances in computational spectral theory with stable algorithms for Gauss-Szego quadrature rules on the unit circle. The second approach leverages new algorithms for quadrature rules that adaptively deform and discretize the contour inside the unit circle, resembling classical resonance expansions for wave operators. We discuss strong convergence rates for the approximate operators, data-driven aspects of the approximation, and highlight the unifying power of quadrature rules derived from rational approximations.
Speaker: Andrew Horning (Rensselaer Polytechnic Institute) -
174
Resolvent compactification methods for spectral approximation of Koopman operators
Koopman operators and transfer operators represent nonlinear dynamics in state space through its induced action on linear spaces of observables and measures, respectively. This framework enables the use of linear operator theory for analysis and modeling of nonlinear dynamical systems, and has received considerable interest over the years from mathematical, computational, and domain-scientific disciplines. In this talk, we present data-driven techniques for spectral approximation of Koopman and transfer operators of measure-preserving ergodic flows that are based on operators with compact resolvent. Our approach performs a bounded transformation of the Koopman generator (an operator implementing directional derivatives of observables along the dynamical flow), followed by smoothing by a Markov semigroup of kernel integral operators. This results in a skew-adjoint, compact operator with compact resolvent whose eigendecomposition is expressible as a variational generalized eigenvalue problem amenable to approximation using Galerkin methods. A key aspect of these methods is that they are physics-informed, in the sense of making direct use of dynamical vector field information through automatic differentiation of kernel functions. Solutions of the eigenvalue problem reconstruct evolution operators that preserve unitarity of the underlying Koopman group while spectrally converging to it in a suitable limit. In addition, the computed eigenfunctions have representatives in a reproducing kernel Hilbert space, enabling out-of-sample evaluation of learned dynamical features. We illustrate this method with numerical experiments on integrable and chaotic systems.
Speaker: Trevor Camper (Dartmouth College) -
175
Learning Enhanced Ensemble Filters: Continuum Limits of Attention on Measures
Many classic methods in data assimilation, like the Ensemble Kalman Filter (EnKF), are limited by its Gaussian ansatz. In this work, we frame the filtering update as learning a nonlinear operator mapping between probability distributions in the mean-field limit. We introduce Measure Neural Mappings (MNMs), a class of neural operators acting on probability measures, implemented via Set Transformers. A central theoretical contribution is establishing the continuum limit for attention mechanisms on measures. We prove that attention layers applied to finite ensembles are consistent with their continuous-measure counterparts, converging in Wasserstein distance as the sample size approaches infinity. This result rigorously links finite-dimensional linear algebra operations to infinite-dimensional operator learning, justifying the use of a single parameterization across different ensemble sizes. We demonstrate that the resulting MNM-enhanced ensemble filter (MNMEF) achieves superior accuracy on chaotic dynamical systems, including the Lorenz '96 and Kuramoto-Sivashinsky models, outperforming leading filtering methods.
Speaker: Bohan Chen (California Institute of Technology)
-
173
-
Computational Advances in Discrete Inverse Problems: E Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
176
Kronecker Rank Bounds for Operator Matrices
Block-structured matrices arise as operators in many contexts, including image deblurring and discretized differential equations. These matrices are often large and computationally difficult to work with. By rewriting these operators as a sum of Kronecker products, we may be able to alleviate these challenges. In this talk, we show how we can use the structure of a matrix to impose bounds on such a sum.
Speaker: Malena Espanol (Arizona State University) -
177
IterativeCUR: One Small Sketch for Big Matrix Approximations
The computation of accurate low-rank matrix approximations is central to improving the scalability of various techniques in machine learning, uncertainty quantification, and control. Traditionally, low-rank approximations are constructed using SVD-based approaches such as truncated SVD or RandomizedSVD. Although these SVD approaches---especially RandomizedSVD---have proven to be very computationally efficient, other low-rank approximation methods can offer even greater performance. One such approach is the CUR decomposition, which forms a low-rank approximation using direct row and column subsets of a matrix. Because CUR uses direct matrix subsets, it is also often better able to preserve native matrix structures like sparsity or non-negativity than SVD-based approaches and can facilitate data interpretation in many contexts. This paper introduces IterativeCUR, which draws on previous work in randomized numerical linear algebra to build a new algorithm that is highly competitive compared to prior work: (1) It is adaptive in the sense that it takes as an input parameter the desired tolerance, rather than an a priori guess of the numerical rank. (2) It typically runs significantly faster than both existing CUR algorithms and techniques such as RandomizedSVD, in particular when these methods are run in an adaptive rank mode. Its asymptotic complexity is $\mathcal{O}(mn + (m+n)r^2 + r^3)$ for an $m\times n$ matrix of numerical rank $r$. (3) It relies on a single small sketch from the matrix that is successively downdated as the algorithm proceeds.
Speaker: Nathaniel Pritchard (The University of Oxford) -
178
Regularized Krylov Subspace Methods in Low Precision
In recent years a substantial amount of work has been done on developing mixed-precision algorithms for linear systems, methods that can exploit capabilities of modern GPU architectures. However, very little work has been done for ill-conditioned problems that arise from large-scale inverse problems. Special considerations, which normally do not arise when solving well-conditioned problems, such as incorporating regularization into the developed methods, need to be considered. In this talk we consider the use of low-precision, regularized preconditioners in Krylov subspace methods.
Speaker: James Nagy (Emory University)
-
176
-
Linear Algebra Foundations for Data-driven Modeling and Model Order Reduction: E Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
179
Numerical linear algebra for data driven nonlinear dynamics
The Dynamic Mode Decomposition (DMD) is a powerful and versatile numerical method for data driven analysis of nonlinear dynamical systems, with a wide spectrum of applications. It can be used for model order reduction, analysis of latent structures in the dynamics, and e.g. for forecasting and control. The theoretical bedrock upon which the more general Extended DMD (EDMD) framework is built is the Koopman composition operator. The EDMD can be described as a data driven Rayleigh-Ritz extraction of spectral information of a Koopman operator associated with the underlying dynamical system. The nonlinear data snapshots are represented using the eigenvectors of the operator resulting in a modal decomposition KMD (Koopman Mode Decomposition). This becomes a model order reduction tool that represents the nonlinear dynamics using selected eigenpairs. It can be used to reveal coherent states and for forecasting.
The numerical realization of the Koopman operator framework (i.e. DMD and EDMD) for computational analysis of nonlinear dynamics is an excellent illustration of the importance and power of numerical linear algebra, and an instructive case study of the software development based on the design principles introduced in the state of the art software packages such as e.g. LAPACK. The main ingredients of all variants of the DMD are the SVD and low rank approximations,the QR factorization and the orthogonal projections, structured least squares approximations,
and approximations of eigenvalues and eigenvectors from subspaces.
The problem of illconditioned eigenvectors is solved using a Koopman-Schur decomposition, based on unitary transformations. The analysis in terms of the eigenvectors is replaced with a modal decomposition in terms of flag of invariant subspaces that correspond to selected eigenvalues. Accuracy is controlled using computable residuals.
We will discuss these and other fine details that are built in robust software solutions.Speaker: Zlatko Drmac (Faculty of Science, University of Zagreb) -
180
Wavelet-Based Observables for Koopman Analysis: An Extended Dynamic Mode Decomposition Framework
We present an in-depth analysis of the Koopman operator using wavelet transform. Based on this analysis, we construct an invariant subspace for the Koopman operator and introduce wavelet-based observables that span this subspace. Moreover, on this subspace, we study the Koopman operator and its eigendecomposition. To approximate its action numerically over this subspace, we combine Extended Dynamic Mode Decomposition (EDMD) with the proposed wavelet-based observables, leading to the Wavelet Dynamic Mode Decomposition via Continuous Wavelet Transform (cWDMD) algorithm. We validate our theoretical results on two numerical examples.
Speaker: Cankat Tilki (Virginia Tech) -
181
Reduced order modeling and numerical linear algebra analogs
Constructing low-order approximations to a high-dimensional manifold is a well-studied field as these types of problems arise naturally from the solution of parametric partial differential equations in multi-query or optimization contexts. Full-order approximations, although the most accurate approach to reconstructing a solution manifold, incur too high of an expense in these scenarios. Results for reduced order modeling (ROM) procedures such as proper orthogonal decomposition (POD) and greedy reduced basis methods are often stated in a continuous, functional analysis setting; however, these algorithms are the continuous analog of well-known discrete linear algebra routines for matrix factorizations. In this talk, we compare widely used ROM techniques with their discrete counterparts: POD with SVD, reduced basis methods with column pivoted QR, and empirical interpolation with full pivoted LU. Results from the continuous and discrete settings are juxtaposed to highlights similarities and allow for the interpretation and development of continuous ROM results in light of their linear algebra decomposition analogs.
Speaker: Amy de Castro (University of Utah)
-
179
-
Where Algebraic Coding Theory and Graph Theory Meet: E Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
182
Spectral analysis of linear codes
Let $\mathcal{C} = \{c_0, c_1, \ldots, c_{q^k}\} \subseteq \mathbb{F}_q^n$ be a $[n,k]_q$-linear code endowed with the Hamming metric. That is, $\mathcal{C}$ is a $k$-subspace of $\mathbb{F}_q^n$. Let $M_{\mathcal{C}}\in\mathbb{R}^{q^k\times q^k}$ be the distance matrix of the code defined as $(M_\mathcal{C})_{i,j} := d_H(c_i, c_j)=|\{i\colon x_i\neq y_i\}|$. We analyze the spectrum of $M_{\mathcal{C}}$, providing examples of spectra for well-known families of linear codes. We compute the multiset of the eigenvalues of $M_{\mathcal{C}}$ and bases for their respective eigenspaces.
Speaker: Valentino Smaldore (Università degli Studi di Padova) -
183
Modern Expander-Based Error-Correcting Codes
This talk explores the interplay between coding theory and expander graphs. We will discuss key developments in the design of expander-based codes, including recent advancements that lead to locally testable codes and efficient quantum codes. This includes a review of work dedicated to designing vertex expanders and unique-neighbor expanders, focusing on the specific properties that make them suitable for modern error correction.
Speaker: Pedro Paredes (Princeton University)
-
182
-
Convex Structures in Quantum Information and Gravity: E Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
184
Today’s Experiments Suffice to Indirectly Verify the Quantum Essence of Gravity
The gravity-mediated entanglement experiments employ concepts from quantum information to argue that if entanglement due to gravitational interaction is observed, then gravity cannot be described by a classical system. However, the proposed experiments remain beyond out current technological capability, with optimistic projections placing the experiment outside of short-term future. Here we argue that current matter-wave interferometers are sufficient to indirectly prove that gravitational interaction creates entanglement between two systems. Specifically, we prove that if we experimentally verify the Schrödinger equation for a single delocalized system interacting gravitationally with an external mass, then, under one of two reasonable assumptions, the time evolution of two delocalized systems will lead to gravity-mediated entanglement.
Speaker: Martin Plávala (Leibniz University Hannover) -
185
Time-evolution in generalized probabilistic theories
Generalized probabilistic theories (GPTs) are a general framework to describe physical theories like quantum mechanics and classical mechanics. At their core, GPTs model prepare-and-measure scenarios by describing preparations and measurements within ordered vector spaces and then predicting measurement outcome probabilities. This framework has been very successful in describing and understanding correlations predicted by quantum theory. More recently, the attention has also been extended to the description of Markovian time evolution within GPTs by using positivity-preserving dynamical maps. This extension is important from a fundamental physics perspective because, for example, hypothetical non-quantum dynamics induced by gravity would be detectable via precise tests of the time evolution. Furthermore, dynamical GPTs can also be used in order to better understand quantum dynamics and may enable, for example, quantum computational advantage. In this talk, I will discuss the structure of dynamical GPTs, how they can be used to perform precision tests of quantum physics and help to understand dynamical quantum effects that may serve as precursors to applications in quantum technologies.
Speaker: Matthias Kleinmann (University of Münster) -
186
Quantum Simulation of Non-Positive-Operator-Valued Measurements in General Probabilistic Theories with Post-Selection and Prior Information.
In quantum theory, POVMs form the maximal class of measurements compatible with the Born rule. Operational reconstructions motivate a broader convex framework—General Probabilistic Theories (GPTs)—specified by a convex state space and its dual cone of affine measurement functionals.
Within GPTs one can define “non-positive POVMs” (N-POVMs): Hermitian effects summing to the unit but not positive semidefinite. Although invalid on the full quantum state space, they are mathematically well-defined on suitable convex domains where all outcome probabilities are nonnegative and can outperform genuine POVMs on selected state families (e.g., in state discrimination).
Are such N-POVMs physically realizable? We give a constructive quantum simulation: for a given N-POVM we build a quantum POVM with a failure outcome whose success-conditioned statistics reproduce the N-POVM on an implementation domain. The simulation has two distinct costs—post-selection (success probability) and domain restriction, interpreted as prior information about admissible states—yet the domain can still contain the families exhibiting an N-POVM advantage, so the separation from POVMs persists under the constraints of the implementation. This provides a quantitative bridge between convex operational theories and quantum mechanics.
Speaker: Hayato Arai (University of Tokyo)
-
184
-
Rational Approximation and Interpolation: Practical Applications, Challenges and Solutions: E Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
187
Concentrated real-pole uniform-in-time approximation of the matrix exponential
We propose an asympotically optimal choice of shared concentrated real poles of a family of rational approximants of time-dependent exponential functions exp(−𝑡𝑧) for 𝑧 ≥ 0 and 𝑡 in a positive time interval 𝑇. Our result extends a classical result by J.-E. Andersson [J. Approx. Theory, 32(2):85–95, 1981] on the asymptotic best rational approximation of exp(−𝑧) with real poles. Numerical experiments demonstrate the near-optimality of our choice for various time ranges and for both small and large approximation degrees.
Speaker: Shuai Shao (The University of Manchester) -
188
Multivariate Rational Approximation of Scattered Data Using the p-AAA Algorithm
The parametric adaptive Antoulas–Anderson (p-AAA) algorithm is an effective method for multivariate rational approximation [Carracedo Rodriguez et al., 2023], inspired by the AAA framework for univariate rational approximation [Nakatsukasa et al., 2018].
In its original formulation, p-AAA aims to approximate a function $\mathbf{f} : \mathbb{C}^d \rightarrow \mathbb{C}$ via a multivariate rational function $\mathbf{r} \approx \mathbf{f}$ based on the samples
$$ \begin{equation*} \mathbf{D} = \{ \mathbf{f}(z^{(1)},\ldots,z^{(d)}) \; | \; (z^{(1)},\ldots,z^{(d)}) \in \mathbf{S} \}, \end{equation*} $$ with the sampling points given by the Cartesian grid $$ \begin{equation*} \mathbf{S} = \{ z_1^{(1)}, \ldots, z_{n_1}^{(1)} \} \times \cdots \times \{ z_1^{(d)}, \ldots, z_{n_d}^{(d)} \} \subset \mathbb{C}^d. \end{equation*} $$ The main idea of p-AAA is to iteratively construct $\mathbf{r}$ such that it interpolates $\mathbf{f}$ on a subgrid of $\mathbf{S}$ and minimizes the least-squares error over the remaining data. Despite its effectiveness, the original p-AAA algorithm can not always be applied in practice due to its grid data requirement, which may not always be available or feasible to obtain in practice. In this talk, we discuss how the p-AAA algorithm can be formulated with sampling points given by scattered data sets, i.e., arbitrary subsets of $\mathbb{C}^d$ that do not necessarily follow a grid structure. Towards this goal, we introduce a novel approach for enforcing interpolation conditions on scattered data sets via multivariate rational functions in barycentric form. To incorporate this approach into the p-AAA framework, a constrained linear least-squares problem is integrated into the algorithm. We show that this novel formulation of p-AAA is a strict generalization of the original algorithm, and demonstrate its effectiveness on several challenging examples.Speaker: Linus Balicki (Novateur Research Solutions) -
189
p-set valued AAA for parametric model order reduction.
We present an extension of the AAA algorithm, named psvAAA. This MOR method combines the multivariate pAAA and uni-variate set valued AAA method. Many physical systems are described by dynamical systems with physical or geometrical parameters. Often, the system's output is strongly dependent on the Laplace variable or the frequency, and less strong on the parameters. In this talks, we present a combination of (p)AAA and set-valued AAA. The 'set' is an ensemble of frequency dependent functions evaluated in points in the parameter domain. This set is used to choose support points. In the talk, we describe ways to determine the weights and derive reduced models for evaluation in frequency or time domain.
Speaker: Karl Meerbergen (KU Leuven)
-
187
-
Low-rank Matrix and Tensor Decompositions: Theory, Algorithms and Applications: E Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
190
Reducing swamp behavior for canonical polyadic decomposition
A novel optimization framework is proposed for solving the low-rank tensor approximation problem using the canonical polyadic decomposition (CPD). This can be a difficult optimization problem for certain tensors, e.g., due to degeneracy, i.e., a tensor that can be approximated arbitrarily closely by an ill-conditioned tensor of lower rank. This is one of the phenomena that are encountered in regions of slow convergence, informally known as swamps.
Numerical experiments with state-of-the-art optimization algorithms indicate that in a swamp, often only a few rank-1 terms are modified while others stagnate. Often, the non-stagnant terms are problematic and form an ill-conditioned decomposition. To address this, we propose to temporarily freeze the stagnant terms. The lower number of terms has several benefits: it simplifies the problem by reducing the number of variables and reduces the cost per iteration significantly. Furthermore, in many cases, the residual tensor can be compressed, and an algebraic (re)initialization can be carried out, even if this was not possible for the original tensor. A refinement step can further improve the accuracy if desired. We provide theoretical insights into why terms can stagnate. More specifically, we prove that terms that are close to the solution are not modified anymore in further optimization steps under certain assumptions.
Extensive numerical experiments show that our framework greatly facilitates escaping from swamps. The resulting algorithm outperforms current state-of-the-art approaches on difficult-to-decompose tensors, both in accuracy and computation time, and has similar performance on easier problems.
Speaker: Charlotte Vermeylen (KU Leuven) -
191
Chiselling Algorithms for Algebraic Computation of Tensor Block Term Decompositions
Block term decomposition (BTD) unifies the two most common tensor decompositions: canonical polyadic and Tucker. While BTDs have found a broad range of applications from machine learning to blind source separation, all known algorithms for computing BTDs were historically optimisation-based, and required the desired block sizes to be specified as input. Recently, algebraic BTD algorithms have been found. Cai & Li (2021) were the first to find such an algorithm, with Domanov et al. (2024) later publishing a second, very similar algorithm. While these algorithms work well, very little was known about their theoretical underpinnings.
In this talk, we explain the underlying mechanism of these algorithms by identifying them as special cases of Brooksbank, Kassabov, & Wilson's Lie-theoretic chiselling framework. This perspective allows us to introduce a third, previously unexplored algorithm variant; for third-order tensors, these three variants constitute the only possible chiselling methods for BTD. Using this framework, we characterise the BTDs recoverable by chiselling: specifically, those where the sum of blocks constitutes a direct sum. We further show that for an arbitrary tensor, the set of detectable tensor product subspaces naturally forms a lattice.
Finally, we address uniqueness via two definitions. The first is subspace identifiability, which we show holds generically for all chiselled BTDs. The second is a stronger definition of uniqueness discussed by the previous authors. In particular, this factor matrix uniqueness requires not only unique subspaces, but also that each subspace should be detected with a unique set of bases. Regarding this, Cai & Li claimed their algorithm yields generic uniqueness but did not provide a proof, while Domanov et al. provided only sufficient conditions. We resolve these gaps by formally proving Cai & Li's claim and deriving necessary and sufficient conditions for generic uniqueness across all three variants.
Speaker: David Thorsteinsson (KU Leuven) -
192
Fast randomized compression of matrix-vector products in tensor-train format and applications to Krylov subspace methods
We present a new technique for efficiently compressing matrix–vector products in the tensor-train (TT) format, avoiding the explicit formation of intermediate tensors arising from standard MPO–MPS multiplication. The proposed method performs the compression in a single pass, leading to significant computational and memory savings.
From a theoretical point of view, the resulting approximation is mathematically equivalent to the streaming tensor-train approximation introduced by Kressner, Vandereycken, and Voorhaar. However, our approach admits a more efficient implementation, tailored to the structure of TT operators and vectors, and better suited for large-scale and iterative computations.
Beyond accelerating the fundamental MPO–MPS product, the new method allows us to produce a compressed representation of the operator that can be computed once and reused across multiple applications. This feature is particularly advantageous in iterative algorithms where the same operator is applied repeatedly. We demonstrate how this leads to substantial speedups in Krylov subspace methods, with a special focus on the synergy between the proposed technique and sketched GMRES.
Speaker: Alberto Bucci (University of Edinburgh)
-
190
-
Advances in Randomized Algorithms and Kernel Methods for Rank-Structured Matrices: E Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
193
Acceleration of Lanczos approximation for PDE discretizations in unbounded domains
We consider the approximation of $B^T (A+sI)^{-1} B$ for large s.p.d. $A\in\mathbb R^{n\times n}$ with a dense spectrum and $B\in\mathbb R^{n\times p}$, $p\ll n$ using block-Lanczos recursion. We target the computations of MIMO transfer functions for large-scale discretizations of problems with continuous spectral measures, such as linear time-invariant (LTI) PDEs on unbounded domains.
While Krylov methods such as Lanczos and CG are near-optimal for problems with discrete, well-separated spectra, their spectral adaptation deteriorates in the dense-spectrum regime. We address this limitation using a square-root terminator framework (originated in quantum physics in 1970s), modifying the final Lanczos recursion coefficient in an s-dependent manner via a Krein–Nudelman damped string representation. The resulting approximants reproduce the same Stieltjes moments while yielding continuous spectral measures and significantly reduced approximation error through adaptive damping, that maximizes the relative outflow of energy.Large-scale experiments for diffusion and wave PDEs demonstrate the competitiveness of the proposed approach for deterministic as well as randomized computations.
Speaker: Vladimir Druskin (Southern Methodist University) -
194
Practical Spectral Density Estimation with Explicit Deflation
Stochastic Lanczos Quadrature (SLQ) is a popular algorithm for approximating the spectral density of a symmetric matrix $A$ using matrix-vector products. We present a variance reduced implementation of SLQ. This implementation has two key ingredients: a faster problem-specific eigensolver and a carefully implemented selective orthogonalization scheme that we use as a deflation criterion. Our eigensolver is observed to be faster, more robust, and to scale better than LAPACK's 'stemr' (MRRR) in the context of SLQ. Equipped with this faster eigensolver, we explicitly track residual information and perform deflation to speed up convergence. This is achieved using an implementation that closely follows the LanSO algorithm described in Parlett's $\textit{The Symmetric Eigenvalue Problem}$.
Speaker: Ibrohim Nosirov (Cornell University) -
195
Eigenvector Approximation via Random Sampling
We study the problem of approximating the eigenvectors of an n x n symmetric matrix A with bounded entries using random column sampling. We show that for any eigenvalue $\lambda$ of A, one can compute a vector v satisfying $||Av-\lambda v||_2 \leq \epsilon n$ by sampling $\tilde{O}(\frac{1}{\epsilon^4})$ columns of A. For the eigenvector corresponding to the largest-magnitude eigenvalue, this sample complexity improves to $\tilde{O}(\frac{1}{\epsilon^2})$ columns.
We further establish a lower bound on the number of entries of A that must be sampled to approximate its top eigenvector up to error $\epsilon n$, showing that our algorithm is optimal up to polylogarithmic factors. When columns of A can be sampled with probabilities proportional to their squared $l_2$-norms, we obtain an improved error guarantee of $\epsilon ||A||_F$ for the approximate eigenvectors, where $||A||_F$ denotes the Frobenius norm.
Our analysis builds on recent results of Bhattacharjee et al. (2022) and Swartworth and Woodruff (2025), which demonstrate that all eigenvalues of A can be approximated up to additive error $\epsilon n$ (and $\epsilon ||A||_F$ under $l_2$ sampling) by sampling a sufficiently large principal submatrix.
Speaker: Rajarshi Bhattacharjee (University of Massachusetts Amherst)
-
193
-
Polynomials, Krylov Methods and Applications: E McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
196
Polynomials, Twin BiCG and Approximating the Inverse
The BiCG method for solving linear equations has a polynomial at its core. The new Twin BiCG method solves multiple right-hand systems using the same polynomial for each system. This polynomial is applied implicitly by using the parameters from solving the first right-hand side for all of the systems. Twin BiCG has automatic stability control from the extra copies of eigenvalues that are produced by the nonsymmetric Lanczos algorithm.
We also discuss how Twin BiCG can give an approximation to the inverse of a matrix. Applications include Multilevel Monte Carlo for determining the trace of the inverse of a large matrix. Also, shift-and-invert Arnoldi for computing eigenvalues can solve the associated linear equations to less accuracy as the iteration proceeds. Twin BiCG can efficiently implement this.
Speaker: Ron Morgan (Baylor University) -
197
Preconditioned log-determinant approximation: one probe vector is almost always enough!
We present randomized algorithms for estimating the log-determinant of regularized symmetric positive semi-definite matrices. The algorithms access the matrix only through matrix vector products, and are based on the introduction of a preconditioner and stochastic trace estimator.
We claim that preconditioning as much as we can and making a rough estimate of the residual part with a small budget achieves a small error in most of the cases. We choose a Nystr\"om preconditioner and estimate the residual using only one sample of stochastic Lanczos quadrature. We analyze the performance of this strategy from a theoretical and practical viewpoint. We also present an algorithm that, at almost no additional cost, detects whether the proposed strategy is not the most effective, in which case it uses more samples for the stochastic Lanczos quadrature part. Numerical examples on several test matrices show that our proposed methods are competitive with existing algorithms.Speaker: Daniele Toni (Scuola Normale Superiore) -
198
Stochastic trace estimation for parameter-dependent matrices
Stochastic trace estimators are a family of widely used techniques for approximating traces of large matrices accessible only via matrix-vector products. These methods have been studied extensively when applied to constant matrices $B$. We analyze three standard stochastic trace estimators—the Girard-Hutchinson, Nyström, and Nyström++ estimators—when they are applied to parameter-dependent matrices $B(t)$ that continuously depend on a real parameter $t \in [a, b]$. Our key observation is that a single set of random vectors can be reused to form the estimators for all values of $t$, yielding estimates whose $L^1$-error bounds match those of the constant-matrix case. Traces of parameter-dependent matrices arise naturally in important applications, including spectral density estimation and partition function estimation. Building on our analysis, we develop algorithms that combine Chebyshev interpolation with parameter-dependent stochastic trace estimation to obtain efficient methods with provable accuracy guarantees for these problems.
Speaker: Fabio Matti (EPFL)
-
196
-
3:15 PM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Contributed Talks: F Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
199
The Smith form of Sylvester and Bézout matrices for zero-dimensional ideals
In the talk, I will go through our recent work regarding the Smith form of the Sylvester and Bézout resultant matrices. The partial multiplicities associated to the eigenvalue of a polynomial matrix tell us about the conditioning of computing the eigenvalue. Since the eigenvalues of the resultant matrices are the roots of the system, the partial multiplicities are connected to the stability of solving the polynomial system. We have also attempted to keep the assumptions as general as possible, and as a result, our coefficient field can be any field.
For bivariate polynomials $f,g \in \mathbb{K}[x,y]$, we make a connection between the partial multiplicities of the eigenvalues of the resultant matrices and the structure of the dual space of the zero-dimensional ideal $\langle f,g \rangle$. We show this by explicitly constructing the root polynomials for the matrix $S(y)$ using a Gauss basis of the dual space, introduced in [3M96].
In the talk, I aim to go through the general idea of the proof, as well as intuitively introduce the results through various examples. We first consider the case where $f$ and $g$ have coprime leading coefficients, as then $\det S(y) = \mathrm{Res}(f,g)$. Then, via means of a Möbius transformation that takes the "infinte roots" in the $x$-coordinate to finite roots, we prove that the Smith normal form stays intact through the transformation, allowing us to lift the earlier assumption.
[3M96] M. G. Marinari, H. M. Möller, T. Mora. On multiplicities in polynomial systems solving. Trans. Amer. Math. Soc., 348(8):3283--3321, 1996.
Speaker: Etna Lindy (Aalto University) -
200
A Unified Framework for Linearizations of Matrix Polynomials in Classical Bases
Over the last two decades, a number of methods for constructing linearizations of matrix polynomials have been developed, including ansatz spaces, Fiedler pencils, and block minimal basis pencils. These methods have also been extended in various ways to apply to matrix polynomials expressed in non-monomial bases. However, these extensions have often been achieved one basis at a time, without systematically investigating the possibility of useful relationships between linearizations for matrix polynomials expressed in different bases. This talk describes a unified framework for establishing strong connections between generalized ansatz spaces associated with different bases, enabling the immediate and explicit construction of large spaces of strong linearizations for matrix polynomials expressed in many of the classical polynomial bases.
Speaker: Steve Mackey (Western Michigan University) -
201
The Polynomial Set Associated with a Fixed Number of Matrix-Matrix Multiplications
We consider the problem of computing matrix polynomials $p(X)$, where $X$ is a large dense matrix, using as few matrix-matrix multiplications as possible. Our approach to this problem involves studying the set $\Pi^*_{2^m}$, defined as the set of polynomials computable with at most $m$ matrix-matrix multiplications, but with an arbitrary number of matrix additions and scaling operations. This set is characterized by a tabular parameterization. We derive equivalence transformations of the tabular parameterization, which can be used to identify and eliminate redundant parameters without changing the image space. This leads to two new scientific contributions: a) a precise determination of the dimension: $\dim\Pi^*_{2^m} = m^2$; b) a first minimal parameterization of the set. Using this parameterization, we develop new methods for constructing specific elements of $\Pi^*_{2^m}$. In addition, we use tools from computational algebraic geometry to determine the largest degree such that all polynomials of that degree belong to $\Pi^*_{2^m}$. We then use this knowledge to compute truncated Taylor polynomials of maximal degree for a given number of matrix-matrix multiplications. This talk is based on the preprint [Jarlebring and Lorentzon, arXiv:2504.01500, 2025]
Speaker: Gustaf Lorentzon (KTH)
-
199
-
New Directions and Challenges in Linear Algebra: F Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
202
Divide-and-Conquer for Nonsymmetric Eigenvalue Problems Part I: Randomization
The fastest method for diagonalizing a nonsymmetric matrix or matrix pencil is pseudospectral divide-and-conquer. This two-step algorithm diagonalizes a matrix/pencil by (1) randomly perturbing the input(s) and (2) running fast (and highly-parallel) spectral divide-and-conquer. The key to this approach is the random perturbation, which with high probability implies a guarantee of pseudospectral shattering – i.e., that the spectrum and pseudospectrum of the perturbed problem is sufficiently well-behaved for divide-and-conquer to succeed and to run efficiently. In this talk, we present the most general formulation of pseudospectral divide-and-conquer and discuss efforts to specialize the algorithm to structured problems (e.g., definite pencils).
Speaker: Ioana Dumitriu (University of California, San Diego) -
203
Divide-and-Conquer for Nonsymmetric Eigenvalue Problems Part II: Implementation
While pseudospectral divide-and-conquer is optimal for nonsymmetric eigenvalue problems (in terms of both arithmetic and communication complexity) it is not currently implemented in any of our standard numerical linear algebra libraries. This is due to both the difficulty of translating the algorithm's technical pseudocode into something practical and to the challenge of preparing users for a randomized eigensolver, which will necessarily output different eigenvalues each time it runs. This talk explores the obstacles to bringing pseudospectral divide-and-conquer to practice and the reasons for pursuing a high-performance implementation in spite of them.
Speaker: Ryan Schneider (University of California Berkeley) -
204
Operator learning without the adjoint
There is a mystery at the heart of operator learning: how can one recover a non-self-adjoint operator from data without probing the adjoint? Current practical approaches suggest that one can accurately recover an operator while only using data generated by the forward action of the operator without access to the adjoint. However, naively, it seems essential to sample the action of the adjoint. We prove that that without querying the adjoint, one can approximate a family of non-self-adjoint infinite-dimensional compact operators via projection onto a Fourier basis. We then apply the result to recovering Green's functions of elliptic partial differential operators and derive an adjoint-free sample complexity bound.
We also discuss the following question in the discrete, transpose-free setting: for what problems in numerical linear algebra is it necessary to access matrix-vector products with both a matrix and its transpose? This question also arises in the "unmatched backprojector" setting of certain problems in imaging. In particular, we describe the importance of the transpose in low-rank approximation, least-squares problems, and different varieties of norm estimation.
Speaker: Diana Halikias (New York University) -
205
Spooky Scary Skeletons
In low rank approximations to kernel matrices, skeletons consist of subsets of rows and columns from which CUR, ID, and related approximations can be formed. We consider applications in which parameter-dependent families of matrices with numerical low rank structures appear, such as in parameter dependent partial differential equations. We develop new techniques for analyzing and constructing universal skeletons. A universal skeleton is a single skeleton useful for the entire family of matrices; it captures shared low rank structures across the group. We also supply a practical and memory-efficient algorithm for constructing and using universal skeletons via proxy-point methods. To make things sufficiently scary, we develop universal skeletons in the hierarchical setting, and then use them to develop new fast direct solvers for applications involving multifrequency Helmholtz and Lippmann Schwinger equations.
Speaker: Heather Wilber (University of Washington)
-
202
-
Advanced Acceleration and Convergence Techniques for Solving Linear and Nonlinear Systems: F McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
206
Natural Gradient Descent for Hyperparameter Estimation in Bayesian Inverse Problems
We propose a natural gradient descent (NGD) method for efficient hyperparameter estimation in Bayesian inverse problems. In this framework, the objective is formulated as a function of a symmetric positive definite (SPD) matrix that depends on the hyperparameters. Rather than optimizing purely in the parameter space, we define a natural gradient that incorporates information about the model space (i.e. the space of SPD matrices) into the gradient descent by equipping the manifold with a Riemannian metric. We consider two metrics, the Log-Cholesky and Log-Euclidean metric, both designed for SPD matrices. Finally, we present a numerical example comparing the performance of NGD based on different metrics to the standard gradient descent.
Speaker: Elle Buser (Emory University) -
207
Flattening and Middle Rank Tensor Approximation
Canonical polyadic tensor decomposition and approximation are fundamental problems in multilinear algebra with broad applications in signal processing, machine learning, and scientific computing. The difficulty of those problems depends on both the rank and order of the tensors. We introduce a new method for middle-rank tensor approximation and prove a new criterion for reshaping higher-order tensors. Our method leverages generating polynomials and utilizes linear algebra to generate a good starting point for the middle rank tensor approximation problem. When the given tensor is sufficiently close to a tensor whose rank is below a certain bound, we prove that our algorithm gives a quasi-optimal tensor approximation. Numerical experiments demonstrate that our algorithm can produce accurate tensor approximations for order 3 and higher-order tensors.
Speaker: Zequn Zheng (Louisiana State University) -
208
Accelerated Multigrid Cycles
Multigrid (MG) methods are efficient and scalable for solving sparse linear systems arising from the discretization of partial differential equations (PDEs). However, the performance of standard V- and W-cycle MG methods often deteriorates as the physical and geometric complexity of the PDEs increases. To remedy this, the Algebraic Multilevel Iteration (AMLI)-cycle was developed, utilizing Chebyshev polynomials to define the coarse-level solver. Despite its theoretical strength, AMLI is not widely used in practice because it requires accurate estimations of extreme eigenvalues and can be difficult to implement.
In this talk, inspired by recent acceleration techniques, we propose new accelerated MG cycles that do not require estimating extreme eigenvalues. We prove that these resulting cycles achieve convergence rates comparable to the Chebyshev-based AMLI-cycle. Furthermore, our approach is more straightforward to implement, making it practical for real-world applications. Numerical experiments are presented to validate our theoretical results.
Speaker: Yunhui He (University of Houston) -
209
Algebraic Multigrid for H-Curl-Systems
An algebraic multigrid (AMG) algorithm is proposed for curl-curl electro-magnetics PDEs that are discretized with 1st order edge elements. The key idea behind the algorithm centers on generating edge interpolation operators such that certain near null space properties of the discrete curl-curl operator are preserved on coarse levels so that good AMG convergence rates can be obtained. Specifically, the algorithm guarantees that the algebraically constructed discrete gradient operator lies in the null space of the algebraically generated discrete curl-curl operator on all grid levels of the multigrid hierarchy.
The new algorithm is based on a varient of energy minimization AMG. It takes an already-computed nodal interpolation operator to then define an edge interpolant that satisfies a related discrete commuting relationship while minimizing the energy of the edge interpolation basis functions. Unlike previous works, this algorithm is general in that it can essentially transform any standard nodal interpolation operator into an edge interpolation operator that is suitable for curl-curl problems. Thus, it is possible to leverage traditional AMG schemes to first generate nodal interpolation operators which can then be adapted to edge interpolation operators. We show that the new algorithm can produce "ideal geometric" edge interpolants in some limited cases. Numerical results are provided showing the overall efficacy, comparing with traditional Reitzinger and Schoberl schemes.
Speaker: Raymond Tuminaro (Sandia National Laboratories)
-
206
-
Matrix Inequalities, Matrix Equations, and Their Applications: F Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
210
T-Eigenvalues of Third-order Quaternion Tensors
In this talk, we investigate the eigenvalue problem for third-order quaternionic tensors. We first introduce the notion of right T-eigenvalues and develop an efficient algorithm for their computation, whose effectiveness is demonstrated through comparative numerical experiments.
For Hermitian quaternionic tensors, we then derive bounds for the eigenvalues of tensor sums and extend Weyl’s classical theorem from matrices to the tensor setting.
Finally, we generalize the Ger\v{s}gorin disc theorem to obtain eigenvalue localization results for such tensors, providing a practical tool for spectral estimation.
Speaker: Meiling Deng (University of Nevada, Reno) -
211
Revisiting the Upper Bild Convexity of Quaternionic Numerical Range
The quaternionic numerical range of a matrix is generally nonconvex, in contrast to the classical complex case. Nevertheless, a theorem of So and Thompson in 1996 asserts that the associated \emph{upper bild} in the complex upper half-plane is always convex.
The original proof of So and Thompson relies on a detailed case-by case and computationally involved analysis, including a reduction to the $2\times 2$ case and an explicit description of boundary curves. A recent note by Kumar in 2019 proposes a shorter and more conceptual argument.
In this talk, we revisit Kumar's approach and identify several gaps in the argument, showing that certain steps require additional justification. These issues point to the importance of incorporating geometric structure into the analysis of the upper bild. We also discuss ongoing work exploring algebraic and geometric features of the upper bild toward a more conceptual understanding of its convexity.
[1] W. So and R. C. Thompson, Convexity of the upper complex plane part of the numerical range of a quaternionic matrix, Linear Multilinear Algebra 41 (1996) 303--362.
[2] P. S. Kumar, A note on convexity of sections of quaternionic numerical range, Linear Algebra Appl. 572 (2019) 92--116.
Speaker: Jie Tian (University of Nevada, Reno) -
212
Riccati Differential Equations, State Transition Matrices, and State Covariance Matrices
In this talk, we first investigate the maximal interval of existence for the solution of a symmetric matrix Riccati differential equation. Then, we apply this result to study the reachability of the closed-loop state transition matrix for a linear time-varying system over a finite time interval. Under a mild assumption, we characterize the set of closed-loop terminal state transition matrices reachable from the identity matrix using controls of the state feedback form. Lastly, we provide the set of terminal state covariance matrices reachable from a given positive definite initial state covariance matrix when the linear system is not necessarily controllable.
Speaker: Fengjiao Liu (FAMU-FSU College of Engineering) -
213
Pressing sequences in nonbinary fields
Pressing sequences of simple undirected graphs totally colored by a field arise naturally in computational phylogenetics, where (over $GF(2)$) they are in bijection with sortings-by-reversal of signed permutations which model gene sequences in related organisms. They can be viewed in several surprising different ways, including as any consecutive initial sequence of rows whose diagonal elements can serve, in natural order, as the pivots of Gaussian elimination performed on a symmetric matrix over the chosen field, with the caveat that every pivot entry is a nonzero square. This gives rise to their connection with positivity: for example, a symmetric matrix $M$ is positive definite iff the graph of which it is the (colored) adjacency matrix is ``pressable'' in the order given by the rows of $M$. Thus, finding such an ordered basis of the row space with this property is a natural question, and one for which we have a simple polynomial time algorithm when the base field is $GF(2)$. However, the only algorithms known for fields of characteristic greater than $2$ have time complexity $O(n!)$ where $n$ is the number of vertices in the graph (and the dimension of the corresponding matrix). We discuss progress on this question and future directions.
Speaker: Joshua Cooper (University of South Carolina)
-
210
-
Spectral Graph Theory: F Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
214
Cycles in directed graphs
Let $F$ be a digraph. What is the largest possible minimum outdegree on an $n$-vertex digraph which does not contain a copy of $F$? I will discuss algebraic approaches to this question and present some of my related work (joint with Michael Tait).
Speaker: John Byrne (University of Delaware) -
215
The factorization norm and an inverse theorem for MaxCut
In this talk, we will present a proof of the fact that Boolean matrices with bounded $\gamma_2$-norm or bounded normalized trace norm must contain a linear-sized all-ones or all-zeros submatrix, verifying a conjecture of Hambardzumyan, Hatami, and Hatami. We will also discuss further structural results about Boolean matrices of bounded $\gamma_2$-norm and applications in communication complexity, operator theory, spectral graph theory, and extremal combinatorics.
As a key application, we establish a theorem for MaxCut which contrasts a celebrated result of Edwards. In particular, we show that if the MaxCut of a graph with $m$ edges is at most $m/2 + O(\sqrt{m})$, then it must contain a clique of size $\Omega(\sqrt{m})$.
Speaker: Igor Balla (Leipzig University) -
216
A graph energy conjecture through the lenses of semidefinite programming
In the 1980s, using the software Graffiti, Fajtlowicz made a conjecture relating the independence number of a graph with its energy, a spectral parameter introduced by Gutman (1978). By formulating the graph energy as a semidefinite program (SDP), we take a step towards Fajtlowicz's conjecture, relating the graph energy to the fractional clique covering number. As a byproduct of the SDP formulation, we obtain several lower bounds for the graph energy that improve and refine previous results by Hoffman (1970) and Nikiforov (2007).
Speaker: Emanuel Juliano (Federal University of Minas Gerais) -
217
Semidefinite programming bounds on fractional cut-cover and maximum 2-SAT for highly regular graphs
We use semidefinite programming to bound the fractional cut-cover parameter of graphs in association schemes in terms of their smallest eigenvalue. We also extend the equality cases of a primal-dual inequality involving the Goemans-Williamson semidefinite program, which approximates MAXCUT, to graphs in certain coherent configurations. Moreover, we obtain spectral bounds for MAX 2-SAT when the underlying graphs belong to a symmetric association scheme by means of a certain semidefinite program used to approximate quadratic programs, and we further develop this technique in order to explicitly compute the optimum value of its gauge dual in the case of distance-regular graphs.
Speaker: Henrique Soares Assumpção e Silva (Universidade Federal de Minas Gerais)
-
214
-
Inverse Problems and Uncertainty Quantification through the Lens of Numerical Linear Algebra: F Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
218
Scalable iterative data-adaptive RKHS regularization for linear inverse problems
In this talk, I will present iDARR, a scalable iterative Data-Adaptive RKHS Regularization method for solving ill-posed linear inverse problems. This method searches for solutions in subspaces where the true solution can be identified, with the data-adaptive reproducing kernel Hilbert space (RKHS) penalizing the spaces of small singular values. At the core of the method is a new generalized Golub-Kahan bidiagonalization procedure that recursively constructs orthonormal bases for a sequence of RKHS-restricted Krylov subspaces. The method is scalable, with a complexity of O(kmn) for m-by-n matrices, where k denotes the number of iterations. Numerical tests on the Fredholm integral equation and 2D image deblurring demonstrate that it outperforms the widely used L^2 and l^2 norms, consistently producing stable and accurate solutions that converge when the noise level decreases.
Speaker: Haibo Li (Huazhong University of Science and Technology) -
219
Stochastic zeroth-order calculation of operator quantities
The maximization of the (generalized) Rayleigh quotient is a central problem in numerical linear algebra.
Conventional algorithms for its computation typically rely on matrix-adjoint products,
making them sensitive to errors arising from adjoint mismatches.
To address this issue, we introduce a stochastic zeroth-order Riemannian algorithm
that maximizes the generalized Rayleigh quotient without requiring adjoint or matrix inverse computations.
Moreover, the construction can be translated to other quantities
as for example $\|A\|$ and $\|A - V\|$,
where only evaluations of the linear map $x \mapsto A x$,
respectively $y \mapsto V^*y$ are available.
Those are of interest in inverse problems and uncertainty quantification.
We provide theoretical convergence guarantees
showing that the iterates converge to the set of global maximizers of the (generalized) Rayleigh quotient
at a sublinear rate with probability one.
Our theoretical results are supported by numerical experiments,
which demonstrate the excellent performance of the proposed method compared to state-of-the-art algorithms.Speaker: Jonas Bresch (Technische Universität Berlin) -
220
Nonlinear OED with Column Subset Selection
We address optimal sensor placement for Bayesian nonlinear inverse problems by formulating the task as a matrix column subset selection problem. The design matrix is derived from the expected information gain criterion. Although the resulting solutions are not necessarily globally optimal, the approach presents a rapid time to solution. The effectiveness of the method is demonstrated on nonlinear model problems.
Speaker: Amit Subrahmanya (Argonne National Laboratory) -
221
A Scalable Sequential Framework for Dynamic Inverse Problems via Model Parameter Estimation
Large-scale dynamic inverse problems are often ill-posed due to model complexity and the high dimensionality of the unknown parameters. Regularization is commonly employed to mitigate ill-posedness by incorporating prior information and structural constraints. However, classical regularization formulations are frequently infeasible in this setting due to prohibitive memory requirements, necessitating sequential methods that process data and state information online, eliminating the need to form the full space-time problem. In this work, we propose a memory-efficient framework for reconstructing dynamic sequences of undersampled images from computerized tomography data that requires minimal hyperparameter tuning. The approach is based on a prior-informed, dimension-reduced Kalman filter with smoothing. While well suited for dynamic image reconstruction, practical deployment is challenging when the state transition model and covariance parameters must be initialized without prior knowledge and estimated in a single pass. To address these limitations, we integrate regularized motion models with expectation-maximization strategies for the estimation of state transition dynamics and error covariances within the Kalman filtering framework. We demonstrate the effectiveness of the proposed method through numerical experiments on limited-angle and single-shot computerized tomography problems, highlighting improvements in reconstruction accuracy, memory efficiency, and computational cost.
Speaker: Aryeh Keating (Virginia Tech)
-
218
-
Numerical Linear Algebra Tools for Model Order Reduction: F Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
222
On linear matrix equations and the Akhiezer iteration
The Akhiezer iteration is a new iterative method for solving indefinite linear systems and computing matrix functions. The iteration uses orthogonal polynomial recurrence coefficients to efficiently compute the action of a matrix polynomial to a vector without computing inner products. It features an a priori computable convergence rate and is often faster in practice than standard Krylov subspace methods.
We present extensions of the Akhiezer iteration to large-scale linear matrix equations, such as the Sylvester equation $AX+XB=C$, arising from model order reduction problems.
Speaker: Cade Ballew (University of Washington) -
223
A tangential low-rank ADI method for solving indefinite Lyapunov equations
Continuous-time algebraic Lyapunov equations are linear matrix equations of the form
$$ \begin{equation*} A X E^{\mathsf{H}} + E X A^{\mathsf{H}} = -W \end{equation*} $$ where $A, E \in \mathbb{C}^{n \times n}$ are large-scale sparse coefficient matrices and $W = B R B^{\mathsf{H}}$ represents an indefinite right-hand side defined by the low-rank factor $B \in \mathbb{C}^{n \times m}$ and the indefinite Hermitian matrix $R = R^{\mathsf{H}} \in \mathbb{C}^{m \times m}$. Such matrix equations have become an essential tool in various applications like model reduction, control or data-driven modeling. In the case of large-scale sparse coefficient matrices and indefinite constant terms, indefinite low-rank factorizations have successfully been used to allow methods like the alternating direction implicit (ADI) iteration to efficiently compute accurate approximations to the solution of the Lyapunov equation. However, classical block-type approaches quickly increase in computational costs when the rank of the constant term grows. In this work, we propose a novel tangential reformulation of the ADI iteration that allows for the efficient construction of low-rank approximations to the solution of Lyapunov equations with indefinite right-hand sides even in the case of constant terms with higher ranks. We provide adaptive methods for the selection of the corresponding ADI parameters, namely shifts and tangential directions, which allow for the automatic application of the method to any relevant problem setting. The effectiveness of the developed algorithms is illustrated by several numerical examples.Speaker: Rudi Smith (Virginia Tech) -
224
Data-Driven Modal Truncation
Modal truncation has long been a fundamental approach of model order reduction: to systematically eliminate eigenmodes of a dynamical system that contribute little to the modeling behavior over a given time/frequency range. Typically, this procedure requires the ability to access and utilize intrusive state-space information about the underlying full-order system, which can be infeasible for larger state dimensions or impossible if the full-order system is not known in closed form. Data-driven approaches like proper orthogonal decomposition address these concerns by identifying empirical modes of the system derived from multiple time-domain trajectories of the full-order model, whereafter model reduction can proceed as in the state-space formulation. In this work, we show how one can truncate empirical eigenmodes with only frequency-domain samples of a system's transfer function by utilizing contour integral methods.
After reviewing the necessary theoretical foundations, we will discuss how contour integration and the pole-residue decomposition of an underlying system can be combined to identify all eigenvalues within an arbitrary domain in the complex plane. This tool allows us to perform modal truncation, or identify a stable-unstable decomposition of the underlying system. Through the use of a key result of Keldysh, contour integral methods can also identify eigenvalues of analytic nonlinear matrix-valued functions. We will show numerical experiments illustrating how our data-driven modal truncation method performs on application problems and compare with other state-of-the-art methods.
Speaker: Dan E. Folescu (Virginia Tech) -
225
Model Reduction For Optimal Control By Balanced Truncation Of State and Gradient Covariance
Data-driven reduced-order models (ROMs) could enable near-optimal control for very high-dimensional nonlinear dynamical systems, with applications in active flow control such as relaminarizing turbulent flows and recovering from aerodynamic stall. With initial conditions far away from the desired steady state solving the resulting Hamilton-Jacobi-Bellman (HJB) equation, which defines the value function necessary for feedback control over the continuous state space, computationally intractable due to the curse of dimensionality. Reduced-order models (ROMs) can help in a variety of ways, but existing methods often fail to capture relevant dynamics for the control problem.
To overcome these challenges, we employ an indirect method of trajectory optimization (which is feasible in high dimensions) to obtain state and costate data offline, along locally optimal trajectories for estimating state and gradient covariance matrices. This method builds on the Pontryagin minimum principle and other related work that establishes the costate (adjoint variables) provided as generalized gradients of the optimal value function satisfying the HJB equation. An oblique projection obtained by balanced truncation of these matrices with initial conditions sampled uniformly from an uncontrolled attractor, is used to identify active directions in the state space along which the value function is most sensitive and states have large variance.
The oblique projection obtained is used to build surrogate models for both the value function and the optimal feedback control law, which is validated on the full order model. ROM-based state estimators are also built with these projections for closed-loop feedback control. We assess the quality of the resulting ROMs across linear and nonlinear flow control problems, benchmarking against existing ROM approaches.
Speaker: Emmanuel Ameh (Cornell University)
-
222
-
Theoretical Advances in Operator Learning: F McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
226
Learning Material Constitutive Laws with Neural Operators
The simulation of multiscale viscoelastic materials poses a significant challenge in computational materials science, requiring expensive numerical solvers that can resolve dynamics of material deformations at the microscopic scale. The theory of homogenization offers an alternative approach to modeling, by locally averaging the strains and stresses of multiscale materials. This procedure eliminates the smaller scale dynamics but introduces a history dependence between strain and stress that proves very challenging to characterize analytically. In the one-dimensional setting, we give the first full characterization of the memory-dependent constitutive laws that arise in multiscale viscoelastic materials. Using this theory, we develop a neural differential equation architecture, that simultaneously across a wide range of material microstructures, accurately predicts their homogenized constitutive laws, thus enabling us to simulate their deformations under forcing. We use the approximation theory of neural operators to provide guarantees on the generalization of our approach to unseen material samples.
Speaker: George Stepaniants (California Institute of Technology) -
227
Data-efficient Adjoint-free Learning for Asymptotically Smooth Integral Operators
The goal underlying our work is to develop provably accurate and data-efficient learning algorithms for non-self-adjoint operators using only input-output pairs. State-of-the-art approximation techniques with fast convergence rates either apply only to self-adjoint operators or require access to the adjoint operator, which is unavailable in experimental settings and often difficult to access computationally. Recent work has shown that provably accurate adjoint-free learning is possible but uses an amount of data that is likely excessive. In this talk, we discuss an adjoint-free learning algorithm for asymptotically smooth integral operators which is based on an extension of the fixed sparsity matrix approximation algorithm [Amsel et al., arXiv, 2024] to the case where a fixed subspace - determined by asymptotic smoothness and the resulting exponentially-convergent separable expansion - is known for each row. The algorithm queries the action of the underlying operator on functions with standard Gaussian random coefficients in a finely discretized finite element space, yielding an approximation of the unknown finite element discretization of the operator. While the number of degrees of freedom in this discretization grows polynomially with the inverse error tolerance, asymptotic smoothness and the hierarchical decomposition of each row ensure that only a polylogarithmic number of queries $\mathcal{O}(\text{polylog}(\varepsilon^{-1}))$ is required to achieve the desired accuracy $\varepsilon$ with high probability.
Speaker: Esther Gallmeier (Cornell University) -
228
Quasi-optimal hierarchically semi-separable matrix approximation
We present a randomized algorithm for producing a quasi-optimal hierarchically semi-separable (HSS) approximation to an $N\times N$ matrix $A$ using only matrix-vector products with $A$ and $A^T$. We prove that, using $O(k \log(N/k))$ matrix-vector products and $O(N k^2 \log(N/k))$ additional runtime, the algorithm returns an HSS matrix $B$ with rank-$k$ blocks whose expected Frobenius norm error $\mathbb{E}[\|A - B\|_F^2]$ is at most $O(\log(N/k))$ times worse than the best possible approximation error by an HSS rank-$k$ matrix. In fact, the algorithm we analyze in a simple modification of an empirically effective method proposed by [Levitt & Martinsson, SISC 2024]. As a stepping stone towards our main result, we prove two results that are of independent interest: a similar guarantee for a variant of the algorithm which accesses $A$'s entries directly, and explicit error bounds for near-optimal subspace approximation using projection-cost-preserving sketches. To the best of our knowledge, our analysis constitutes the first polynomial-time quasi-optimality result for HSS matrix approximation, both in the explicit access model and the matrix-vector product query model.
Speaker: David Persson (New York University & Flatiron Institute) -
229
L2-Optimal Reduced-Order Modeling Using Parameter-Separable Forms
We introduce a unified approach to $\mathcal{L}_2$-optimal reduced-order modeling that applies to both linear time-invariant dynamical systems and stationary parametric problems. The framework leverages parameter-separable representations to obtain gradient information for the $\mathcal{L}_2$ objective with respect to the reduced operators, enabling a fully nonintrusive, data-driven, gradient-based construction of optimal reduced models from output data alone. By selecting an appropriate measure, the formulation naturally includes both continuous and discrete cost functions. The proposed methodology is validated through representative numerical examples, and conditions guaranteeing a projection-based realization of the data-driven approximant are established.
Speaker: Serkan Gugercin (Virginia Tech)
-
226
-
Convex Structures in Quantum Information and Gravity: F Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
230
Comparing measurement incompatibility via convex subsets of states
In quantum theory, certain observables cannot be measured simultaneously, a feature known as measurement incompatibility. This concept captures a fundamental limitation of quantum measurements and has deep connections to quantum phenomena. In this talk, we propose an operational framework to characterize measurement incompatibility using restricted sets of states, modeled as convex subsets. We introduce an operational quantification and ordering defined by the detectability of incompatibility over these subsets. We show that these notions establish a fine-grained hierarchy for measurement incompatibility, providing new structural insights into mutually unbiased qubit observables beyond standard noise robustness measures.
Speaker: Ryo Takakura (University of Osaka) -
231
Definitely Not That One: The Art of Antidistinguishing Quantum States
Identifying quantum states is one of the oldest problems in quantum information theory. In this work, we explore a variation of this task: rather than determining which state a system is in, we seek to identify a state that it is not. A set of quantum states is said to be antidistinguishable if this inverse guessing game can be won with certainty. We establish tight bounds characterizing when a set of pure states must be antidistinguishable, as well as when it is impossible.
This is joint work with Vincent Russo, Nathaniel Johnston, and Benjamin Lovitz.
Speaker: Jamie Sikora (Virginia Tech) -
232
Information-theoretic finite de Finetti theorems for quantum and beyond
We develop quantitative de Finetti representation theorems beyond standard quantum settings, driven by the principle that permutation symmetry enforces approximate independence at finite extension level. First, using a GPT-motivated notion of relative entropy (via an integral representation) we define mutual information for general convex state spaces and prove a uniform monogamy bound for multipartite extensions: for permutation-invariant states, the total mutual information is bounded by a constant depending only on the A-system. This yields an information-theoretic finite de Finetti theorem for convex bodies, asserting that the $AB$-marginal of an $n$-extendible (max-tensor) state is close to a separable (min-tensor) state.
Second, for constrained separability problems arising in quantum information, we establish constrained de Finetti theorems compatible with additional linear marginal/fixed-point constraints, including a Bose-symmetric variant that operates directly on symmetric-subspace-supported extensions. These results provide quantitative trace-norm closeness of constrained symmetric (or Bose-symmetric) extensions to convex mixtures of constrained product states, enabling de Finetti control in settings where extremal decompositions into pure products are unavailable. The work is based on arXiv:2507.12326, 2507.12302 and 2601.15184.Speaker: Gereon Koßmann (RWTH Aachen University) -
233
Convergent Inner-outer Approximation Schemes From De Finetti Theorems For Games And Quantum Error Correction
We apply information-theoretic de Finetti principles to build convergent approximation schemes with explicit finite-level guarantees, yielding both outer relaxations and certified inner points. For polynomial optimization over convex bodies with local equality and inequality constraints, an information-theoretic monogamy argument yields a convergent conic hierarchy whose approximation error decays as $\mathcal{O}[1/\sqrt{n}]$ with the extension level, resolving the lack of finite-level guarantees for inequality-constrained de Finetti-based methods. A constructive rounding scheme converts outer feasible points into certified interior approximations, producing matching inner/outer bounds that “sandwich” the true optimum. As an application, we express the optimal GPT value of two-player non-local games as a polynomial optimization problem with local constraints, enabling finite-convergence approximation guarantees in the GPT setting. In the quantum setting of fixed-size free games with bounded entanglement assistance, we obtain SDP outer hierarchies with $\varepsilon$-additive guarantees and combine them with measurement-based rounding to generate inner sequences of feasible strategies. Crucially, we achieve overall $\operatorname{poly}[1/ \varepsilon]$ complexity by exploiting representation-theoretic symmetry reduction (symmetric subspace/Bose symmetry, Schur–Weyl duality, block decompositions) to construct the SDPs directly in symmetry-adapted bases. Finally, for approximate quantum error correction under symmetric noise, we round extendability-based outer bounds into explicit encoder–decoder pairs and develop a framework that combines noise symmetries with extendability symmetry (via commuting group actions) to make higher hierarchy levels computationally feasible. The work is based on arXiv:2507.12326, 2507.12302 and 2601.15184.
Speaker: Julius Alexander Zeiss (RWTH Aachen)
-
230
-
Advances and Challenges in Eigensolvers: F Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
234
The Restarted Block Two-Level Orthogonal Arnoldi Algorithm
The two-level orthogonal Arnoldi algorithm, abbreviated as TOAR, proposed by Lu, Su and Bai, is a Krylov method for the solution of large sparse quadratic eigenvalue problems (QEPs). Traditionally, such eigenproblems are first linearised with an appropriate companion form, then fed into the standard Arnoldi algorithm. This approach has the advantage of being simple, but suffers from large memory requirements when storing the basis. In contrast, TOAR works compactly, requiring (to highest order) half the memory of Arnoldi. Similar algorithms existed prior to TOAR, for example the SOAR (second-order Arnoldi) algorithm, which had the same major benefit as TOAR but was prone to numerical instability.
Interest has been shown in block Arnoldi methods for some time because of their ability to exploit BLAS 3 subroutines. Standard block Arnoldi is widely used and some block SOAR algorithms have shown up in the literature, but not block TOAR. In my work I have extended TOAR to block form and have shown that it has similar numerical stability to non-block TOAR.
Stewart's Krylov-Schur algorithm for restarting Arnoldi is an improvement on the implicitly restarted Arnoldi method (IRAM) in terms of numerical stability. Campos and Roman provide an extension of Krylov-Schur to the TOAR algorithm, which I have further extended to block TOAR.
In this talk, I will briefly explain the TOAR algorithm, then cover my work extending TOAR to block form and talk about a numerical instability issue with the restart method.
Speaker: Michael Jones -
235
Integration of the GPLHR Method Within LS-DYNA
LS-DYNA is a multiphysics simulation software package. It targets a wide range of industrial applications, such as modal analysis problems. These are solved with a variety of homegrown eigensolvers, tailored over decades to industrial models. In this talk, we will focus our attention on the quadratic eigenvalue problem underlying the rotational dynamics' framework of the Jeffcott Rotor model. LS-DYNA currently relies on the implementation of IRAM within ARPACK to target this problem. As an alternative, we have developed an in-house extension of the Locally Optimal Block Preconditioned Conjugate Gradient (LOBPCG) method to the complex case, known as the Generalized Preconditioned Locally Harmonic Residual (GPLHR) method. We present our studies on the numerical behavior and performance of GPLHR, including strong-scaling results (with comparisons to LOBPCG where applicable), sensitivity to tolerances and subspace dimension size, and convergence bottlenecks that trade off preconditioner cost against orthogonalization. Finally, we verify the correctness of the implemented damping contributions by isolating various damping terms in the equation of motion, demonstrating consistent eigenvalue behavior across these cases.
Speaker: Daniel Bielich (Synopsys) -
236
Inexact subspace projection methods for tensor eigenvalue problems
I will present two projection methods for solving high-dimensional tensor eigenvalue problems with low-rank structure: an inexact Lanczos method and an inexact polynomial-filtered subspace iteration. The inexactness arises from low-rank compression, which enables efficient representation of high-dimensional vectors in low-rank tensor formats. A central challenge is that standard operations, such as matrix–vector products, increase tensor rank, necessitating rank truncation and thereby introducing approximation errors. Our numerical results show that subspace iteration is significantly more robust to truncation errors than the Lanczos method. Comparisons with the density matrix renormalization group (DMRG) method further demonstrate that subspace iteration can converge in regimes where DMRG stagnates. Overall, these results highlight inexact subspace iteration as a robust and effective approach for computing multiple eigenpairs of rank-structured tensor operators. A convergence analysis of inexact subspace iteration will also be presented.
Speaker: Alec Dektor (Lawrence Berkeley National Lab) -
237
Adaptive filtered subspace iteration for self adjoint eigenvalue problems on moving domains
Filtered subspace iterations can be used to approximate a finite cluster of eigenvalues of a lower semi-bounded selfadjoint operator in a Hilbert space. Prototype examples of such operators are Schrödinger operators with short-range potentials. A rational function (filter) of the operator is constructed such that the eigenspace of interest (eigenvalues below the infimum of the essential spectrum) is its dominant eigenspace, and a subspace iteration procedure is used to approximate this eigenspace. To approximate an operator in an unbounded domain we use a sequence of finitely truncated domains whose union is the whole space. We present an adaptive multispace algorithm based on a posteriori error estimation. Numerical experiments with spectral and finite element approximation methods confirm the theoretical results. We also discuss an application of the results for other moving domain eigenvalue problems.
Speaker: Luka Grubisic (University of Zagreb, Faculty of Science, Department of Mathematics)
-
234
-
Linear Algebra Education: F Torgersen Hall 2150
Torgersen Hall 2150
Virginia Tech
-
238
Linear Algebra Education from an APOS perspective
In the first part of the talk, I will briefly explain the APOS (Action—Process—Object—Schema) theory and how it is applied to the teaching of mathematical concepts as well as to research on learning. In this approach, the mental construction of a mathematical concept is modeled by means of structures that give the theory its name, and mental mechanisms that allow the passage to new structures. The framework helps interpret sources of student difficulties and intuitive ideas in terms of the elements of the theory. It also offers pedagogical suggestions to overcome the challenges.
In our research group, special importance is given to the design of mathematical tasks that promote progress in the learning spiral described by APOS theory. These activities also have the purpose of identifying and differentiating between different mental structures. I will give examples of such questions involving the concept of linear transformation together with empirical data from students’ productions while engaging with the problems. Characteristics of Action, Process and Object conceptions will be discussed both from the point of view of design principles as well as actual student work.
A phenomenon that often happens in linear algebra classes, which can also be observed in textbooks, is that students are expected to deal with Processes and Objects too soon in the instructional development, while they are still constructing Actions. With suitable pedagogical strategies this situation can be dealt with so that students have the opportunity to develop the necessary conceptions for constructing linear algebra concepts.
Speaker: Dr Asuman Oktaç (Cinvestav) -
239
Engaging Undergraduates in Computational Linear Algebra Through Data-Driven Projects
This talk will discuss the integration of undergraduate research and project-based learning in both abstract and computational linear algebra. Specifically, it will highlight a research training course focused on the mathematical foundations of machine learning and the analysis of human brain connectivity.
Throughout a summer research program and a Python-based numerical linear algebra course, students are trained to apply computational linear algebra, matrices, and graph theory concepts to real-world datasets. The central project involves investigating structural brain networks derived from $90\times 90$ connectivity matrices obtained from MRI data. Furthermore, students analyze the spectra of adjacency matrices and use computational linear algebra tools to identify structural patterns and hubs within human brain networks.
We will discuss that by engaging students in data-driven projects, students develop proficiency in numerical methods in linear algebra while learning the concept of eigenvalues, eigenvectors, and graph representations. Also, this presentation will reflect on student learning outcomes and the pedagogical benefits of using data to fill the gap between classroom theory and interdisciplinary research.
Speaker: Sima Ahsani (Washington and Lee University) -
240
You sunk my Battleship! Exploring Matrix Multiplication with a Linear Algebra Video Game
In this talk, we will discuss a game within the video game Vector Unknown:Echelon Seas (VUES) that focuses on matrix multiplication. VUES was designed by a group of students at Arizona State University in collaboration with mathematics educators and computer scientists. The 3D video game allows students to select one of four matrices (represented by cannonballs) and one of four vectors (represented by a cannon) and fire the cannonball at a ship (represented with a target vector) to practice and interpret matrix multiplication that could be leveraged in a first-year linear algebra course. In addition to a symbolic matrix equation, players are also presented with a geometric representation of the matrix as a lattice generated by the column vectors of the chosen matrix.
We conducted several paired teaching experiments with three students from a northeastern university who had already taken linear algebra. During the paired teaching experiments we asked the students to play the game and explain their thinking as they progressed through the level and asked follow-up questions when clarification was needed. We analyzed the results qualitatively and developed a list of student strategies which we highlight during the talk. We divided these strategies into two categories depending on the information from the game the students leveraged for their reasoning: numeric and geometric. One example a student’s numeric strategy was to selective compute an part of the matrix multiplication to eliminate potential combinations of matrices and vectors by focusing on a coordinate in the target vector. Another strategy involved students leveraging the geometric supports provided to check the result of the multiplication.
In this talk, we discuss the student strategies we observed and they contributed to students not only learning about matrix multiplication computationally, but also deepening their understanding of the underlying geometric at play.
Speaker: Matthew Mauntel (Tarleton State University) -
241
Symplectic Linear Algebra in Honors Linear Algebra - A Proposal
Symplectic linear algebra has had profound connections in many areas of math and physics. Despite this, the subject remains outside the scope of standard linear algebra education. In this talk, I propose motivating the subject via the study of complex coordinate space as the canonical symplectic vector space and argue its approachability for an undergraduate audience. I argue this by highlighting notable theorems and sketching out a hypothetical curriculum.
Speaker: Matthew Park (Virginia Tech)
-
238
-
Recent Advances in Tensor Decompositions for Model and Data Reduction: F Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
242
High-Performance Implementation of Star-M SVD for Big Data Compression
In the era of big data, effectively compressing large datasets while performing complex mathematical operations is crucial. Tensor-based decomposition methods have shown superior compression capabilities with minimal loss of accuracy compared to traditional matrix methods. Under the $\star_M$ tensor framework, tensors can be decomposed in a matrix-mimetic way, including using the $\star_M$ SVD. This tensor SVD has optimality guarantees and has shown exceptional performance on specific types of data, but software implementations have been mostly limited to productivity-oriented languages. In this work, we present our development of a shared-memory parallel, high-performance solution designed to efficiently implement the underlying algorithms. This software will enable optimal compression of extensive scientific datasets, paving the way for enhanced data analysis and insights.
Speaker: Md Taufique Hussain (Wake Forest University) -
243
Streaming Tensor BM-Decomposition
Tensor decomposition is widely used for analyzing multi-way data in various applications that often involve continuously generated data. Efficient methods to process and extract meaningful patterns dynamically are hence essential for these applications. In this talk, we consider the problem of computing the streaming tensor BM-decompositions (BMD). An incremental algorithm, OnlineBMD, is presented, which is the first work that addresses third-order tensor BMD in a dynamic setting.
Speaker: Fan Tian -
244
POD-DEIM in the starM-product framework
We propose using the starM tensor product framework for constructing Proper Orthogonal Decomposition (POD) and Discrete Empirical Interpolation Method (DEIM) reduced order models. By exploiting the inherent multidimensional relationship structure of snapshot data, the approach enables efficient computation of the reduced bases. Operating directly on tensor representations reduces storage and computational costs while maintaining accuracy of the full-order dynamics compared to the more standard vector/matrix based methods. The resulting *M POD–DEIM models are well suited for large-scale, high-dimensional systems, and we demonstrate it on a number of different test problems.
Speaker: Vishwas Rao -
245
Streaming data tensors efficiently and accurately
In this talk, I will present a new method for maintaining low-rank CP decompositions of tensorial data streams. Numerical results indicate that the approach has acceptable computational costs at scale, while significantly improving accuracy and adaptivity to changes in the data stream compared to existing methods. Convergence theory will be provided, in addition to demonstrations on real datasets. Joint work with Yifan Zhang (UT Austin).
Speaker: Prof. Joe Kileel (University of Texas, Austin)
-
242
-
Combinatorial Matrix Theory: F Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
246
Frobenius theorem for Cullis determinant
The first results on transformations preserving matrix invariants is due to Frobenius. This result describes the structure of linear maps $T$ preserving the determinant function, i.e., $\det X = \det T(X)$ for all $X$. Later on there were several extension of this result which are due to Diedonnie, Schur, Dynkin and others.
In 1913 Cullis and then in 1966 independently Radi\'c introduced an analog of the determinant function for the rectangular matrices. Namely, given an injection $\sigma \in S^I_K,$ we denote by ${\rm sgn}_{nk}(\sigma)$ the product
$$s \cdot (-1)^{\sum_{l = 1}^{k} (\sigma(l) - l)} ,$$ where $s={\rm sgn}(\pi)$ is the sign of the permutation $$ \pi = \begin{pmatrix} i_1 & \ldots & i_k\\ \sigma(1) & \ldots & \sigma(k) \end{pmatrix}, $$ here $\sigma(K)={i_1, \ldots, i_k}$, and $i_1<i_2<\ldots <i_k$. Let $n \ge k$, $X \in M_{nk} (\mathbb F)$. Then Cullis determinant $\det_{n\, k}(X)$ of $X$ is defined to be the function: $$ {\det}_{n\, k} (X) = \sum_{\sigma \in S_{k}^{I}} {\rm sgn}_{nk} (\sigma) X_{(\sigma(1),1)}X_{(\sigma(2),2)}\ldots X_{(\sigma(k),k)}. $$ Cullis determinant has many interesting properties of the usual determinant as well as many new properties. It is interesyting as from theoretical point of view, as due to the number of applications. In this talk we discuss an analog of Frobenius theorem for Cullis determinant, namely we characterize linear preservers of this function. It appears that the preservers depend on the parity of $k+n$. Our proof is based on analogs of Flander's results and matroid theory.The talk is based on the series of joint works with Andrey Yurkov.
Speaker: Alexander Guterman (Bar-Ilan University) -
247
Geometric and Combinatorial Properties of the ASM Polytope
The polytope $ASM_n$, the convex hull of the $n\times n$ alternating sign matrices, was introduced by Striker and by Behrend and Knight. A face of $ASM_n$ corresponds to an elementary flow grid defined by Striker, and each elementary flow grid determines a doubly directed graph defined by Brualdi and Dahl. We show that a face of $ASM_n$ is symmetric if and only if its doubly directed graph has all vertices of even degree. We show that every face of $ASM_n$ is a 2-level polytope. We show that a $d$-dimensional face of $ASM_n$ has at most $2^d$ vertices and $4(d-1)$ facets, for $d\ge 2$. We show that a $d$-dimensional face of $ASM_n$ satisfies $vf\le d2^{d+1}$, where $v$ and $f$ are the numbers of vertices and facets of the face. If the doubly directed graph of a $d$-dimensional face is 2-connected, then $v\le 2^{d-1}+2$. We describe the facets of a face and a basis for the subspace parallel to a face in terms of the elementary flow grid of the face. We prove that no face of $ASM_n$ has the combinatorial type of the Birkhoff polytope $B_3$. We list the combinatorial types of faces of $ASM_n$ that have dimension 4 or less.
Speaker: Elizabeth Dinkelman (George Mason University) -
248
Sparsity of a matrix and its inverse
This talk discusses problem of determining the minumum number of nonzero entries in a pair of matrices (A,A^(-1)) for A in various families of matrices (e.g. irreducible, fully indecomposable, primitive, positive definite, orthogonal, symmetric with connected graph, irreducible covariance matrices). Some of the results are from work with H. Gupta, L. Hogben and T. Wong.
Speaker: Bryan Shader (University of Wyoming) -
249
A new combinatorial rank concept and its challenges
We introduce a new rank concept for $(0,\pm 1)$-matrices, called the $\pm$-rank of a $(0,\pm 1)$-matrix. This ``generalizes'' the binary rank and the term rank of (0,1)-matrices. We establish several inequalities relating the different ranks, including ordinary real rank. Moreover, the $\pm$-rank is discussed for certain classes of $(0,\pm 1)$-matrices, such as alternating sign matrices (ASMs), network matrices etc. We also focus on some challenging problems concerning this rank concept, both of theoretical and computationally nature.
Speaker: Geir Dahl
-
246
-
250
Mentoring Panel McBryde Hall 100
McBryde Hall 100
Virginia Tech
This panel aims to provide guidance and support to early-career researchers through a discussion of mentoring experiences, challenges, and best practices across different academic systems. This activity is part of the broader service component of an National Science Foundation Early Career Award, through which we have received support to fund participation of U.S.-based early-career conference speakers.
Speakers: Anna Ma (University of California, Irvine), Daniela Calvetti (Case Western Reserve University), Francoise Tisseur (The University of Manchester), Heike Faßbender, Mirjeta Pasha, Misha Kilmer (Tufts University), Roel Van Beeumen (Lawrence Berkeley National Laboratory), Shaun Fallat (University of Regina)
-
125
-
-
251
Announcements McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
Plenary Talks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
252
Quantum Numerical Linear Algebra Doesn't Have to be Hard: A Matrix Algebra Oriented Approach
The field of quantum computing offers a unique opportunity to revolutionize numerical linear algebra and scientific computing. This stems from the ability of quantum computers to efficiently model complex structures, and to represent and manipulate high-dimensional vectors and matrices using exponentially fewer qubits. These advantages arise from the fundamental principles of superposition and entanglement inherent to qubits.
Yet, the current landscape of quantum computing research is dominated by intricate, tailor-made circuit designs developed in an ad-hoc manner for specific mathematical challenges. Although such state-of-the-art quantum algorithms provide a powerful means of translating diverse computations into circuits, their development is far from straightforward. It often demands extensive ``circuit engineering'' to achieve desired mathematical outcomes. This also extends to quantum numerical linear algebra.
In this talk, I will discuss our recent progress on developing a unified and systemic approach to utilizing quantum computing for numerical linear algebra. Our research centers around a novel Quantum Linear Algebra (qLA) framework offering fundamental matrix algebra building blocks, akin to BLAS -- but for Quantum Computers.
Speaker: Haim Avron (Tel Aviv University) -
253
The Eigenvalue Method in coding theory
One of the main goals in spectral graph theory is to deduce the principal properties and structure of a graph from its graph spectrum. In this talk we will show how spectral graph theory provides powerful methods for obtaining results concerning substructures of graphs, and also how these results can be useful in other mathematical fields such as coding theory. In particular, we will derive sharp eigenvalue bounds for the k-independence number of a graph (or equivalently, the independence number of the k-th graph power), which is known to be very hard to compute. We will see how to use polynomials and mixed integer linear programming in order to optimize and compute such spectral bounds. Finally, we will illustrate some recent applications of the obtained eigenvalue bounds to coding theory. The obtained results are encouraging and strongly suggest that spectral graph theory can uncover structural properties of ambient spaces that are relevant to coding theory, but that are often not captured by classical coding theory techniques. There is no question that eigenvalues play a central role in our fundamental understanding of graphs. The goal of this talk is to show that we can also use them for deepening our understanding of codes.
Speaker: Aida Abiad (Eindhoven University of Techonolgy)
-
252
-
10:15 AM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Matrix Geometries: G Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
254
Stiefel Optimization is NP-hard Goodwin Hall 135
Goodwin Hall 135
We show that linearly constrained linear optimization over a Stiefel or Grassmann manifold is NP-hard in general. We show that the same is true for unconstrained quadratic opti- mization over a Stiefel manifold. We will show that unless P = NP, these optimization problems over a Stiefel manifold do not have FPTAS. As an aside we extend our results to flag manifolds. Combined with earlier findings, this shows that manifold optimization is a difficult endeavor — even the simplest problems like LP and unconstrained QP are already NP-hard on the most common manifolds.
Speaker: Dr Tianyun Tang (University of Chicago) -
255
Differential-Geometric View of the Schur--Horn Theorem and Related Convexity Phenomena Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
We present a differential--geometric view of the Schur--Horn theorem and related convexity phenomena. For an $n\times n$ Hermitian matrix $A$ with simple spectrum, the Schur--Horn map
$$ \mu: {\mathrm U}(n) \to \mathbb R^n,\quad \mu(U)=\mathrm{diag}(UA U^{-1}), \qquad U\in {\mathrm U}(n), $$ is shown to be a proper submersion over the relative interior of the Schur--Horn polytope, where ${\mathrm U}(n)$ is the unitary group. We obtain a smooth path-lifting property and a global smooth selection along any line segment in the interior, providing a geometric strengthening of the classical majorization theorem and a proof of Westwick's $c$-numerical range theorem. The talk will emphasize the matrix geometry and topology.Speaker: Tin-Yau Tam (University of Nevada, Reno) -
256
Stochastic approximations with operator means Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
In this talk we investigate zeros of nonlinear operators on the cone of positive definite operators over a Hilbert space. The unique zero of such nonlinear operators define means of positive definite operators. Moreover these nonlinear operators generate strictly exponentially contracting semigroups in some metric defined on positive operators. We survey recent results that establish the operator norm convergence of deterministic and stochastic resolvent and proximal type algorithms, in particular versions coming from a Trotter-Kato type splitting formula. Applications include generalizations of results proved under strong moment conditions for generalized Karcher means. The talk is based on recent joint work with Zoltán Léka and is a continuation of our earlier project.
Speaker: Pálfia Miklós (Corvinus University of Budapest)
-
254
-
The Inverse Eigenvalue Problem of a Graph and Zero Forcing: G Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
257
Leaky Forcing of Unicyclic Graphs
Zero forcing is a graph coloring process in which a set of initially blue vertices force the remaining vertices in the graph to be colored blue after repeated applications of a color change rule. Leaky forcing is a fault-tolerant variant of zero forcing in which some set of $\ell$ vertices, called leaks, are forbidden from forcing. The $\ell$-leaky forcing number is the size of the smallest set of initially blue vertices for which the graph is successfully forced despite the placement of $\ell$ leaks. In this talk, we will characterize the $\ell$-leaky forcing number for unicyclic graphs.
Speaker: Carolyn Reinhart (Swarthmore College) -
258
Orientable forcing and relationships with linear algebra
In this talk, we introduce a new parameter, the orientable forcing number of an undirected graph $G$, which is the maximum zero forcing number among all oriented graphs that have $G$ as their underlying undirected graph. We establish some properties of the orientable forcing number, including extreme values, and discuss how the parameter relates to matrices.
Speaker: Bonnie Jacob (Rochester Institute of Technology) -
259
Parameters connected to the strong nullity interlacing property
Given a simple graph $G$, $\mathcal{S}(G)$ is the set of real symmetric matrices indexed by the vertices in $G$ and with off-diagonal zeros corresponding to non-edges in $G$. The problem of finding the maximum nullity of a matrix in $\mathcal{S}(G)$ has been extensively studied. We consider the maximum nullity of a matrix $A$ and its principal submatrix $A(i)$ corresponding to deleting the vertex $i$. The strong Arnold property possessed by some matrices is one of the useful tools for studying maximum nullity. The parameter $\xi(G)$ is the maximum nullity of a matrix in $\mathcal{S}(G)$ with the strong Arnold property, and has been shown to be minor monotone. We defined a new parameter which we call $\xi\xi(G)$ for the maximum sum of the nullities of $A$ and $A(i)$ for matrices $A \in \mathcal{S}(G)$. We show that $\xi\xi(G)$ is also minor monotone. In this talk I will show how $\xi\xi(G)$ is a refinement of $\xi(G)$ and minimal minors for small values of $\xi\xi(G)$.
Speaker: Mary Flagg (University of St. Thomas)
-
257
-
Numerical Ranges and Numerical Radii: G Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
260
Norms on tensors in quantum information related to numerical radii
In this talk we survey some recent results on entanglement and separability of general, symmetric (bosons), skew-symmetric (fermions) tensors, and their computability. All these results are related to corresponding numerical radii. This talk is based on the arXiv preprint
S. Friedland, Tensors, entanglement, separability, and their complexity, arXiv:2509.21639, 2025.
Speaker: Prof. Shmuel Friedland (University of Illinois Chicago) -
261
Linear maps preserving products of involutions
Linear preserver problems study linear maps on matrix spaces that leave certain functions, subsets, or relations invariant, whereas matrix decomposition problems focus on expressing matrices as products of matrices with prescribed structural properties. An element of the algebra $M_n(\mathbb{F})$ of $n \times n$ matrices over a field $\mathbb{F}$ is called an involution if its square equals the identity matrix. A classical result of Gustafson, Halmos, and Radjavi asserts that any product of involutions in $M_n(\mathbb{F})$ can be written as a product of at most four involutions. It is also known that a matrix is a product of two involutions if and only if it is similar to its inverse. These results naturally lead to the following linear preserver question: which linear maps on $M_n(\mathbb{F})$ preserve products of involutions?
In this talk, we address this question and present a characterization of bijective linear maps that preserve matrices expressible as products of two, three, or four involutions in $M_n(\mathbb{F})$. We also outline possible directions for further research. This is joint work with Chi-Kwong Li and Sushil Singla.
Speaker: Dr Tejbir Lohan (Indian Statistical Institute, Delhi Centre, India) -
262
On the $k$-numerical ranges of matrices
This study considers some problems involving the $k$-numerical range. Following the idea of the zero-dilation index, the notion of the zero-trace index is introduced, which is defined as the largest zero-trace compression of a matrix. Alternative characterization of the zero-trace index is given, and zero-trace indices of certain classes of matrices are identified. The study also considers recent results on the numerical range of cyclic shift matrices. A recent solution to a conjecture proved that a certain arrangement of the weights of a given cyclic shift matrix maximizes the classical numerical range. This work explores whether analogous optimal arrangements exist for the $k$-numerical range.
Speaker: Kennett Dela Rosa (University of the Philippines Diliman)
-
260
-
Inverse Problems and Uncertainty Quantification through the Lens of Numerical Linear Algebra: G Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
263
Randomized Generalized Error Minimizing Method for Linear Ill-Posed Problem
For solving noisy linear ill--posed problems arising from the practical applications, the residual based iterative methods may suffer semi-convergence phenomenon, where the iterates initially get closer to the desired solution but then degrade as the iteration continues. Building upon the randomized Gram--Schmidt algorithm, a random sketching technique known to reduce inner product computational costs over classical Gram--Schmidt and numerical stability comparable to modified Gram--Schmidt, we develop a novel randomized generalized error minimizing (GMERR) Krylov subspace method. This process extends the successful application of randomized Gram--Schmidt in methods such as randomized GMRES and LSQR. We further introduce a block variant, resulting in a block randomized Arnoldi process and a block GMERR method for large-scale ill-posed problems. A theoretical analysis of the regularization properties and numerical stability of the proposed methods is provided, leveraging random projection theory. Numerical experiments demonstrate the efficacy of the new algorithms.
Speaker: Prof. Ning Zheng (Tongji University) -
264
Mixed-to-Low Precision Iterative Methods for Linear Inverse Problems
Many problems in science and engineering give rise to linear systems of equations that are commonly referred to as large-scale linear discrete ill-posed problems. The matrices that define these problems are typically severely ill-conditioned and may be rank deficient. Because of this, regularization is often needed to stem the effect of perturbations caused by error in the available data. In this talk we consider the solution of the regularized least-squares problem using both mixed-to-low precision and Krylov subspace projection techniques. We utilize a filter factor analysis to investigate the regularizing behavior of the proposed iterative methods as well as numerical experiments to verify their efficacy.
Speaker: Lucas Onisk (Emory University) -
265
Separable Nonlinear Bayesian Inverse Problems
Separable nonlinear inverse problems arise in many applications where a forward model depends linearly on some unknowns and nonlinearly on others, including semi-blind deconvolution. We adopt a Bayesian framework with Gaussian noise and Gaussian priors on the linear variables, leading to regularized formulations of the inverse problem. We examine prior models for the nonlinear parameters and show that maximum a posteriori (MAP) estimation yields regularized separable nonlinear least squares problems that can be efficiently solved using variable projection (VarPro) methods, as demonstrated through numerical examples.
Speaker: Malena Espanol (Arizona State University)
-
263
-
Code-based Cryptography: G Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
266
Decoding from the Other Side: Primal vs. Dual Attacks
The security of code-based cryptography relies fundamentally on the computational hardness of decoding random linear codes. Until recently, the most efficient known algorithms for the decoding problem were Information Set Decoding (ISD) algorithms, which we refer to as primal attacks in this presentation.
In 2001, a new class of decoding algorithms, known as dual attacks, was introduced and has been significantly improved over the past four years. These attacks rely on a reduction of the decoding problem to the Sparse Secret Learning Parity with Noise (sparse-LPN) problem, where samples are constructed using low-weight codewords of the dual (orthogonal) code. As a consequence, dual attacks require, as a core subroutine, the ability to find low-weight codewords in the dual code.
This observation can be seen as a first step toward addressing a long-standing open question: can the decoding problem be dualized, that is, polynomially reduced to a decoding problem in the dual code? This question is of particular importance in code-based cryptography, since a positive answer would imply that decoding at rate R can be reduced to decoding at rate 1 - R. Such a result would be highly significant, as the complexity of existing primal attacks is not symmetric with respect to the code rate.
In this presentation, after a brief overview of primal and dual attacks, we examine the parameter regimes in which dual attacks currently outperform the best known primal attacks. In particular, we focus on the relevance of these attacks for cryptosystems such as BIKE, McEliece, and HQC, where the target decoding distance grows sub-linearly with the code length.
Speaker: Kévin Carrier (École Polytechnique (France)) -
267
Post-Quantum Blind Signatures from Matrix Code Equivalence
We construct a novel code-based blind signature scheme, using the Matrix Equivalence Digital Signature (MEDS) group action. The scheme is built using similar ideas to the Schnorr blind signature scheme and CSI-Otter, but uses additional public key and commitment information to overcome the difficulties that the MEDS group action faces: lack of module structure (present in Schnorr), lack of a quadratic twist (present in CSI-Otter), and non-commutativity of the acting group. We address security concerns related to public key validation, and prove the security of our protocol in the random oracle model, using the security framework of Kastner, Loss, and Xu, under a variant of the Inverse Matrix Code Equivalence problem and a mild heuristic assumption. We also discuss alternative techniques for constructing a code-based blind signature and consider possible parameter sets and corresponding performance metrics.
Speaker: Veronika Kuchta (Florida Atlantic University) -
268
Half is Enough: halving keys through optimal representation of self-orthogonal codes
Given two linear codes, the Permutation Equivalence Problem (PEP) asks to find a permutation that maps one code onto the other.
The state-of-the-art solvers for PEP take time that is either exponential in the code length or in the dimension of the hull, which is the intersection between a code and its dual.
To avoid the latter type of attacks, PEP-based cryptosystems employ linear codes with large hull such as self-orthogonal codes, i.e., codes contained in their dual.
Due to the need of representing such codes, whose communication cost grows quadratically with their length, such cryptosystems suffer from large public keys.
We present an efficient compression technique that allows to represent self-orthogonal codes with fewer bits, taking advantage of the pairwise orthogonality of their codewords.
Among the cryptosystems that benefit from this work there are SPECK, LESS (and its threshold variant LEAST), which is a candidate in the ongoing NIST standardization process, and the updatable public key encryption scheme proposed by Albrecht, Benčina and Lai, which is based on a special instance of the Lattice Isomorphism Problem (LIP).
Remarkably, the compression technique nearly halves the public key for the former two schemes.Speaker: Mr Rahmi El Mechri (Univeristà Politecnica delle Marche, Scuola IMT Alti Studi Lucca)
-
266
-
Linear Algebra Foundations for Data-driven Modeling and Model Order Reduction: G Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
269
Dynamical Tensor Train Approximation for Kinetic Equations
The numerical solution of kinetic equations is challenging due to the high dimensionality of the underlying phase space. In this paper, we develop a dynamical low-rank method based on the projector-splitting integrator in tensor-train (TT) format. The key idea is to discretize the three-dimensional velocity variable using tensor trains while treating the spatial variable as a parameter, thereby exploiting the low-rank structure of the distribution function in velocity space. In contrast to the standard step-and-truncate approach, this method updates the tensor cores through a sweeping procedure, allowing the use of relatively small TT-ranks and leading to substantial reductions in memory usage and computational cost. We demonstrate the effectiveness of the proposed approach on several representative kinetic equations.
Speaker: Geshuo Wang (University of Washington) -
270
Tensor-based Dynamic Mode Decomposition for Complex Dynamics
In this talk, I will present recent work on analysing time series data for complex dynamics. Extended dynamic mode decomposition (EDMD, Williams et al, 2015) is a widely used algorithm to learn a linear surrogate model for the statistics of an evolving dynamics, based on the Koopman operator framework. For high-dimensional systems, choosing a suitable basis set can become challenging, as traditional methods are subject to the curse of dimensionality. Tensor-based versions of EDMD were first presented by Klus and co-workers in 2018, and then extended by Nüske et al in 2021. A further extension to approximate the infinitesimal generator of the dynamics was developed by Lücke et al in 2022.
In this talk, I will provide an overview of these developments, and then present some recent results on the application of tensor-based EDMD to simulation data of complex bio-molecules. In particular, I will show how combining dimensionality reduction methods with tensor-based models helps unravel complex structure in high-dimensional time series data.
Speaker: Feliks Nueske (Max-Planck-Institute DCTS Magdeburg) -
271
Learning magnetic field structure from trajectories
Stellarators are non-axisymmetric magnetic field configurations used to confine plasmas. Within a stellarator, particles roughly follow magnetic field lines, and the magnetic fields in stellarators can be organized into different regions according to the dynamics of field line flows, with regions of nested flux surfaces potentially interspersed with islands or regions of chaos. While can visualize these structures with Poincare plots, such visualizations are not wholly satisfactory if we wish to numerically optimize such structures. In this talk, we describe an approach to automatically inferring the structure of magnetic field line flows by building models of the dynamics of field line trajectories. We give connections to dynamical systems theory, numerical extrapolation methods, and the theory of adaptive filtering.
Joint work with Max Ruth (UT Austin).
Speaker: David Bindel (Cornell University)
-
269
-
Hierarchical Low-Rank Approximations: Algorithms and Applications: G McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
272
Extension of hierarchical matrices to hierarchical tensors with butterfly compression
The development of hierarchical matrix techniques has been essential for many modern scientific computing frameworks including fast direct solvers for PDEs and integral equations, scalable kernel methods and Gaussian processes, and second-order optimization and inverse problems, etc. These algorithms oftentimes lead to optimal computational and memory complexities. That said, when dealing with higher-dimensional problems, e.g., PDEs and kernels in at least 3D spaces, matrix algorithms can still be expensive and tensor algorithms become more attractive. Despite several recent pioneering works attempting to extend hierarchical matrices to hierarchical tensors, the field of hierarchical tensors remains largely unexploited. This work attempts to tackle the computational challenge of fast tensor algorithms for highly-oscillatory and high-dimensional integral operators. We develop hierarchical tensors with tensor butterfly compression, and demonstrate its efficiency for computing Green's function tensors in volume integral equations and Babich ansatz for high-frequency wave equations. This is a joint work with Jianliang Qian, Yuxiao Wei, Tianyi Shi, Hengrui Luo and Paul Michael Kielstra.
Speaker: Yang Liu (Lawrence Berkeley National Laboratory) -
273
Estimating High-Dimensional Covariance Matrices with Hierarchical Rank Structure
Many algorithms in data assimilation and model order reduction rely on sample-based estimates for a covariance matrix associated with the trajectory of a high-dimensional dynamical system. The number of available samples is often far less than the dimension of the underlying state space because of computational constraints. Under these circumstances, extracting meaningful covariance information requires that the noisy statistics of the sample are regularized with a structural assumption. This talk will describe a regularization technique that exploits the rank structure of submatrices representing cross-covariances between well-separated domains of space. We will establish that low-rank truncations of these submatrices can be estimated from fewer samples than the submatrices themselves. We will then show how this fact can be used to encode physics-informed regularizing assumptions onto the sample statistics, resulting in a hierarchically rank-structured covariance estimator. Through numerical experiments with a variety of dynamical systems, we will demonstrate that these techniques are effective at reducing sampling errors in the covariance.
Speaker: Robin Armstrong (Cornell University) -
274
Parametric Hierarchical Matrix Approximations to Kernel Matrices
Kernel matrices arising in applications such as Gaussian processes may not always admit a low-rank approximation. Important examples are kernel matrices induced by certain members of the Matérn family of covariance kernels, with smaller length scales and values of $\nu$. Still, they can often be approximated by a hierarchical matrix ($\mathcal{H}$-matrix or $\mathcal{H}^{2}$-matrix), which consists of a hierarchy of small near-field blocks (sub-matrices) stored in a dense format and large low-rank far-field blocks that are efficiently stored in factored form. A hierarchical matrix approximation of a kernel matrix can be constructed, stored, and used to perform matrix-vector multiplication in log-linear or linear complexity with respect to $n$. Standard methods for approximating kernel matrices with hierarchical matrices do not account for the following: kernel matrices often depend on certain hyperparameters that must be optimized over a fixed parameter space. For example, in Gaussian processes and Bayesian inverse problems, estimating the hyperparameters from the data involves solving an optimization problem, which requires repeatedly forming or approximating the kernel matrices for a range of parameters. To address this computational challenge, we introduce a new class of hierarchical matrices, namely, parametric (parameter-dependent) hierarchical matrices. The construction of a parametric hierarchical matrix follows an offline-online paradigm. In the offline stage, the near-field and far-field blocks are approximated by using polynomial approximation and tensor compression. In the fast online stage, for a particular hyperparameter, the parametric hierarchical matrix is instantiated efficiently as a standard hierarchical matrix. Numerical experiments show speedups of over 100 times compared with existing techniques.
Speaker: Abraham Khan (North Carolina State University)
-
272
-
Eigenvalues of Nonnegative and Stochastic Matrices: G Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
275
C-realizability in the Symmetric Nonnegative Inverse Eigenvalue Problem: a combinatorial characterization
The Real Nonnegative Inverse Eigenvalue problem (hereforth, RNIEP) consists, for a given positive integer n, in characterizing the lists of n real numbers which are the spectrum of some n × n matrix with real entries. C-realizability was originally introduced in [1] as a sufficient condition for the RNIEP. It was shown back then that C-realizability included as particular cases most of the known sufficient conditions for the RNIEP. It was not until 2017 that it was shown in [2] that C-realizability was more closely related to the Symmetric Nonnegative Inverse Eigenvalue problem (SNIEP)) than to the RNIEP. The combinatorial nature of the original definition of C-realizability has conduced over the years to combinatorial characterizations of the set of C-realizable lists. In this talk we first review a partial one (see [4]) for real lists with zero sum, and then the most general combinatorial characterization, obtained in [5] for arbitrary real lists. One of the most remarkable consequences of the latter is that it proves the monotonicity of C-realizability, i.e., that the operation of increasing any positive entry of a C-realizable list preserves C-realizability.
References
[1] A. Borobia, J. Moro & R. L. Soto (2008). A unified view on compensation criteria in the real nonnegative inverse eigenvalue problem. Linear Algebra Appl., 428, 2574-2584.
[2] R. Ellard & H. Smigoc (2017). Connecting sufficient conditions for the Symmetric Nonnegative Inverse Eigenvalue Problem. Linear Algebra Appl, 498, 521-552.
[3] C. R. Johnson, T. J. Laffey & R. Loewy (1996). The Real and the Symmetric Nonnegative Inverse Eigenvalue Problem are different. Proc. Am. Math. Soc., 124, 3647-3651.
[4] C. Marijuán & J. Moro. (2021) A characterization of trace-zero sets realizable by compensation in the SNIEP. Linear Algebra Appl., 615, 42-76.
[5] C. Marijuán & J. Moro. (2024) A characterization of sets realizable by compensation in the SNIEP. Linear Algebra Appl., 693, 425–447.
Speaker: Julio Moro (Universidad Carlos III de Madrid) -
276
Characteristic Polynomials of Nonnegative Hessenberg Matrices
The nonnegative inverse eigenvalue problem (NIEP) seeks to characterize the multisets of complex numbers that can be realized as the spectra of nonnegative matrices. Hessenberg matrices have been used in several central constructive results within this field, including the resolution of the NIEP for $4 \times 4$ matrices, the characterization of spectra where all non-Perron eigenvalues possess negative real parts, and the constructive proof of the Boyle-Handelman theorem.
In this talk, we examine the set of characteristic polynomials realized by nonnegative Hessenberg matrices. To analyze the underlying zero-nonzero patterns of this family of matrices, we introduce an adaptation of the non-symmetric strong spectral property. We apply this modified spectral property to address the following problem: given a prescribed pattern of a nonnegative Hessenberg matrix $A$, we seek to identify perturbations of the characteristic polynomial of $A$ that maintain its realizability with nonnegative Hessenberg matrices.
Speaker: Helena Šmigoc (University College Dublin) -
277
NIEP: positive and irreducible realizations
The Nonnegative Inverse Eigenvalue Problem (NIEP) is the problem of characterizing the lists $\sigma$ of $n$ complex numbers (counting multiplicities) that are the spectrum of a nonnegative matrix of size $n$. A list $\sigma$ is said to be realizable if there exists a nonnegative matrix whose spectrum is $\sigma$.
Another way of facing the NIEP is to focus the attention on the characteristic polynomial $P(x)$ of $\sigma$, and to characterize the real monic polynomials of degree $n$ that are the spectrum of a nonnegative matrix of size $n$. Note that a nonnegative matrix $A$ can be seen as the adjacency matrix of a weighted digraph. A real monic polynomial $P(x)$ is said to be realizable if there is a weighted digraph, or equivalently a nonnegative matrix, whose characteristic polynomial is $P(x)$.
Our interest here lies in identifying, among realizable spectra/polynomials, those that are realizable by an irreducible matrix/a strongly connected weighted digraph. We use a special type of digraph structure, called EBL, in which the cycles are especially simple. The EBL digraphs are a tool that allow to relate the information contained in the cyclic structure of the digraph associated to $A$ to the coefficients of its characteristic polynomial $P(x)$, and vice versa, in a suitable way.
After giving some general background, we make some useful new observations and show the existence of irreducible nonnegative and positive realizations in some general cases. We discuss both types of realizations for $n< 5$, where the NIEP is solved. Finally, we study the trace 0 case and, using graph theoretic methods and EBL digraphs, characterize nonnegative irreducible realizability among realizable polynomials.
This talk is based on a joint work with C.R. Johnson and C. Marijuán.
Speaker: Miriam Pisonero (Universidad de Valladolid, Spain)
-
275
-
Rational Approximation and Interpolation: Practical Applications, Challenges and Solutions: G Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
278
A refined nonlinear least-squares method for the rational approximation problem
The adaptive Anderson-Antoulas (AAA) algorithm is capable of generating highly accurate rational approximations to given data. Though AAA almost always produces an approximation to a given target accuracy, the degree of the resulting rational function may be larger than actually required to meet the accuracy tolerance. In this talk, we introduce the nonlinear least-squares adaptive Anderson-Antoulas (NL-AAA) algorithm, which aims to solve the nonlinear least-squares problem arising in the AAA algorithm, as opposed to the linear approximation solved in AAA. The nonlinear problem is solved efficiently with iteratively reweighed least-squares methods. In addition to better accuracy at lower degrees, solving the nonlinear least-squares problem allows us to guarantee monotonic convergence of the NL-AAA method. Further, we provide an analysis of the gradients of each minimization problem, which gives insight into scenarios where AAA is observed to converge sub optimally. Finally, we test our algorithm on numerical examples including classic function approximation problems and applications to reduced order modeling, a field where attaining high accuracy with minimal degree is required.
Speaker: Michael Ackermann (Virginia Tech) -
279
Descriptor realizations of multi-parameter systems and nonlinear eigenvalue problems
Purpose of this presentation is to discuss novel descriptor realizations of linear multiple-parameter systems and their connection to nonlinear eigenvalue problems. The work is based on recent developments of the Loewner framework as exposed in a SIAM Review paper, published in November 2025.
Speaker: Athanasios Antoulas (Rice University) -
280
The Loewner Framework Beyond Linear Outputs
In recent years, structured reduced-order modeling has become an essential component in meaningful applications across engineering and the physical sciences whenever mathematical models are unavailable, but input-output data are abundant. For linear time-invariant systems, the Loewner framework provides a non-intrusive methodology for the construction of minimal interpolants from rational transfer function data. Its key ingredients are the Loewner and shifted Loewner matrices, which are constructed solely from transfer function evaluations. Significantly, the rank of the Loewner matrix reveals the underlying order (McMillan degree) of the system that generated the data, enabling a trade-off between the accuracy-of-fit and the complexity of the data-based reduced-order model.
In this work, we generalize the Loewner framework to the class of linear quadratic-output systems
$$E\dot{x}(t)=Ax(t)+Bu(t),~~y(t)= M(x(t)\otimes x(t)).$$
Dynamical systems with quadratic-output functions arise naturally in applications where one is interested in observing or simulating response quantities computed as the product of time- or frequency-components of the state, such as in vibro-acoustic problems or energy-based modeling. We introduce appropriately defined Loewner and shifted Loewner matrices that are built from samples of the multivariate rational transfer function of the underlying quadratic-output model. These matrices retain the hallmark features of the linear time-invariant Loewner approach; namely, their rank exposes the order of the system used to generate the data, and the reduced-order model computed using the proposed Loewner and shifted Loewner matrices is guaranteed to satisfy certain multivariate rational interpolation conditions. We also discuss practical considerations, such as obtaining quadratic-output transfer function data from frequency-response measurements of an equivalent, multiple-output linear time-invariant system.Speaker: Sean Reiter (Courant Institute of Mathematical Sciences, New York University)
-
278
-
Linear Algebra Education: G Torgersen Hall 2150
Torgersen Hall 2150
Virginia Tech
-
281
Visualizing the Spectral Theorem for Symmetric Matrices in a Dynamic Geometry Environment
Symmetric (real) and Hermitian (complex) matrices occupy a central role in linear algebra due to their well-known spectral properties, including real eigenvalues and mutually orthogonal eigenvectors. While these results are theoretically elegant, students often struggle to develop geometric intuition for eigenvalues, eigenvectors, and matrix transformations, especially in higher dimensions.
This talk presents a sequence of interactive visualizations of symmetric matrices using a dynamic geometry-assisted learning environment, illustrating how eigenvalues and eigenvectors behave in both two and three dimensions. Four representative cases are examined: a 2×2 invertible symmetric matrix with distinct eigenvalues, a 2×2 singular symmetric matrix with a zero eigenvalue, a 3×3 invertible symmetric matrix with three distinct eigenvalues, and a 3×3 singular symmetric matrix with repeated eigenvalues.
Through these examples, we visually demonstrate key theoretical properties of symmetric matrices, including the parallelism between eigenvectors and their matrix images, the scalar-multiplication effect of eigenvalues, and the orthogonality of eigenvectors corresponding to distinct eigenvalues. Special attention is given to the geometric interpretation of zero and repeated eigenvalues, highlighting how singular matrices collapse eigenvectors to the zero vector while preserving orthogonality. These visualizations directly illustrate the Spectral Theorem for symmetric matrices by making eigenvalues, eigenvectors, and orthogonal diagonalization visible through dynamic geometry software. By linking algebraic computations with concrete geometric transformations, the approach provides an intuitive and conceptually coherent framework for understanding the theorem, making it particularly effective for teaching and learning linear algebra.
Speaker: Günhan Caglayan (New Jersey City University, Mathematics Department) -
282
A Free Online Linear Algebra Textbook with Explorations that may Help your Students
In January 2023, the authors published the second edition of Linear Algebra: An Interactive Introduction, which the speaker shared at the ILAS Education Session at JMM in Boston.
Thanks to a $ 2.1 million grant from the United States Department of Education entitled “Fortifying Open Education: Scaling Ximera for Enduring Impact”, the authors have been able to enhance the online textbook in numerous ways. The book is now on its third edition, and the authors hope to release a fourth edition this fall.
This talk will showcase the Explorations which utilize the capabilities of Ximera. Some of the Octave exercises will also be demonstrated. Instructors who use other textbooks may benefit from using one or two of these activities in their classroom.
Linear Algebra: An Interactive Introduction can be accessed at https://sites.google.com/view/lin-alg-interactive-intro/homeSpeakers: Anna Davis (The Ohio State University), Paul Zachlin (Lakeland Community College) -
283
Balancing Theory and Application in Numerical Linear Algebra with Modern Computational Tools
This talk describes the design and features of a newly developed undergraduate course in numerical linear algebra and an accompanying instructor-led research study that examines how students engage with theoretical structure and computational behavior in AI-supported environments.
This course offered an introduction to numerical linear algebra, focusing on its theoretical foundations, practical applications, and computational techniques. Topics included matrix factorization (LU, QR, Cholesky, SVD), eigenvalues and eigenvectors, iterative methods, and least squares. Computational explorations in MATLAB were central to the course, allowing students to learn theories through computation and real-world applications, bridging the numerical and abstract linear algebra.
The instructor conducted educational research by collecting students’ work, regular reflections, and an exit survey, along with the instructor’s daily teaching reflections, to document classroom interactions and instructional decisions. The goal is to examine how students interpret numerical output, reason about algorithmic reliability, and incorporate AI-assisted tools while remaining grounded in theory. Some preliminary data and analysis will be shared in this talk.
The course was designed and created in collaboration with Mike Michailidis and Jon Loftin from MathWorks.
Speaker: Prof. Sepideh Stewart (University of Oklahoma)
-
281
-
Advances in Randomized Algorithms and Kernel Methods for Rank-Structured Matrices: G Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
284
Low-Rank Approximation by Randomly Pivoted LU
I will present Randomly Pivoted LU (RPLU), a randomized variant of Gaussian elimination with complete pivoting that samples pivots proportional to squared Schur-complement entries, and analyze its low-rank approximation properties. I will highlight two regimes where RPLU is particularly effective at low-rank approximation: (i) memory-limited settings, where a rank-$k$ approximation can be computed with $\mathcal{O}(k^2+m+n)$ storage and $\mathcal{O}(k^3+m+n+k\mathcal{M}(\mathbf{A}))$ work (with $\mathcal{M}(\mathbf{A})$ the cost of a matvec with $\mathbf{A} \in \mathbb{C}^{m \times n}$ or $\mathbf{A}^* $), and (ii) structured problems where $\mathbf{A}$ and its Schur complements admit fast updates (e.g., Cauchy-like matrices). I will discuss applications to fast computation of high-degree rational approximants.
Speaker: Marc Aurèle Gilles (Princeton University) -
285
Proxy Points and Rational Approximation Using Contour Integration
Due to the size of many kernel matrices that arise in applications, it is often necessary to work with their low-rank approximations in order to efficiently perform many computations. Low-rank matrix decompositions to such matrices may be quickly obtained by exploiting the analytic structure of the underlying kernel, for example by using Taylor expansions or an integral representation; such ideas trace back to the fast multipole method. However, there is often a gap between the theoretical epsilon-rank of a given kernel matrix, explicitly obtained using its eigenvalue decomposition, and the existing analytic decompositions of the same rank. In this work, we aim to bridge this gap by exploring the proxy point method, which approximates complex-analytic kernel matrices using a discretized contour integral representation. In particular, we combine this approach with a judicious choice of conformal map applied to the set of points at which the kernel is evaluated in order to obtain effective analytic approximations. We apply this new method to some well-studied kernel matrices, such as the Hilbert and Cauchy matrices, in order to compare it with existing analytic methods and with the best available theoretical bounds. In the process, we demonstrate a novel way of obtaining one kind of theoretically-optimal solution to a related problem in rational approximation.
Speaker: Mikhail Lepilov (Rensselaer Polytechnic Institute) -
286
Kernel-Based Variational Formulations of Nonlocal Mean-Field Games
We study mean-field games with nonlocal interactions modeled through kernel-based representations. Using feature-space expansions inspired by kernel methods, the resulting models admit a variational saddle-point formulation that is well suited for efficient primal–dual algorithms such as the primal-dual hybrid gradient method. We also discuss inverse problems in which interaction mechanisms are recovered from partial and noisy observations of population dynamics, emphasizing optimization-based formulations and splitting methods for stable and efficient reconstruction.
Speaker: Siting Liu (University of California, Riverside)
-
284
-
Recent Advances in Tensor Decompositions for Model and Data Reduction: G Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
287
Streaming Compression of Scientific Data through Weak-SINDy and POD Integration
The exponential growth of scientific data from simulations and experiments demands efficient compression techniques for storage and processing. This talk introduces a novel streaming weak-SINDy algorithm designed for real-time compression of streaming scientific data. Leveraging the underlying structure of physical systems, the algorithm constructs memory-efficient feature matrices and target vectors in real-time, enabling model recovery in an offline regression stage. For high-dimensional data, we integrate a streaming proper orthogonal decomposition (POD) process, dynamically reducing data dimensions and augmenting the streaming weak-SINDy algorithm to handle evolving POD bases. Proof-of-concept examples, including applications to the Lorenz system and fluid-flow data, demonstrate both the memory efficiency and reconstruction accuracy of the approach. Join us to explore how these advancements pave the way for scalable, real-time compression of scientific data.
Speaker: Dr Rick Archibald (Oak Ridge National Laboratory) -
288
Improving prediction for a low rank tensor ROM via continuous data assimilation
We extend a low rank tensor ROM recently developed by Olshanskii et al by enhancing it with continuous data assimilation (CDA). We show how CDA is easily incorporated into the ROM, and analytically show that it provides for theoretical long time error estimates. Numerical tests illustrate the theory and show it is an effective tool for simulating incompressible flow over a wide range of Reynolds numbers.
Speaker: Prof. Leo Rebholz (Clemson University) -
289
Derivative informed Tucker tensor train Taylor series surrogate models
We introduce Tucker tensor train Taylor series (T4S) surrogate models for high dimensional mappings that depend implicitly on the solution of a partial differential equation. Traditionally, Taylor series are intractable here because the derivative tensors are enormous, and are only accessible through multilinear actions. We overcome these challenges by approximating each derivative tensor with a Tucker decomposition composed with a tensor train, fitting each Tucker tensor train to symmetric tensor actions via Riemannian manifold optimization. We present theory and numerical experiments that validate the model and numerical method.
Speaker: Dr Nick Alger (University of Texas, Austin)
-
287
-
Polynomials, Krylov Methods and Applications: G McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
290
Spectral density estimation for random matrices
We introduce a new approach for estimating the asymptotic spectral distribution (ASD) of a random matrix using a single, sufficiently high-dimensional sample, without computing the full spectrum. The method builds on the Lanczos algorithm, together with asymptotic analysis and perturbation theory for orthogonal polynomials, and enables efficient and accurate estimation of the ASD. We illustrate the approach through an application to spectral density estimation in spiked covariance models.
Speaker: Charbel Abi Younes (University of Washington) -
291
On block Krylov and matrix polynomials
The deep connection between Krylov methods, scalar orthogonal polynomials, and moment matrices is well established, particularly for Hermitian and unitary matrices. In this talk, we extend this framework to block Krylov methods and orthogonal matrix polynomials.
By representing the elements of a block Krylov subspace via matrix polynomials, we consider the matrix-valued inner product introduced in [1], which, under a non-degeneracy assumption, defines a linear isometry. This establishes a one-to-one correspondence between orthonormal matrix polynomials and orthonormal bases of the block Krylov subspace.
For normal matrices, the block Gauss discretization [1,2] of such an inner product admits an integral representation familiar from the theory of orthogonal matrix polynomials. As an application, we extend a Szegő-type short recurrence, originally developed for matrix polynomials on the unit circle [3], to the block Arnoldi algorithm applied to unitary matrices.
Finally, we analyze the structure of the block moment matrix and explore its connection to orthogonal matrix polynomials and recurrence coefficients via a Cholesky-like factorization.
[1] Lund, K.: A New Block Krylov Subspace Framework with Applications to Functions of Matrices Acting on Multiple Vectors, Phd thesis, Temple University and Bergische Universität Wuppertal, (2018)
[2] Zimmerling, J., Druskin, V., Simoncini, V.: Monotonicity, bounds and acceleration of block Gauss and Gauss-Radau quadrature for computing $B^T\phi(A)B$, J. Sci. Comput., 103 (2025)
[3] Sinap, A., Van Assche, W.: Orthogonal matrix polynomials and applications, J. Comput. Appl. Math. 66 (1996)
Speaker: Michele Rinelli (KU Leuven) -
292
A Generalized Framework for Orthogonal Rational Functions applied to Rational Approximation
Rational functions are fundamental to several non-linear approximation problems in, for example, model reduction, system identification, and PDE problems. Consequently, one is often interested in constructing an orthonormal basis of rational functions to ensure numerical stability and conditioning.
In this talk, we present a generalized framework for constructing such bases for rational function vectors of arbitrary length k. We show that the underlying recurrence relations can be represented by a specific structured pair of k-Hessenberg matrices. We propose two efficient algorithms to compute these bases:
- An updating algorithm using rotations,
- A rational Arnoldi-type algorithm.
To demonstrate the robustness of this framework, we apply it to the rational approximation of $\sqrt{z}$ on [0,1]. We show that our method successfully handles exponentially clustered poles and recovers the optimal lightning + polynomial convergence rates.
Speaker: Robbe Vermeiren (KU Leuven)
-
290
-
12:00 PM
Excursion Departure Goodwin Hall Parking Lot
Goodwin Hall Parking Lot
Virginia Tech
-
12:30 PM
Excursions
-
6:00 PM
Departure to Conference Dinner Hyatt Place Parking Lot (Hyatt Place)
Hyatt Place Parking Lot
Hyatt Place
650 University City Blvd, Blacksburg, VA 24060 -
6:45 PM
Conference Dinner Eastern Divide Brewing Company (Eastern Divide)
Eastern Divide Brewing Company
Eastern Divide
3175 Commerce St, Blacksburg, VA 24060
-
251
-
-
293
Announcements McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
Plenary Talks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
294
When Structured Matrices Pay Off: Linear Algebra at the Heart of Wireless Communication
Millimeter waves within the sub-terahertz band offer a plethora of applications in next-generation wireless communication. However, they also introduce severe real-time and hardware limitations, making conventional wideband multi-beam beamforming exceedingly complex. For example, an $N$-element array using true-time delay beamformers needs $\mathcal{O}(N^2)$ time delays or phase shifts, while FFT-based beamformers experience the beam-squint problem. Thus, this lecture addresses how structured matrix analysis can lead to algorithmic and architectural solutions to address these obstacles.
The central object of the talk is the delay Vandermonde matrix (DVM), a new structured matrix whose elements are defined via true-time delays in the spatio-temporal domain. By exploring its structure, we derive fast DVM algorithms that reduce computational and hardware complexity from $\mathcal{O}(N^2)$ to $\mathcal{O}(N \log N)$ and also to $\mathcal{O}(N)$ for $N$ beams. These results position DVM-based beamformers as a superclass of FFT-based multibeam beamformers. Unlike FFT beams, DVM beamformers are inherently wideband and eliminate beam squint, while simultaneously reducing chip area, power consumption, and system cost.
Beyond this, DVM factorization enhances AI-chip design by using structured weight matrices in neural network layers, enabling real-time wireless communication. This model-driven approach constrains learning through a structured neural network, enabling real-time inference on massive data streams at gigahertz rates. Taken together, the lecture emphasizes that structured matrices connect algorithmic complexity, hardware feasibility, and future wireless communication powered by AI.
Speaker: Sirani M. Perera (Embry-Riddle Aeronautical University, USA) -
295
Numerical Ranges and Dilations: Theoretical Advances and Applied Perspectives
The numerical range is a fundamental tool for understanding the properties of matrices and operators. In this talk, we discuss recent advances in the study of the numerical range and its generalizations, specifically focusing on their utility in analyzing operator dilations. We demonstrate how these theoretical frameworks provide critical insights into applied topics, including quantum information theory and modern data science.
Speaker: Chi-Kwong Li (College of William and Mary)
-
294
-
10:15 AM
Conference Photo McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
10:25 AM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Contributed Talks: H Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
296
A-posteriori error estimates for randomized low-rank approximations Torgersen Hall 3100
Torgersen Hall 3100
Randomized algorithms in numerical linear algebra have proven to be effective in ameliorating issues of scalability when working with large matrices, efficiently producing accurate low-rank approximations. A key remaining challenge, however, is to efficiently assess the approximation accuracy of randomized methods without additional expensive matrix accesses.
In this talk, we discuss a posteriori error estimation strategies for randomized low-rank approximations, with a focus on estimators that can be constructed from the same data used to compute the approximation, including leave-one-out type estimators. These can serve both as certification tools and as algorithmic building blocks, enabling adaptive approximations and informed trade-offs between accuracy and computational cost.
Speaker: Lorenzo Lazzarino (University of Oxford) -
297
Randomized algorithms for operator trace estimation Torgersen Hall 3100
Torgersen Hall 3100
The implicit trace estimation is a problem of approximating the trace of a matrix $A$ accessed only through matrix-vector multiplication $x \mapsto Ax$, with the goal of using as few multiplications as possible to obtain an accurate approximation. Girard-Hutchinson's estimator computes the $\varepsilon$-approximation using $\mathcal{O}(\varepsilon^{-2})$ products, while its' variance-reducing improvement, known as Hutch++, only requires $\mathcal{O}(\varepsilon^{-2})$ products. In this talk, we propose an extension of those algorithms to estimate the trace of a trace-class operator by using operator-function multiplications. Our estimator is based on the notion of a Gaussian distribution on an arbitrary Hilbert space, so precise high-probability error bounds can be derived analogously to the matrix case estimators. We compare the performance of our method with an existing algorithm of this type, the ContHutch++ algorithm by Zvonek, Horning & Townsend. The theoretical results we prove concern mostly positive definite operators, but we provide empirical results that show potential in applications to general trace-class operators.
Speaker: Hrvoje Olić (Faculty of Science, University of Zagreb)
-
296
-
Matrix Geometries: H Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
298
Locally Diffeomorphic Logarithm of Special Orthogonal Matrices
The special orthogonal group $\mathbb{SO}_n$ is a Lie group whose geometry and local structure are encoded by the exponential map on its Lie algebra $\mathbf{Skew}_n$, the set of skew-symmetric matrices. The associated inverse problem---the matrix logarithm---exhibits a highly nontrivial local diffeomorphism structure, and the notion of a nearby logarithm arises naturally as a local inverse of the skew-restricted exponential. In this work, we establish a complete description of the local diffeomorphism structure of the exponential on the complement of tangent conjugate locus, the skew-symmetric matrices where the derivative of the exponential is rank-deficient. We show that the connected components of this complement admit a systematic organization into countably many diffeomorphic regions. By introducing a canonical alignment rule for Schur decompositions, we obtain a labeling and interpretation of these components. Based on this geometric framework, we propose an efficient and numerically stable algorithm for computing the nearby logarithm with two Schur decompositions. We further show that whenever a nearby logarithm exists, it coincides with the nearest Frobenius-norm preimage, providing a practical and reliable alternative for local inversion.
Speaker: Zhifeng Deng (Xiamen University) -
299
Special orthogonal, special unitary, and symplectic groups as products of Grassmannians
We describe a curious structure of the special orthogonal, special unitary, and symplectic groups that has not been observed, namely, they can be expressed as matrix products of their corresponding Grassmannians realized as involution matrices. We will show that $\text{SO}(n)$ is a product of two real Grassmannians, $\text{SU}(n)$ a product of four complex Grassmannians, and $\text{Sp}(2n, \mathbb{R})$ or $\text{Sp}(2n, \mathbb{C})$ a product of four symplectic Grassmannians over $\mathbb{R}$ or $\mathbb{C}$ respectively.
Speaker: Xiang Lu (University of Chicago) -
300
On the Identifiability of Transformer Self-Attention
Multi-head self-attention is a fundamental building block of the transformer architecture in modern machine learning, enabling large language models and much of modern generative AI as we know it. However, some aspects of the self-attention function space remain poorly understood. In particular, its parameterization is non-unique: continuous families of unique weight matrices can induce the same input–output map. This talk studies the self-attention function space, analyzing this non-uniqueness.
We study two common variants of self-attention. In linear attention, the attention map is a matrix of cubic polynomials in the inputs and tri-linear in the parameters, making it natural to use tools from algebraic geometry to describe generic fibers in the space. In softmax attention, a normalization breaks symmetries, leading to a different notion of generic identifiability, which we study with complex analysis.
Speaker: Nathan Henry (University of California, Berkeley)
-
298
-
The Inverse Eigenvalue Problem of a Graph and Zero Forcing: H Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
301
Resolving Inverse Singular Value Problems with Spoiler Spaces
Let $A$ be an $m \times n$ matrix. The spoiler space of $A$ is the set of all $m\times n$ matrices $X$ such that, $XA^\top$ is symmetric, $A^\top X$ is symmetric, and $X\circ A = O$, where "$\circ$" denotes the entrywise Schur product. If the spoiler space of $A$ contains only the 0 matrix, we say that $A$ has the Strong Singular Value Property (SSVP). The SSVP gives us access to a rich toolset for resolving inverse singular value problems. In this talk, we will discuss results pertaining to the spoiler space, its orthogonal complement, the tangent space, as well as the singular value multiplicity lists of various matrices and how they are connected.
Speaker: Caleb Cheung (University of Wyoming) -
302
The Inverse Symplectic Eigenvalue Problem and Coupled Zero Forcing for Graphs
Symplectic geometry appears in many areas of mathematics, physics, and applications, and naturally gives rise to interesting matrix families and properties. Symplectic eigenvalues extend the classical notion of eigenvalues to the symplectic setting and are guaranteed to exist for positive definite matrices by Williamson's theorem. We introduce the inverse symplectic eigenvalue problem for positive definite matrices whose zero and nonzero pattern is described by a labeled graph (ISEPG). In this talk, we define the ISEPG and present key tools developed to address it. We focus particularly on coupled graph zero forcing, a combinatorial technique used to bound the maximum symplectic eigenvalue multiplicity for a given graph. This is joint work with Leslie Hogben, Bryan Shader, and Tony Wong.
Speaker: Himanshu Gupta (University of Regina) -
303
An inverse eigenvalue problem for structured matrices determined by graph pairs
For vertex-labelled graphs $G$ and $H$ on $n\geq 1$ vertices, we consider matrices of the form $C(A,B) = \left[\begin{array}{c|c} A&B\\ \hline I&O\\\end{array}\right]\in\mathbb{R}^{2n\times 2n}$ where $A,B\in\mathbb{R}^{n\times n}$ are a pair of real symmetric matrices with nonzero patterns determined by the edges of the graph pair $G, H$. We denote the set of all such matrices by $\mathcal{S}(G,H)$. Our aim is to determine all possible spectra for $C(A,B)\in\mathcal{S}(G,H)$. We conjecture that $C(A,B)$ can attain any spectrum invariant under conjugation regardless of the chosen vertex-labelled graphs $G$ and $H$. In this talk, we highlight some results that support our conjecture.
Speaker: Dr Minerva Catral (Xavier University)
-
301
-
Numerical Ranges and Numerical Radii: H Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
304
On the ellipticity of the higher rank numerical range
The higher rank numerical range is investigated for 2-by-2 block matrices with associated Kippenhahn curves consisting of ellipses and eventually points. As a consequence, elliptical higher rank numerical range results are derived in a unified way, using an approach developed by Spitkovsky et al.
Speaker: Rute Lemos (CIDMA, University of Aveiro, Portugal) -
305
Numerical Range of Positive Hermitian Hankel Matrices
In 2022, Hwa-Long Gau and Pei Yuan Wu discussed the numerical range of various Hankel matrices with an emphasis on which subsets of the complex plane are attainable. We will be discussing as well as expanding upon these results when the Hankel matrices are positive and Hermitian.
Speaker: Brooke Randell (UCSC) -
306
The Cauchy Equation and Norm-Additive Mappings on Positive Cones of Commutative $C^{*}$-Algebras
The Cauchy functional equation plays a fundamental role in the study of additive and linear structures arising from numerical and norm-related information in functional analysis. In this talk, we investigate preserver problems on positive cones of commutative C$^{*}$-algebras, where a norm identity of Fischer--Muszély type, arising from the Cauchy functional equation, determines the underlying algebraic and geometric structure.
Let $A_i$ $(i=1,2)$ be commutative C$^{*}$-algebras, and let $A_i^{+}$ denote their positive cones. We consider surjective mappings $T:A_1^{+}\to A_2^{+}$, not assumed to be continuous, satisfying the norm identity
$$ \|T(a+b)\|=\|T(a)+T(b)\| \qquad (a,b\in A_1^{+}). $$ We show that every such mapping is necessarily additive and positive homogeneous. Moreover, in the unital case, if $T$ is injective, then it can be normalized to a composition operator, which in turn induces an isometric isomorphism between the underlying commutative C*-algebras $A_1$ and $A_2$. These results demonstrate that norm identities of Fischer--Muszély type on positive cones contain rich numerical and order-theoretic information, which completely determines the structure of commutative $\text{C}^*$-algebras.Speaker: Daisuke Hirota (National Institute of Technology, Tsuruoka College)
-
304
-
Inverse Problems and Uncertainty Quantification through the Lens of Numerical Linear Algebra: H Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
307
Projected Regularization in Low Precision
In recent years, mixed-precision and reduced-precision algorithms for solving large-scale linear systems have emerged as an effective approach for exploiting modern GPU architectures. While much of this work has focused on well-conditioned systems, comparatively little attention has been given to ill-posed inverse problems, where regularization is essential. In this talk we consider projected iterated Tikhonov regularization methods in reduced precision and show that these methods can produce reconstructions comparable to their high-precision counterparts. In addition, we discuss a secant-type update for automatic regularization parameter selection within the Golub–Kahan bidiagonalization framework, and demonstrate its effectiveness in the reduced-precision setting.
Speaker: Chelsea Drum (Emory University) -
308
Bayesian inference for rough feature reconstructions
Edges in imaging, that is sharp discontinuities in intensity, pose a significant challenge for inverse problems algorithms that often rely on Gaussian assumptions. Non-Gaussian heavy-tailed priors, which can better model the sparsity and sharp transitions inherent in edges, offer an alternative for edge-preserving image reconstructions. We consider the inherent difficulties in handling edges and highlight the potential of heavy-tailed prior models to transform this problem into a practical engineering solution.
Speaker: Prof. Lassi Roininen (LUT University) -
309
Data-driven discovery of chemical reaction networks
We propose a unified framework that allows for the full mechanistic reconstruction of chemical reaction networks (CRNs) from concentration data. The framework utilizes an integral formulation of the differential equations governing the chemical reactions, followed by an automatic procedure to recover admissible mass-action mechanisms from the equations. We provide theoretical justification for the use of integral formulations using analytical and numerical error bounds. The integral formulation is demonstrated to offer superior robustness to noise and improved accuracy in both rate-law and graph recovery when compared to other commonly used formulations. Together, our developments advance the goal of fully automated, data-driven chemical mechanism discovery.
Speaker: Abraham Reyes Velazquez (University of Manchester)
-
307
-
Code-based Cryptography: H Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
310
Relative generalized Hamming weights and their applications in cryptography
The relative generalized Hamming weights of a nested pair of linear codes are a generalization of the minimum distance. We will see how these parameters characterize the security of ramp secret sharing schemes, and how this can be adapted for private information retrieval. The computation of these parameters for a linear code is NP-hard in general, and we will study the most efficient current algorithm to compute them for any nested pair of linear codes.
Some of these results are joint work with Diego Ruano and Gonzalo Rodríguez-Pajares.
Speaker: Rodrigo San-José (Virginia Tech) -
311
Code Rigidity in Characteristic 2
Two linear codes are equivalent if there exists a monomial matrix that transforms one to the other. The problem of finding a monomial transformation from one code to another underlies the Linear Equivalence Signature Scheme (LESS). An automorphism of a linear code is a monomial matrix which fixes the code. When a code has a large number of automorphisms, it is easier to solve the linear code equivalence problem; thus, it is desirable for cryptography to have codes with as few automorphisms as possible. A code with the smallest possible number of automorphisms is called rigid.
Prior results show that almost all binary codes of dimension k and length 2k are rigid as k goes to infinity. We extend this result to arbitrary finite fields of characteristic 2.
Speaker: Adam Downs (Virginia Tech) -
312
Sample Complexity of the Matrix Code Equivalence Problem
The Matrix Equivalence Digital Signature (MEDS) is a code-based digital signature that was submitted to the NIST call for quantum-resistant protocols. It is currently considered as a candidate for building advanced group action signatures schemes.
The hard problem behind this digital signature is the Matrix Code Equivalence problem. Namely, given two matrix codes $C_1$ and $C_2$, suppose that there is a matrix pair $(A,B)$ where $AC_1B=C_2$ (left and right multiplication of matrices in $C_1$ by $A$ and $B$), how do we find the matrices $A$ and $B$?
This talk studies the matrix code equivalence problem from a sample complexity point of view. Since there is no known polynomial-time algorithm for finding $A$ and $B$ given one instance of the $(C_1, C_2)$ pair, how many such instances are enough to make a polynomial-time solution possible? In particular, we try to establish a relationship between the number of samples and the complexity of finding the secret pair $(A,B)$ using the XL (Extended Linearization) Algorithm.
Speaker: Wendi Gao
-
310
-
Linear Algebra Foundations for Data-driven Modeling and Model Order Reduction: H Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
313
Adaptive kernel methods
Kernel methods approximate nonlinear maps in a data-driven way by projecting the target map onto a finite-dimensional Hilbert space called the solution space. Traditionally, this space is a subspace of a fixed ambient reproducing kernel Hilbert space (RKHS), determined solely by the chosen kernel and the dataset, whose elements identify the basis elements. Consequently, the projection operator underlying the kernel method depends on the loss function, the dataset, and the choice of ambient RKHS. In this talk, we consider kernel methods whose solution spaces also depend on learnable parameters that are independent of the dataset. The resulting methods can be viewed as variable projection operators that depend on the loss function, the dataset, and the new learnable parameters instead of a fixed RKHS. We propose an efficient approximation of kernels associated with infinite-dimensional RKHSs, commonly used to reduce the solution-space dimension for large datasets. Second, we construct fixed-dimensional, parameter-dependent solution spaces that enable highly efficient kernel models suitable for large-scale problems without the need to approximate kernels of infinite-dimensional RKHSs. Our novel family of adaptive kernel methods generalizes earlier approaches, including Random Fourier Features, and we demonstrate their effectiveness through several numerical experiments.
Speaker: Matthias Voigt (UniDistance Suisse) -
314
Non-intrusive reduced-order models for parameterized partial differential equations using kernel methods
This talk presents an interpretable, non-intrusive reduced-order modeling technique for parameterized problems using regularized kernel interpolation. Parameterized reduced-order models (ROMs) enable the rapid approximation of PDE solutions corresponding to a given parameter, thus accelerating uncertainty quantification or inverse problem workflows requiring many PDE solves. Existing non-intrusive parameterized ROM approaches approximate the ROM dynamics by solving a data-driven least-squares regression problem for low-dimensional matrix operators. However, these approaches typically assume affine parametric dependence, which may not be satisfied by the underlying full-order model (FOM). To overcome this limitation, our approach leverages regularized kernel interpolation, which yields an optimal approximation of the ROM dynamics from a user-defined reproducing kernel Hilbert space and allows for arbitrary parametric dependence. We further show that our kernel-based approach can produce interpretable ROMs whose structure mirrors the parameterized FOM structure by embedding judiciously chosen feature maps into the kernel. The approach is demonstrated in several numerical experiments.
Speaker: Alejandro Diaz (Sandia National Laboratories) -
315
Matrix Structures for Certified Stability of Nonlinear Reduced Models
Learning compact surrogate models from data has become a major application area of machine learning techniques. Such models are required to describe dynamical behavior of processes in the presence of time series data and the absence of explicit mechanistic models. This may be the case if only measurement data is available or simulation data is obtained via proprietary software. Prominent techniques in this area are dictionary learning via sparse regression or operator inference which allow to infer nonlinear models describing the data in some optimal way. If stable behavior of the underlying process is expected, be it Lyapunov or global stability, or some sort of attractor dynamics, the classical system identification approaches can not guarantee this intrinsic property of the physical process. We discuss how several stability concepts can be encoded in the inferred model structure using structured matrices so that the obtained surrogate models are guaranteed to have the desired stability property. We discuss in particular quadratic and cubic nonlinearities as it is known that these are sufficiently expressive for many nonlinear dynamical systems. We also discuss the extension of the suggested encodings to parametric and control systems.
This talk is based on prior work with Pawan K. Goyal, Siddarth Mamidisetti, and Igor Pontes Duff.
Speaker: Peter Benner
-
313
-
Hierarchical Low-Rank Approximations: Algorithms and Applications: H McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
316
Superfast and stable divide-and-conquer singular value decomposition for hierarchical rank-structured matrices
This talk gives a superfast divide-and-conquer algorithm for computing the full singular value decomposition (SVD) of hierarchical rank-structured matrices with small off-diagonal ranks. The method achieves nearly linear complexity while delivering all singular values and singular vectors in structured forms. The structured representation of singular vectors enables near-linear operations with reduced storage. In contrast, classical algorithms for the full SVD require cubic computational cost and quadratic storage. The proposed method directly handles nonsymmetric or rectangular matrices by reducing them to a hierarchical block broken arrow form via stable QL factorizations. This form is then repeatedly decomposed through rank-1 SVD updates in the conquering stage. Several stability-preserving mechanisms are incorporated, including deflation, splitting and shifting, and orthogonality-preserving perturbations, to ensure the robustness of this stage. Efficiency is enhanced via fast kernel methods, such as the fast multipole method (FMM). A rigorous numerical error analysis establishes the backward stability of the overall process. Numerical experiments demonstrate significant advantages in both accuracy and efficiency, yielding the first fast and stable SVD for general hierarchical rank-structured matrices.
Speaker: Chenyang Cao (Purdue University) -
317
Unknown hierarchies, hyperbolic PDE, and randomized rank detection
We describe a problem arising from operator learning for hyperbolic PDEs, in which one would like to recover an unknown, non-standard low-rank hierarchical partition of a linear operator using only input-output data, or, in the finite-dimensional case, matrix-vector products. We provide a solution to the operator learning problem by employing a continuous analogue of the randomized SVD (RSVD) to decide whether the operator, restricted to a given subdomain, is numerically low-rank or not. Doing so requires the RSVD to obtain good singular subspace estimates, which in theory depends on the sizes of gaps between singular values of the operator. Thus, in the second part of this talk, we derive exact descriptions for the angular error of the approximate singular subspaces returned by the RSVD, which helps explain why large singular value gaps are typically not required in practice.
Speaker: Christopher Wang (Cornell University)
-
316
-
Eigenvalues of Nonnegative and Stochastic Matrices: H Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
318
New Results on the Doubly Stochastic Inverse Eigenvalue Problem
Stochastic matrices are matrices with nonnegative entries whose rows each sum to $1$. When a matrix and its transpose are both stochastic, it is said to be \emph{doubly stochastic}. In 1938, Kolmogorov proposed the problem of characterizing the region of possible eigenvalues of an $n \times n$ stochastic matrix, and Karpelevich gave a complete description thirteen years later. This talk concerns the doubly stochastic analogue: characterizing the region $\omega_n$ of eigenvalues of $n \times n$ doubly stochastic matrices, which is contained in the unit disk.
Perfect and Mirsky (1965) conjectured that $\omega_n$ is the union of the regions $\Pi_k$ (the convex hulls of the $k$-th roots of unity) for $k = 1,\dots,n$. This conjecture holds for $n = 1,2,3,4$, but fails for $n = 5$. The case $n \ge 6$ remains open. In response to the scarcity of progress over the past 60 years, Levick, Pereira, and Kribs proposed a related conjecture. In this talk, I prove a stronger version of this conjecture.
The approach, based on majorization theory and geometric properties, also provides a potential general framework for characterizing $\omega_5$, as well as a numerical method for testing the conjecture for various values of $n$.
Speaker: Ludovick Bouthat (Université Laval) -
319
Eigenvalue regions and realising matrices for monotone stochastic matrices
This talk presents spectral properties of monotone stochastic matrices which are characterised by the fact that each row stochastically dominates the preceding one, and which arise in contexts such as intergenerational mobility, equal-input models, and credit-rating systems.
In analogy with the stochastic matrices, for the monotone stochastic matrices both the individual eigenvalues as the spectrum as a whole are examined. Individually, the eigenvalue region for all $n \times n$ monotone matrices with $1 \leq n \leq 3$ is completely determined, and realising matrices are provided. Collectively, the set of possible pairs of non-trivial eigenvalues arising from $3 \times 3$ monotone matrices is characterised, accompanied by realising matrices. In both perspectives, the resulting regions are substantially smaller than those for general stochastic matrices. Finally, a reduction theorem is presented, stating that, for $n \geq 4$, the eigenvalue region of $n \times n$ monotone matrices is contained within that of $(n-1) \times (n-1)$ stochastic matrices.
Speaker: Brando Vagenende (Department Business Technology and Operations, Data Analytics Laboratory, Vrije Universiteit Brussel (VUB)) -
320
Spectrally Complete Subsets and Eigenvalue Regions of classes of Stochastic Matrices
Karpelevich’s theorem describes the single-eigenvalue region
$ \Theta_n=\{\lambda\in{\bf C}:\lambda\in\sigma(A)\ {\rm for\ some\ }A\in{\bf R}^{n\times n} \ {\rm row\mbox{-}stochastic}\}. $
The set of row-stochastic matrices is the polytope $\mathrm{conv}(V_n)$, where $V_n$ consists of the $n^n$
deterministic Markov kernels ($0$-$1$ matrices with exactly one $1$ in each row).
For a stochastic class $C=\mathrm{conv}(V)$, call $W\subseteq V$ spectrally complete if
$ \Theta(\mathrm{conv}(W))=\Theta(C), $ and minimal if it is inclusion-minimal.
Since these eigenvalue regions are typically star-shaped about $0$, spectral completeness is essentially radial, writing $ \rho_C(\theta)=\sup\{r\ge0:\, r e^{i\theta}\in\Theta(C)\}, $ equality $\rho_{\mathrm{conv}(W)}(\theta)=\rho_C(\theta)$ for all $\theta$ implies $\Theta(\mathrm{conv}(W))=\Theta(C)$.I develop this subset-selection viewpoint in three settings.
(1) Doubly stochastic matrices: vertices are permutation matrices; simulations up to $n=25$ support the Harlev-Johnson-Lim Boundary Conjecture, with no new phenomena beyond the exceptional $n=5$ behavior, and indicate that a very small family of permutation pairs is already spectrally complete.
(2) Monotone stochastic matrices: building on work of Vagenende et al., I focus on the
explicit boundary description for $n=3$ and explain how AI-assisted was used to
identify extremal families and supporting inequalities that were then verified rigorously.
(3) Prescribed zero patterns: in the nonnegative inverse eigenvalue problem with fixed support (Ran-Teng, dimension $3$), restricting the zero pattern yields another spectral subset-selection problem.I conclude by outlining how AlphaEvolve-style workflows could systematize the discovery ofbspectrally complete subsets and hopefully guide us towards alternative Karpelevich-type proofs.
Speaker: Brecht Verbeken (Vrije Universiteit Brussel)
-
318
-
Rational Approximation and Interpolation: Practical Applications, Challenges and Solutions: H Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
321
Rational quasi-Hermite approximation for computing acoustic quasiresonances in transmission problems
We introduce an AAA-type method for rational quasi-Hermite approximation formulated in barycentric form. A stacked Hermite–Löwner matrix is assembled from function values and derivative data at adaptively selected support nodes, and the barycentric weights are determined through a homogeneous least-squares procedure. This approach eliminates the need for external test points as required in the classical AAA algorithm. To maintain computational efficiency, the barycentric weights are computed using randomized nullspace methods, which accelerate the SVD of the stacked Hermite–Löwner matrices.
To further control computational costs and mitigate the appearance of spurious poles, the support set is adaptively partitioned into pieces. This yields piecewise rational quasi-Hermite approximants where inter-piece smoothness is enforced through penalization rather than exact spline constraints, significantly reducing the size of the underlying SVD problems and suppressing unwanted poles. The refinement strategy estimates the location of the maximal residual by constructing a Hermite interpolant of the residual data over the current support nodes. A complete MATLAB implementation is provided, and numerical experiments demonstrate that the method achieves accuracy and efficiency comparable to continuum AAA while requiring substantially fewer function evaluations when derivative information is available.
Speaker: Luka Marohnić (Zagreb University of Applied Sciences) -
322
Tensor-based multivariate rational approximation
Recent contributions for rational approximation include the p-AAA method (and variations) and the extension of the Loewner framework to multiple dimensions. These contributions are recent and are inspiration for further analysis and algorithmic improvements.
In this talk, we combine two ingredients that have proven to be successful in their respective contexts, i.e., the AAA method for univariate rational approximation, and tensor decompositions for the representation of low rank discrete multidimensional data. We present a tensor-based framework for multivariate rational approximation from function samples given on tensor-product grids. The approach combines a low-rank tensor decomposition with the set-valued AAA algorithm. We compare the proposed method against the current state-of-the-art p-AAA by numerical experiments. In particular, we examine both robustness and computational cost on challenging function classes, including nonsmooth and highly oscillatory functions. We also investigate performance in the presence of noise in the sampled data.
Speaker: Leonie Van Pottelberghe (KU Leuven) -
323
Greedy rational approximation: Analysis and algorithms of sketched resolvents
We discuss analytical estimates for greedy construction of rational approximations, where the underlying function is the linear sketch of an operator resolvent. The canonical example of this setup is the transfer function of a linear dynamical system. Under a sectorial assumption for the operator, this analysis immediately reveals corresponding algorithms, and provides explicit estimates of transfer function rational approximability. The theoretical constructions correspond to explicit Galerkin projection-based reduced order models of the original system, and additionally provide concrete estimates of Hankel singular value decay. The framework we discuss is a paradigm that forges concrete relationships between rational approximation, Kolmogorov n-widths, and Galerkin projection-based model reduction. We will discuss extensions to parametric systems, where this corresponds to multivariate rational approximation through a greedy approach. Finally, we will identify algorithmic challenges and opportunities, investigating numerical procedures that improve on stability and accuracy of the procedure.
Speaker: Akil Narayan (University of Utah)
-
321
-
Polynomials, Krylov Methods and Applications: H McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
324
Computing Functions of Rank-structured or Telescopic Matrices.
Appearing in a wide variety of applications, often in the context of discretized (fractional) differential and integral operators, Hierarchically Semiseparable (HSS) matrices have a number of attractive properties facilitating the development of fast algorithms [6,4].
For HSS matrices, the rank-structure is numerically preserved if $f(z)$ is well-approximated by a rational function. The underlying reason for this is that if $A$ is HSS, $A^{-1}$ shares the same structure [5]. Recently, Casulli, Kressner and Robol proposed an algorithm for computing matrix functions $f(A)$ where $A$ is HSS and symmetric [3]. This result is based on the equivalence between HSS structure and the existence of telescopic decompositions
$$A=A^{(L)}= D^{(L)}+U^{(L)}A^{(L-1)}{V^{(L)}}^T.$$ In our work we extend the procedure to general (non-symmetric) HSS matrices. The algorithm works recursively, viewing the telescopic decomposition as a sequence of low-rank updates to block-diagonal matrices $D^{(L-i)}$ for $i = 0,\dots,L$. At each step, after some careful permutation of the diagonal blocks, we employ Block Rational Krylov methods [1] to approximate the effect of the low-rank update while keeping the size of the matrices involved small. Moreover, this technique allows the efficient computation of $f(T)$ where $T$ is a Toeplitz. Up to an FFT, the off-diagonal blocks of a Toeplitz matrix satisfy a displacement equation. The low-rank of the right-hand side is linked to fast singular value decay, which, in turn, implies that $\Omega T\Omega^*$ can be well-approximated by an HSS matrix [2].[1] B. Beckermann et al., SIAM J. Numer. Anal., 59(3), 2021.
[2] B. Beckermann et al., Numerical Algorithms 100(4), 2025.
[3] A. A. Casulli et al., SIAM J. Matrix Anal. Appl., 45(4), 2024.
[4] S. Chandrasekaran et al., SIAM J. Matrix Anal. Appl., 27(02), 2005.
[5] A. Cortinovis et al., SIAM J. Matrix Anal. Appl., 43(1), 2022.
[6] W. Hackbusch, Hierarchical Matrices: Algorithms and Analysis, 49, 2015.Speaker: Andrea Baleani (Scuola Normale Superiore di Pisa) -
325
(Block) Lanczos Function Approximation for Quasi-Newton Optimization Algorithms
The Lanczos process is a well-known Krylov subspace method for the orthogonal tridiagonalization of a hermitian matrix $\mathbf{Z}$. Equipped with a suitable function $f$, Lanczos function approximation (LFA) can be used as a powerful tool for approximating the matrix-function $f(\mathbf{Z})$ or matrix-function-vector products $f(\mathbf{Z})\mathbf{\omega}$.
We discuss an application of LFA in computing iterates of the form
$$\mathbf{x}_{(k+1)} = \mathbf{x}_{(k)} - \eta_{(k)}\left(\left(\nabla^2g(\mathbf{x}_{(k)})\right)^2 + \lambda_{(k)}\mathbf{I}\right)^{-1/2} \nabla g(\mathbf{x}_{(k)})$$ arising in the regularized saddle-free Newton (R-SFN) algorithm for $g\in\mathcal{C}^2$. LFA facilitates efficient update steps yielding competitive performance against adaptive regularization with cubics (ARC), which is among the best available algorithms for non-convex optimization.
Additionally, we will detail ongoing work for improving LFA in the general setting, motivated by our application in optimization. We will discuss residual estimation for iterative refinement and block Lanczos for implicit preconditioning.
Speaker: Cooper Simpson (University of Washington) -
326
Everything is Vecchia: Unifying low-rank and sparse inverse approximations
The partial pivoted Cholesky approximation accurately represents matrices that are close to being low-rank. Meanwhile, the Vecchia approximation accurately represents matrices with inverse Cholesky factors that are close to being sparse. What happens if a partial Cholesky approximation is combined with a Vecchia approximation of the residual? We show how the sum can be rewritten as a Vecchia approximation of the original matrix with an amplified sparsity pattern. Thus, the Vecchia approximation is a superset for other factored matrix approximations and it has broad applicability.
Speaker: Robert Webber (University of California San Diego)
-
324
-
New Advancements in Tensor Decomposition and Computation: H Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
327
Multiscale Grassmann Manifolds for Single-Cell Data Analysis
Single-cell data are typically represented in Euclidean space, which limits the ability to capture intrinsic correlations and multiscale geometric structure. We propose a multiscale framework based on Grassmann manifolds that represents single-cell RNA-seq data through subspace geometry across multiple scales.
Speaker: Dr Xiang Xiang Wang (Michigan State University) -
328
Revisit CP Tensor Decomposition: Statistical Optimality and Fast Convergence
Canonical Polyadic (CP) tensor decomposition is a fundamental technique for analyzing high-dimensional tensor data. While the Alternating Least Squares (ALS) algorithm is widely used for computing CP decomposition due to its simplicity and empirical success, its theoretical foundation, particularly regarding statistical optimality and convergence behavior, remain underdeveloped, especially in noisy, non-orthogonal, and higher-rank settings.
In this work, we revisit CP tensor decomposition from a statistical perspective and provide a comprehensive theoretical analysis of ALS under a signal-plus-noise model. We establish non-asymptotic, minimax-optimal error bounds for tensors of general order, dimensions, and rank, assuming suitable initialization. To enable such initialization, we propose Tucker-based Approximation with Simultaneous Diagonalization (TASD), a robust method that improves stability and accuracy in noisy regimes. Combined with ALS, TASD yields a statistically consistent estimator. We further analyze the convergence dynamics of ALS, identifying a two-phase pattern—initial quadratic convergence followed by linear refinement. We further show that in the rank-one setting, ALS with an appropriately chosen initialization attains optimal error within just one or two iterations.
Speaker: Mr Augustine (Runshi) Tang (University of Wisconsin-Madison) -
329
The Nondecreasing Rank
Matrix or tensor data often has structured rows, columns, or more generally modes. In particular, a mode may have a natural ordering that can be leveraged to obtain parsimonious representations of the data. To this end, the concept of the nondecreasing (ND) rank is introduced in this talk. A tensor has an ND rank of r if it can be represented as a sum of r outer products of vectors, with each vector satisfying a monotonicity constraint. It is shown that for certain poset orderings finding an ND factorization of rank r is equivalent to finding a nonnegative rank r factorization of a transformed tensor. However, not every tensor that is monotonic has a finite ND rank. Theory is developed describing the properties of the ND rank, including typical, maximum, and border ND ranks. Highlighted also are the special settings where a matrix or tensor has an ND rank of one or two. As a means of finding low ND rank approximations to a data tensor a variant of the hierarchical alternating least squares algorithm is introduced.
Speaker: Dr Andrew McCormack (University of Alberta)
-
327
-
12:15 PM
Lunch Break
-
330
ILAS Education Workshop: Bridging Abstract and Numerical Linear Algebra Torgersen Hall 2150
Torgersen Hall 2150
Virginia Tech
This workshop will discuss ways of bridging abstract and numerical linear algebra. Our topics are included, but not limited to, the following:
• How to introduce numerical linear algebra to students (some examples of coding and programming will be provided),
• The role of AI in teaching and learning linear algebra,
• The role of technology in teaching linear algebra,
• Collaboration, creating and sharing resources (GitHub), lesson plans, and publications
Everybody is welcome!See the associated website
Speakers: Mike Michailidis (Mathworks), Rachel Quinlan (University of Galway), Sepideh Stewart (University of Oklahoma) -
Contributed Talks: I1 McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
331
On the Choice of Sign Defining Householder Transformations
It is well known that, when defining Householder transformations, the correct choice of sign in the standard formula is important to avoid cancellation and hence numerical instability. In this talk we point out that when the "wrong" choice of sign is used, the extent of the resulting instability depends in a somewhat subtle way on the data leading to cancellation.
Speaker: Michael Overton (New York University) -
332
Fast computation of eigenvalues of periodic CMV matrices
Periodic CMV matrices are unitary matrices that can be specified by $O(n)$ data. Their eigenvalues can be computed by standard methods, storing them as conventional matrices (using $O(n^{2})$ data) in $O(n^{3})$ time. Since periodic CMV matrices can be specified by $O(n)$ data, one would hope to find a method that computes the eigenvalues in $O(n^{2})$ time instead of $O(n^3)$. This is indeed possible, and we will show how to do it.
Speaker: David S Watkins (Washington State University)
-
331
-
Contributed Talks: I2 Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
333
Newton's Method for Computing the CP Decomposition
In this talk, I will address using Newton's Method to compute the CP tensor decomposition. The CP optimization problem is a nonlinear least squares problem with factor matrices as the variables. The most common methods for solving CP are Alternating Least Squares (ALS) and Gauss-Newton optimization combined with an iterative method for solving linear systems. I will show that one iteration of Newton's Method, when combined with an iterative linear solver, can achieve similar computational cost to ALS and Gauss-Newton for large, dense tensors. I will discuss the derivation for using Newton's method, how we achieve efficient computation, as well as experimental comparisons among these three methods.
Speaker: DuBose Tuller -
334
A Superfast Direct Solver of Type-III Inverse Nonuniform Discrete Fourier Transform Problem
The nonuniform discrete Fourier transform (NUDFT) and its inverse are widely used in various fields of scientific computing. In this article, we propose a novel superfast direct inversion method for type-III NUDFT. The proposed method approximates the type-III NUDFT matrix as a product of a type-II NUDFT matrix and an HSS matrix, where the type-II NUDFT matrix is further decomposed into the product of an HSS matrix and an uniform discrete Fourier transform (DFT) matrix as in [Wilber, Epperly, and Barnett, SIAM Journal on Scientific Computing, 47(3):A1702-A1732, 2025]. This decomposition enables both the forward application and the backward inversion to be accomplished with quasi-linear complexity. The fast inversion can serve as a high-accuracy direct solver or as an efficient preconditioner. Additionally, we provide an error bound for the approximation under specific sample distributions. Numerical results are presented to verify the relevant theoretical properties and demonstrate the efficiency of the proposed methods.
Speaker: Mr Jingyu Liu (Fudan University) -
335
Quantum Spectral Methods for Poisson and Heat Equations
We present a quantum framework for solving a large class of elliptic and parabolic Partial Differential Equations (PDEs) endowed with periodic conditions. We solve the Poisson equation $\Delta u = f$ and the heat equation on the $d$-dimensional flat torus using a Fourier spectral method implemented on quantum circuits.
The main contribution is an efficient use of block encoding to load the inverse diagonal spectral filter into a quantum circuit. Unlike standard quantum algorithms based on general matrix inversion, our approach constructs the reciprocal of the diagonal entries directly. This procedure requires only one ancillary qubit and can be implemented in polylogarithmic time with respect to the system size $N$.
Different strategies are adopted for solving the problem, depending on the type of equation. For elliptic equations, the solution is obtained by directly applying the inverse spectral filter to the source term. For parabolic equations, we employ an iterative implicit scheme where the spatial variables are handled via our quantum spectral method, while the update for the next time step is performed classically. We validate our approach by
comparing the energy evolution and steady-state solutions against standard classical spectral schemes. The numerical results demonstrate that our quantum solver achieves high fidelity, with errors comparable to machine precision limits.Speaker: Giacomo Antonioli (Universita di Pisa)
-
333
-
Spectral Graph Theory: I Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
336
Spectral theory of $K_t$-decomposable graphs
Given a real number $q$, the $q$-Laplacian of a graph $G$ is the matrix $A+qD$ where $A$ is the adjacency matrix and $D$ the diagonal degree matrix. If the edge set of $G$ can be partitioned into edge-disjoint copies of $K_t$, then $G$ is called $K_{t}$-decomposable.
In this talk, we generalize some results from a survey paper of Cvetković, Rowlinson, and Simić about the signless Laplacian $A+D$ to the $\frac{1}{t-1}$-Laplacian of $K_t$-decomposable graphs. In particular, we generalize the correspondence between the signless Laplacian and line graphs, positive semi-definiteness, and characterization of graphs with a zero eigenvalue.
Speaker: Mr Matt Burnham (Iowa State University) -
337
Connectivity of distance-regular graphs
In 1985, Brouwer and Mesner proved that the vertex-connectivity of a strongly regular graph equals its valency and the only disconnecting sets of this size are point neighborhoods. In 2009, Brouwer and Koolen generalized this result to distance-regular graphs. In 1996, Brouwer conjectured that the minimum size of a disconnecting set of vertices whose removal disconnects a connected strongly regular graph into non-singleton components equals the size of the neighborhood of an edge. In 2014, Cioaba, Kim, and Koolen disproved Brouwer's conjecture, but also showed that the conjecture is true for many families of strongly regular graphs. In their 2016 survey, van Dam, Koolen, and Tanaka asked whether Brouwer's conjecture is true for distance-regular graph with diameter at least three. In this talk, I will describe our progress regarding this restricted vertex-connectivity of distance-regular graphs.
Speaker: Isabel Byrne (University of Delaware) -
338
The non-existence of Moore polygons and spectral Moore bounds
The second largest eigenvalue of a graph's adjacency matrix captures important structure and spectral properties including connectivity and expansion. From the Alon-Boppana theorem, we know that if $\theta<2\sqrt{k-1}$, then there are only finitely many $k$-regular graphs with the second largest eigenvalue at most $\theta$. This motivates the following natural question posed by Richey, Stover, and Shutty in 2013. Given $k\geq 3$ and $\theta<2\sqrt{k-1}$, what is the maximum number of vertices $v(k,\theta)$ of a $k$-regular graph $G$ with $\lambda_2(G)\leq \theta$? This problem can be seen as a spectral version of the Moore (or degree-diameter) problem. In this talk, I will present recent progress on this question. This is joint work with Cioab\u{a}, Nozaki, and Xiang.
Speaker: Vishal Gupta (University of Rochester) -
339
Delsarte designs and Galois groups
Delsarte theory has been extended to the study of subsets in an increasing variety of association schemes in recent years, with many different motivations and applications. Tools originally developed for the study of error-correcting codes in the Hamming scheme and combinatorial $t$-designs in the Johnson scheme apply equally well in association schemes with irrational eigenvalues. The goal of the talk is to study the Delsarte $T$-designs in an assocaition scheme with irrational eigenvalues. We show that, for any subset $T$ of eigenspaces, a $T$-design must also be a $T'$-design where $T'$ the closure of $T$ under the action of the Galois group of the splitting field of the association scheme, acting on primitive idempotents. Conjugacy class schemes of finite groups will be the main examples explored. This is based on joint work with Jesse Lansdown (Galway).
Speaker: William Martin (Worcester Polytechnic Institute)
-
336
-
The Inverse Eigenvalue Problem of a Graph and Zero Forcing: I Goodwin Hall 155
Goodwin Hall 155
Virginia Tech
-
340
Multi-digraphs with maximum nullity at most one
For a multi-digraph $D$ with vertex-set $V=\{1,\ldots,n\}$ and arc-set $A$, let $Q(D)$ be the set of all real $n\times n$ matrices $A=[a_{i,j}]$ with $a_{i,j}\not=0$ if $i\not=j$ and there is a single arc from $i$ to $j$, $a_{i,j}\in \mathbb{R}$ if $i\not=j$ and there are multiple arcs from $i$ to $j$, $a_{i,j}=0$ if $i\not=j$ and there is no arc from $i$ to $j$, $a_{i,i}\not=0$ if there is no loop at vertex $i$, and $a_{i,i}\in \mathbb{R}$ if there is at least one loop at vertex $i$. The maximum nullity $M(D)$ of a multi-digraph $D$ is the maximum nullity of any $A\in Q(D)$. In this talk, we show progress on the characterization of multi-digraphs $D$ with $M(D)\leq 1$.
Joint work with Marina Arav and Cherine Fons
Speaker: Dr Hein Van der Holst (Georgia State University) -
341
Tools for Determining the Minimum Rank of a Graph
The minimum rank of a graph $G$ of order $n$ is the smallest possible rank over all real symmetric $n\times n$ matrices $A$ whose $(i,j)$th entry, for $i\neq j$, is nonzero whenever $ij$ is an edge of $G$ and zero otherwise. We discuss some refinements of techniques currently in the literature to determine the minimum rank of a graph, some new tools to bound this value, and an approach for understanding the gaps in values between parameters used to bound the minimum rank of a graph. We also discuss implementations of known techniques, algorithmic improvements, and applications in computer assisted experimentation for the minimum rank problem.
Speaker: Mark Hunnell (Winston-Salem State University) -
342
Graph Products to Achieve few Distinct Eigenvalues
One of the most challenging aspects of graph inverse eigenvalue problems is knowing when a graph admits a matrix with few distinct eigenvalues. We will discuss results where we construct matrices with only two distinct eigenvalues corresponding to graphs that arise from various kinds of product structures. We will discuss also some directions for future work along these lines.
Speaker: Mark Kempton (Brigham Young University) -
343
A Graphical Approach to Isospectral Unfoldings
An isospectral reduction is a method of shrinking a large matrix into a smaller matrix while preserving properties of the original's spectrum. The inverse, an isospectral unfolding, takes a matrix of an isospectral reduction and expands it into a larger matrix that has that reduction. We present a system of nonlinear equations forming the foundation of general isospectral unfolding. Graphs naturally tie in to isospectral reductions via their adjacency matrices. Thus, automorphic orbits, equitable partitions, and their relations to graphs sharing the same isospectral reductions will be explored, theorems given, and proofs discussed. Connections to quantum walk matrices will also be made.
Speaker: Dallin Seyfried (Brigham Young University)
-
340
-
Inverse Problems and Uncertainty Quantification through the Lens of Numerical Linear Algebra: I Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
344
Discretization-free Bayesian inverse problems
In this talk, we revisit the Bayesian inverse problems formalism in infinite-dimensional distribution spaces, where function evaluations are replaced by evaluations by test functions. It is shown that linear inverse problems can be formulated without a reference to any infinite-dimensional representation of the unknown, e.g., in terms of basis vectors, and therefore, the forward problem has a matrix-free form. In the Gaussian case, the numerical evaluation of the posterior mean and covariance matrix does not rely on finite-dimensional approximations of the unknown, but rather on numerical quadratures. The formalism is flexible, allowing a posteriori probing of the unknown without recalculation of any matrix inverses.
Speaker: Erkki Somersalo (Case Western Reserve University) -
345
Parametric Hyperbolic Conservation Laws: Learning Hyperbolic Conservation Laws from Data through Symmetrization.
We propose a parametric hyperbolic conservation law (SymCLaw) for learning hyperbolic systems directly from data while ensuring conservation, entropy stability, and hyperbolicity by design. Unlike existing approaches that typically enforce only conservation or rely on prior knowledge of the governing equations, our method parameterizes the flux functions in a form that guarantees real eigenvalues and complete eigenvectors of the flux Jacobian, thereby preserving hyperbolicity.
At the same time, we embed entropy-stable design principles by jointly learning a convex entropy function and its associated flux potential, ensuring entropy dissipation and the selection of physically admissible weak solutions. A corresponding entropy-stable numerical flux scheme provides compatibility with standard discretizations, allowing seamless integration into classical solvers. Numerical experiments on benchmark problems, including Burgers’, shallow water, Euler, and KPP equations, demonstrate that SymCLaw generalizes to unseen initial conditions, maintains stability under noisy training data, and achieves accurate long-time predictions, highlighting its potential as a principled foundation for data-driven modeling of hyperbolic conservation laws.Speaker: Dr Lizuo Liu (Dartmouth College) -
346
Interpolation-Based Estimation and Uncertainty Quantification of Periodic Time-Varying Parameters
Many applications in modern day science involve unknown system parameters that must be estimated from limited data. A subset of these problems involves parameters that vary with time but have unknown evolution models and cannot be directly observed. In this work, we formulate time-varying parameter estimation in deterministic dynamical systems as an interpolation problem, where the function values at the node locations are unknown and must be estimated to obtain a polynomial-based approximation of the unknown parameter. We discuss the numerical considerations that arise when solving the linear systems in these approximations, along with several computed examples.
Speaker: Dr Andrea Arnold (Worcester Polytechnic Institute) -
347
Streaming Algorithms for Big Data Inverse Problems
Big data applications are becoming ever more prominent, and in many applications we need to solve very large linear or nonlinear inverse problems while handling only a relatively small amount of data at a time. Moreover, we are interested in distributed, possibly asynchronous, algorithms that solve large problems while only exchanging limited information. We need algorithms that combine approximate (partial) solutions with incoming data or data read from secondary memory to incrementally further improve the solution. We will discuss several algorithmic variations and their convergence.
Speaker: Eric de Sturler (Virginia Tech)
-
344
-
Theoretical Advances in Operator Learning: I McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
348
Operator Learning at Machine Precision
Neural operator learning methods have garnered significant attention in scientific computing for their ability to approximate infinite-dimensional operators. However, increasing their complexity often fails to substantially improve their accuracy, leaving them on par with much simpler approaches such as kernel methods and more traditional reduced-order models. In this article, we set out to address this shortcoming and introduce CHONKNORIS (Cholesky Newton--Kantorovich Neural Operator Residual Iterative System), an operator learning paradigm that can achieve machine precision. CHONKNORIS draws on numerical analysis: many nonlinear forward and inverse PDE problems are solvable by Newton-type methods. Rather than regressing the solution operator itself, our method regresses the Cholesky factors of the elliptic operator associated with Tikhonov-regularized Newton--Kantorovich updates. The resulting unrolled iteration yields a neural architecture whose machine-precision behavior follows from achieving a contractive map, requiring far lower accuracy than end-to-end approximation of the solution operator. We benchmark CHONKNORIS on a range of nonlinear forward and inverse problems, including a nonlinear elliptic equation, Burgers' equation, a nonlinear Darcy flow problem, the Calder\'{o}n problem, an inverse wave scattering problem, and a problem from seismic imaging. We also present theoretical guarantees for the convergence of CHONKNORIS in terms of the accuracy of the emulated Cholesky factors. Additionally, we introduce a foundation model variant, FONKNORIS (Foundation Newton--Kantorovich Neural Operator Residual Iterative System), which aggregates multiple pre-trained CHONKNORIS experts for diverse PDEs to emulate the solution map of a novel nonlinear PDE. Our FONKNORIS model is able to accurately solve unseen nonlinear PDEs such as the Klein--Gordon and Sine--Gordon equations.
Speaker: Aras Bacho (California Institute of Technology) -
349
Operator Learning via Learned Differential Operators
We present an alternative perspective on operator learning for problems in which the operator is implicitly defined by a partial differential equation. Rather than learning the solution operator directly as a high-dimensional mapping, we propose to first learn the underlying PDE operator as a local differential operator and then numerically invert it to evaluate the associated solution operator.
Speaker: Juan Felipe Osorio Ramirez (University of Washington) -
350
Towards High-Precision Optimizers for Scientific Machine Learning
An increasing number of theoretical results are available to characterize the extent to which neural networks can (i) represent scientifically relevant functions and operators, and (ii) learn these functions and operators from data. However, even with the right network architecture and the right dataset, optimization is a bottleneck. On the one hand, popular machine learning optimizers such as SGD and Adam are not designed to attain the levels of precision needed for scientific applications. On the other hand, traditional approaches to high-precision optimization, such as Newton or quasi-Newton methods, are not designed for the highly stochastic training regimes that are characteristic of machine learning. In this talk, I will argue that these limitations can be overcome by adopting randomized linear algebra as a paradigm for high-precision scientific machine learning. As an example, I will share some recent work which uses this paradigm to both explain and improve a promising class of optimizers known as subsampled natural gradient algorithms.
Speaker: Gil Goldshlager (UC Berkeley) -
351
Estimation for intrinsic Gaussian processes
Gaussian processes (GPs) defined through intrinsic random fields provide a flexible framework for modeling spatial phenomena, and have been advocated in a variety of applications over the past several decades. Nevertheless, their adoption has lagged behind traditional, covariance-based approaches, in part because the intrinsic formulation has lacked an accompanying toolkit of computational methods and dependence specifications that facilitate fitting and prediction. This work develops a systematic framework for modeling intrinsic GPs and introduces practical algorithms for working with dependence/variogram models for modeling, inference and computation that parallel those of traditional, stationary GPs, highlighting the advantages of intrinsic-field modeling in terms of robustness, interpretability, and computational efficiency.
Speaker: Christopher Beattie (Virginia Tech)
-
348
-
Linear Algebra Foundations for Data-driven Modeling and Model Order Reduction: I Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
352
Approximate Deconvolution and Spatial Filtering of Reduced Order Models for Fluid Flow
This talk focuses on the numerical analysis of regularized projection-based reduced-order models (ROMs) for turbulent fluid flows. Direct numerical simulations are well known to be computationally infeasible for routine simulations in computational fluid dynamics, particularly at high Reynolds numbers. Reduced-order models offer an efficient low-dimensional framework capable of producing fast and reasonably accurate approximations of the full-order dynamics.
For under-resolved flows, which are typical in high-Reynolds-number regimes, standard POD-ROMs often suffer from loss of accuracy and numerical instability. Regularization techniques based on spatial filtering have been shown to improve both stability and accuracy with negligible additional computational cost. However, depending on the choice of filter and regularization strategy, these methods may introduce excessive numerical diffusion, leading to over-smoothing of the resolved dynamics.
To address this issue, we propose an approximate deconvolution approach designed to recover attenuated flow features and mitigate the over-smoothing effects induced by spatial filtering. The resulting framework enhances ROM accuracy while retaining the stabilizing benefits of regularization. Analytical properties and numerical results are presented to demonstrate the effectiveness of the proposed method for turbulent flow simulations.
Speaker: Jorge Reyes (Virginia Tech) -
353
Randomized Linear Algebra for Symplectic Model Order Reduction of Hamiltonian Systems
Large-scale Hamiltonian dynamics are governed by
$$ \dot x = J\nabla H(x),\quad J=\begin{bmatrix}0&I\\-I&0\end{bmatrix}, $$ and arise from spatial discretizations of conservative PDEs as well as in molecular and multibody models. In multi-query, control, and real-time settings, projection-based model order reduction (MOR) is essential, but generic reduced spaces may destroy the symplectic geometry, leading to energy gain or loss and poor long-time behavior. Symplectic MOR therefore seeks a symplectic trial basis $V\in\mathbb{R}^{2n\times 2r}$ with $V^TJV=J_{2r}$, so that the reduced system remains Hamiltonian. A key offline bottleneck of symplectic MOR is symplectic basis generation from snapshot data, commonly via the complex SVD (cSVD) or the SVD-like decomposition, which can dominate the cost when $n$ and/or the number of snapshots are large.Building on randomized numerical linear algebra, we present basis construction methods that use random sketching to accelerate the matrix factorizations underlying these basis generation techniques, while preserving symplecticity by construction. In particular, the randomized cSVD (rcSVD) and a randomized SVD-like decomposition yield substantial speedups while retaining accuracy close to their deterministic counterparts in numerical tests [1].
Beyond efficiency, we highlight a priori error analysis for rcSVD-based symplectic MOR. The resulting bounds relate the projection error to sketch size (oversampling) and power-iteration depth, thereby quantifying the runtime--accuracy trade-off, motivating practical parameter choices [2].
References:
[1] R. Herkert et. al. Randomized symplectic model order reduction for Hamiltonian systems. In LSSC 2023, pages 99--107. Springer, Cham, 2024.
[2] R. Herkert et. al. Error analysis of randomized symplectic model order reduction for Hamiltonian systems. LAA, 729:67--99, 2026.
Speaker: Robin Herkert (University of Stuttgart) -
354
Reduced Order Modeling of Conservation Laws via CDT
In this work, we propose a reduced-order modeling (ROM) framework for conservation laws that operates in the Cumulative Distribution Transform (CDT) domain. The CDT maps nonnegative, unit-mass states to an $L^2$ Hilbert space, in which pure translations become affine lines and $W_2$ (optimal transport) distances coincide with Euclidean distances. This linearization dramatically improves the approximability of advective dynamics: for linear transport, the solution manifold is exactly two-dimensional in CDT space $(d_2 = 0)$, and for advection–diffusion we show that, once an initial layer is excluded, the CDT trajectory has uniformly bounded curvature, implying quadratic Kolmogorov $n$-width decay with $n$.
Building on these insights, we develop a CDT–POD algorithm: (i) map snapshots to CDT space, (ii) compute a Proper Orthogonal Decomposition (POD) basis in that space, (iii) evolve reduced coefficients via a transform-space surrogate, and (iv) invert the CDT to recover physical states. Numerical experiments on linear advection, advection–diffusion, and both viscous and inviscid Burgers equations show that CDT–POD attains rank-2 exactness for transported signals, robustly resolves moving shocks and contact discontinuities with orders of magnitude fewer modes than Eulerian POD, and remains competitive in smoothly diffusive regimes. The method preserves mass by construction and is compatible with positivity.
Speaker: Sarswati Shah -
355
Solution of generalized tensor Lyapunov equations arising in optimal control and model reduction
Hamilton-Jacobi-Bellman partial differential equations (HJB PDEs) arise in various settings in optimal control and model order reduction, and their solutions are notoriously difficult to acquire. For linear time-invariant systems, the HJB PDEs of interest typically simplify to matrix algebraic equations, such as the algebraic Riccati equation or the matrix Lyapunov equation, for which many mature, reliable, and scalable solvers exist.
For weakly nonlinear systems with polynomial dynamics, a popular approach for locally approximating solutions to the HJB PDEs is the method of Al'brekht, which amounts to computing a Taylor series approximation. The first term in the Taylor expansion is given by the solution to a matrix Riccati (or Lyapunov) equation, and then the higher-order terms in the expansion are given by solutions to very large linear algebraic equations.
Despite being dense, it turns out that these equations exhibit a surprising amount of structure reminiscent of the matrix Lyapunov equation, and they can be considered as tensor Lyapunov equations. In this talk, we expand on related works by considering the generalized tensor Lyapunov equation that arises due to the presence of an invertible but non-identity mass matrix. We describe the changes to the problem associated with the inclusion of the mass matrix, some surprising beauty that emerges in the form of the fractal structure of the Sierpinski triangle, and the resulting obstacles that are introduced that preclude the use of existing solvers, before concluding with a new proposed algorithm for solving these structured systems efficiently. We also highlight the open problems and opportunities for future research on the topic.Speaker: Boris Kramer (University of California San Diego)
-
352
-
Topics in Randomized Numerical Linear Algebra: I Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
356
Robust hybrid infinite and finite dimensional tensor factorizations
The canonical polyadic (CP) tensor decomposition represents a multidimensional data array as a sum of rank-one outer products of latent factors. Building on CP-HiFi, the hybrid infinite- and finite-dimensional CP framework of Larsen et al. (2024), which introduces quasitensors by modeling selected modes as smooth functional factors in a reproducing kernel Hilbert space, we replace the standard least-squares objective with a quantile-based loss. This modification enables estimation of conditional quantile functions (rather than conditional means) along functional modes. The resulting approach yields a distribution-aware tensor factorization that is robust to outliers and captures heterogeneity beyond the mean structure.
Speaker: Jocelyn Chi (University of Minnesota Twin Cities) -
357
Randomized Numerical Linear Algebra for Tensor-Based Transformers
Attention mechanisms are a central component of transformer models that capture contextual relationships between tokens in large language models. Although many of the underlying computations (e.g., query, key, and value embeddings in multi-head attention) are inherently multi-way, classical transformer models are built on matrix-based formulations.
In this talk, we discuss several ways that tensorial structure can be imposed on and exploited in attention mechanisms of transformer models.
We describe how tensor-based attention can capture higher-order contextual relationships among tokens, vs. pairwise or dot-product attention. We then explore how randomized algorithms in numerical linear algebra may be used to accelerate tensor-based attention computations and reduce storage requirements.Speaker: Katherine Pearce (Oden Institute, UT Austin) -
358
Blocked Leverage Score Sampling in the Randomized Alternating Least Squares CP Tensor Decomposition
The CANDECOMP/PARAFAC (CP) decomposition is a powerful tool used for multiway data analysis and to break the “curse of dimensionality” associated with higher-order tensors. The most common way to compute the CP decomposition of a tensor is via a standard alternating least squares (CP-ALS) algorithm. With the CP-ALS, one must iteratively solve a set of overdetermined least squares problem which can quickly become expensive in time and memory. Recently, Battaglino et al[1] developed a randomized CP-ALS algorithm that significantly reduced the cost of the optimization procedure with little impact on the CP decomposition's accuracy. The strategy leverages the structure of the Khatri-Rao product to efficiently compute a sampling probability distribution for each least squares sub-problem. Unfortunately, on modern computer architecture sampling a tensor is a memory-bound operation, i.e. the computational performance of the randomized CP-ALS method can become bound by the bandwidth and latency of a machine. This work seeks to improve on the computational performance of the randomized, sampled CP-ALS strategy by studying the effects of sampling blocks of columns in each least squares problem. In this talk, we illustrate how block-wise sampling impacts the theoretical performance of the randomized CP-ALS algorithm and provide numerical results comparing the block and standard sampled CP-ALS decomposition.
References:
[1]. Casey Battaglino, Grey Ballard, and Tamara G. Kolda. A practical randomized CP tensor decomposition. SIAM J. Matrix Anal. Appl., 39(2):876–901, 2018.Speaker: Dr Karl Pierce (University of Maryland, College Park) -
359
Accelerating Ptychographic Reconstruction via Stochastic Multilevel Optimization
Ptychography is a powerful coherent diffraction imaging technique essential for reconstructing high-resolution, complex-valued images from intensity-only measurements. However, the reconstruction poses significant challenges due to its nonconvex and ill-posed nature. We propose a novel multilevel optimization framework emphasizing stochastic learning principles to efficiently address these challenges. By formulating the inverse problem as iterative minimization of quadratic surrogate models at varying resolution levels, our method strategically utilizes stochastic gradient evaluations, significantly reducing computational overhead. The multilevel structure exploits hierarchical, multi-scale information intrinsic to the imaging data, enabling faster and more stable convergence compared to traditional deterministic approaches. Hyperparameters are automatically adjusted across resolution levels, ensuring robustness and scalability. Numerical results demonstrate that our stochastic multilevel optimization substantially enhances both reconstruction accuracy and computational efficiency, presenting a robust solution for large-scale ptychographic imaging applications.
Speaker: Zichao Wendy DI (Argonne National Lab)
-
356
-
Symplectic Linear Algebra and Applications: I Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
360
Structure-preserving Krylov Subspace Approximations for the Matrix Exponential of Hamiltonian Matrices
It is well known that the matrix exponential $\exp(H)$ is symplectic whenever
$H \in \mathbb{R}^{2n \times 2n}$ is a Hamiltonian matrix. A matrix $H$ is called
Hamiltonian if it satisfies $HJ = (HJ)^{T},$
while a matrix $S$ is called symplectic (or $J$-orthogonal) if $S^{T} J S = J.$
Here, $J \in \mathbb{R}^{2n \times 2n}$ denotes $J = \left [\begin{smallmatrix} 0 & I_n \\ -I_n & 0\end{smallmatrix}\right].$In this talk, we study structure-preserving Krylov subspace methods for approximating the matrix–vector products $f(H)b,$ where $H$ is a large Hamiltonian matrix and $f$ denotes either the matrix exponential or the related $\varphi$-function ($\varphi(z) = \frac{e^{z} - 1}{z}$). Such computations are central to exponential integrators for Hamiltonian systems.
Assume that a suitable projection $\Pi = V W^{T} \in \mathbb{R}^{2 n \times 2 n}$ is given, where $V, W \in \mathbb{R}^{2n \times 2m}$, $m \leq n$, have full column rank and $W^{T} V = I_{2m}$.
Then $f(H)b$ for $b \in \mathbb{R}^{2n}$ can be approximated by $f(H)b \approx V f(\widetilde{H}) W^{T} b$ with $\widetilde{H} = W^{T} H V \in \mathbb{R}^{2m \times 2m}.$ When $m\ll n$, evaluating $f(\widetilde{H})b$ is far less computationally expensive than evaluating $f(H)b$. Typically, $V=W$ and the columns of $V$ form an orthonormal basis of the standard Krylov subspace $\mathcal{K}_{2m}(H,b)=\operatorname{span}\{b,Hb,\ldots,H^{2m-1}b\}$ of order $2m$, but then $\widetilde{H}$ is in general not a Hamiltonian matrix. This motivates the use of Krylov bases with $J$-orthogonal columns that yield Hamiltonian projected matrices and symplectic reduced exponentials. We compare several such structure-preserving Krylov methods on representative Hamiltonian test problems, focusing on accuracy, efficiency, and structure preservation, and briefly discuss adaptive strategies for selecting the Krylov subspace dimension.Speaker: Heike Faßbender (TU Braunschweig, Institute for Numerical Analysis) -
361
On symplectic reduction of a matrix to upper $J$-Hessenberg form
In the context of computations of eignevalues and eignevectors, structure-preserving of a class of specific structured matrices, the reduction of a matrix to a $J$-Hessenberg condensed form is needed.\
Such reduction is based on symplectic similarity transformations.
It is a crucial step in the $SR$-algorithm (which is a $QR$-like algorithm), structure-preserving, for computing eigenvalues and vectors of such structured matrices.
The algorithm JHESS, or its variant JHMSH are the main algorithms used for such reduction.These algorithms may meet fatal breakdowns, causing brutal stops of the computations or encounter near-breakdowns, which are source of serious numerical instability.
In this talk, we point out the source of these breakdowns and present strategies
for curing them.Speaker: Ahmed Salam (Université du Littoral-Côte d'Opale) -
362
The $\phi-$Reversibility Problem for the Real Symplectic Group
Let $G$ be a matrix group over a field $\mathbb{F}$ and $\phi: M_n(\mathbb{F}) \rightarrow M_n(\mathbb{F})$ be a map such that $\phi(A) \in G$ for all $A \in G.$
An element $A \in G$ is said to be $\phi-$reversible if there exists $P \in G$ such that $PAP^{-1}=\phi(A).$ If $P$ can be chosen to be an involution (i.e., $P^2=I),$ then $A$ is said to be strongly $\phi-$reversible. The strongly $\phi-$reversible elements of the complex symplectic group
$$\operatorname{Sp}(2n, \mathbb{C}):=\left\{A \in GL_{2n}(\mathbb{C}) \, \Bigg\lvert \, A^T\begin{bmatrix} 0 & I_n \\ -I_n & 0 \end{bmatrix}A= \begin{bmatrix} 0 & I_n \\ -I_n & 0 \end{bmatrix}\right\}$$ for the map $\phi: A \mapsto -A$ have been completely classified. In this talk, we consider the real case. We show that several results from the complex setting no longer hold over $\mathbb{R}$. In particular, for each $n$, we construct a real symplectic matrix that is strongly $\phi-$reversible as an element of $\operatorname{Sp}(4n,\mathbb{C})$, but not as an element of $\operatorname{Sp}(4n,\mathbb{R})$. Using a suitable canonical form for real symplectic matrices, we classify the strongly $\phi-$reversible elements of the real symplectic group $\operatorname{Sp}(4,\mathbb{R})$.Speaker: Nicole Joy Datu (University of the Philippines Diliman) -
363
On generalization of Williamson's theorem to real symmetric matrices
Williamson's theorem states that if $A$ is a $2n \times 2n$ real symmetric positive definite matrix then there exists a $2n \times 2n$ real symplectic matrix $M$ such that $M^T A M=D \oplus D$, where $D$ is an $n \times n$ diagonal matrix with positive diagonal entries known as the symplectic eigenvalues of $A$. The theorem is known to be generalized to $2n \times 2n$ real symmetric positive semidefinite matrices whose kernels are symplectic subspaces of $\mathbb{R}^{2n}$, in which case, some of the diagonal entries of $D$ are allowed to be zero.
In this work, we further generalize Williamson's theorem to $2n \times 2n$ real symmetric matrices by allowing the diagonal elements of $D$ to be any real numbers, and thus extending the notion of symplectic eigenvalues to real symmetric matrices. Also, we provide an explicit description of symplectic eigenvalues, construct symplectic matrices achieving Williamson's theorem type decomposition, and establish perturbation bounds on symplectic eigenvalues for a class of $2n \times 2n$ real symmetric matrices denoted by $\operatorname{EigSpSm}(2n)$. The set $\operatorname{EigSpSm}(2n)$ contains the set of $2n \times 2n$ real symmetric positive semidefinite matrices whose kernels are symplectic subspaces of $\mathbb{R}^{2n}$. Our perturbation bounds on symplectic eigenvalues for $\operatorname{EigSpSm}(2n)$ generalize known perturbation bounds on symplectic eigenvalues of positive definite matrices given by Bhatia and Jain [J. Math. Phys. 56, 112201 (2015)].Speaker: Hemant Mishra (Indian Institute of Technology (ISM) Dhanbad)
-
360
-
Model- and Data-driven Reduced-order Models and Their Applications in Inverse Problems: I Goodwin Hall 145
Goodwin Hall 145
Virginia Tech
-
364
Multiparameter Waveform Inversion via Reduced Order Modeling
Waveform inversion seeks to estimate an inaccessible heterogeneous medium from data gathered by sensors that emit probing signals and measure the generated waves. The traditional full waveform inversion (FWI) formulation estimates the unknown coefficients via minimization of the nonlinear, least squares data fitting objective function. For typical band-limited and high frequency data, this objective function has spurious local minima near and far from the true coefficients. Thus, FWI implemented with gradient based optimization algorithms may fail, even for good initial guesses, a phenomenon known as cycle skipping. Recently, it was shown that data driven reduced order models (ROMs) can be used to obtain a better behaved objective function for wave speed estimation. We introduce ROMs for waves obeying a first order hyperbolic system. They are defined via Galerkin projection on the space spanned by the wave snapshots, evaluated on a uniform time grid with an appropriately chosen time step. The proposed ROMs are data driven, as they are computed directly from the sensor measurements. The ROM computation applies to any linear waves in lossless and non-dispersive media. We present an example of acoustic waveform inversion in a medium with both wave speed and density unknown. Numerical examples show that both quantities can be estimated efficiently via minimization of an objective function that employs ROM based approximation of the wave field inside the unknown medium.
Speaker: Alexander Mamonov (University of Houston) -
365
Characterization of the response of electric circuits with two kinds of passive elements
The response matrix of an electrical circuits composed of two kinds of passive elements (e.g. LC, RC, CL...) is a matrix-valued rational function of the frequency that associates time-harmonic voltages at some terminal nodes to time-harmonic currents. We present necessary conditions that must be satisfied by the response matrix. These conditions can be used to construct circuits whose response matches prescribed poles, with some exceptions. Progress towards a full characterization will be presented.
Speaker: Fernando Guevara Vasquez (University of Utah) -
366
Adaptive data-driven reduced-order models of port-Hamiltonian dynamical systems for nonlinear inverse scattering applications
The inverse scattering problem formulated for the Schrödinger operators arises in various fields, including quantum mechanics, radars, viscoelasticity, Biot problems, remote sensing, geophysical, and medical imaging. The goal of imaging is to find medium properties in the domain using near-field measured data. The model based nonlinear optimization which is the method of choice for the solution of the inverse problems can be unreliable and particularly expensive for such problems. Data driven nonlinear transforms can be an opening, however it was recently shown that the ReLU networks are intractable for reliable solution of the inverse problems in continuum using conventional digital computers. In the present work, following the success of data-driven reduced-order models (ROMs) developed in recent years, we propose a robust direct method for solving inverse scattering problems for the Schrödinger equation. Our approach is based on a Lippmann-Schwinger algorithm with a crucial component composed of adaptive data-driven ROMs in the frequency domain and efficient learning the internal solutions. In the talk we will discuss the details of the algorithms as well as some bottlenecks.
Speaker: Mikhail Zaslavskiy (Southern Methodist University)
-
364
-
New Advancements in Tensor Decomposition and Computation: I Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
367
Transform-Based Multilinear Algebra via Tensor Decompositions
Transform-based tensor products, including the T-product and its more general form, namely the higher-order tensor-tensor product, have been widely used in image processing, signal reconstruction, and robotics. While invertible transforms enable tensor computations to be carried out via matrix operations in the transform domain, the resulting storage and computational costs remain prohibitive for high-dimensional tensors. To address this challenge, we leverage tensor decomposition techniques, including tensor train decomposition (TTD) and hierarchical Tucker decomposition (HTD), to accelerate transform-based multilinear algebra. In particular, we develop TTD- and HTD-based formulations for the T-product and its associated algebra by operating directly on the factor matrices or tensors of the decompositions. The framework is further generalized to the higher-order tensor-tensor product and applied to multilinear control problems. We demonstrate the effectiveness of our framework with numerical examples.
Speaker: Ms Yidan Mei (Yale University) -
368
Iterative methods for tensor systems under the t-product
Solving linear systems is a crucial subroutine and challenge in data science and scientific computing. Classical approaches to solving linear systems assume that data is readily available and sufficiently small to be stored in memory. However, in the large-scale data setting, data may be so large that only partitions (e.g., single rows/columns of the matrix/tensor) can be utilized at a time. In this presentation, we discuss our recent progress designing and analyzing tensor-based iterative methods, which in settings that are heavily memory-constrained, when: (i) only row/column slices can be used, (ii) only frontal slices can be used, and (iii) only randomly accessed elements can be used in each iteration.
Speaker: Anna Ma (University of California, Irvine) -
369
Tensor Methods for Multi-omics Data
High-throughput genomics and omics technologies generate data with intrinsic multi-way structure arising from multiple samples, molecular features, experimental conditions, and biological contexts. Standard matrix-based methods often obscure this structure through flattening or aggregation. Tensor-based representations provide a natural mathematical framework for preserving and exploiting the underlying multilinear organization of such data.
In this talk, we present tensor decomposition models for the analysis of high-dimensional omics datasets and demonstrate how these models yield low-rank representations that enable dimensionality reduction while maintaining interpretability through mode-specific latent factors. Applications include multi-sample gene expression analysis, multi-omics integration, and higher-order relational modeling of cell–cell interactions using multi-omics data.
We discuss key mathematical and computational challenges arising in these applications, including rank determination, identifiability and uniqueness, robustness to noise, scalability, and the incorporation of structural constraints motivated by domain knowledge. We conclude by outlining open problems at the interface of multilinear algebra, optimization, and statistical modeling, highlighting opportunities for theoretical advances driven by modern omics data.
Speaker: Dr Neriman Tokcan (University of Massachusetts Boston)
-
367
-
Combinatorial Matrix Theory: I Goodwin Hall 135
Goodwin Hall 135
Virginia Tech
-
370
Spectral bounds for the clique number of a graph
Given an integer k, deciding whether a graph has a clique of size k is an NP-complete problem. Wilf's inequality provides a spectral lower bound for the clique number (i.e., the order of a largest clique) in terms of the largest adjacency eigenvalue. In 2018, Elphick and Wocjan conjectured a stronger spectral bound using positive adjacency eigenvalues. We introduce a spectral bound using Laplacian eigenvalues. We show that this is an improvement for almost all graphs.
Speaker: Sudipta Mallik (Marshall University) -
371
Hermitian Spectra of Merged Subdivision Mixed Graphs
Let $G$ be a mixed graph and let $(H_1, H_2)$ be an ordered pair of mixed graphs whose orders coincide with the order and size of $G$, respectively. We introduce the subdivision mixed graph $S(G)$ and the $(H_1,H_2)$-merged subdivision mixed graph. We investigate the Hermitian spectrum and the Hermitian energy of these graphs, deriving spectral properties that relate merged subdivision mixed graphs to their underlying mixed and undirected counterparts. In particular, we compare the Hermitian energies of certain classes of mixed graphs with those of their underlying graphs, illustrating how the merging process influences spectral behavior. These results show how graph operations influence Hermitian spectral invariants in mixed graphs.
This is joint work with G. Infante, M. Robbiano, Universidad Católica del Norte, Chile,and Eber Lenes, Universidad del Sinú, ColombiaSpeaker: Enide Andrade (Center for Research and Development in Mathematics and Applications (CIDMA), Department of Mathematics, University of Aveiro, Portugal) -
372
How to Combinatorially Prove Matrices are Mutually Inverse
Let $A = (A_n)_{n\geq 0}$ and $B = (B_n)_{n\geq 0}$ be families of "recursive'' rectangular matrices, that is, the entries of $A_n$ can be expressed as a linear combination of entries of $A_m$ for $m < n$, and similarly for $B$. Such families arise in algebraic combinatorics as change-of-basis matrices between the basis of symmetric functions and their generalizations. The entries of such matrices can be computed as sums over signed-weighted combinatorial objects which themselves can be constructed recursively. In this talk, I will describe a framework that reduces the condition $A_nB_n = I$ for all $n\geq 0$ on the matrix families to an equivalent local condition on the objects which can be handled using combinatorial methods such as sign-reversing involutions. This framework has applications in producing canonical bijective proofs for the Kostka matrix case, the orthogonality of symmetric group characters, and Mobius inversion for the lattice of sets. All combinatorial concepts will be defined in the talk.
Speaker: Aditya Khanna (Virginia Tech)
-
370
-
3:40 PM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
373
ILAS Business Meeting McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
293
-
-
374
Announcements McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
Contributed Talks: J Torgersen Hall 1040
Torgersen Hall 1040
Virginia Tech
-
375
A More Robust Streaming Dynamic Mode Decomposition
Dynamic Mode Decomposition (DMD) is a data-driven tool for capturing complex nonlinear dynamics. It can be used to identify, analyze and forecast dynamical systems $x_{k+1}=F(x_k)$ governed by an unknown or complex mapping $F$ using only observed snapshots $s_1,s_2,...,s_{n+1}$. If we denote $X := (s_1 \ s_2 \ \cdots \ s_n)$, $Y := (s_2 \ s_3 \ \cdots \ s_{n+1})$, finite-dimensional approximation of the governing function $F$ is a matrix $A$ such that $ A X \approx Y$ in the least-squares sense.
In a streaming application DMD has to be recomputed over a widening data window. If we denote a rank-revealing orthonormal decomposition of $X = Q_x\tilde X$ and $Y = Q_y \tilde Y$, Hemati et al. (2014) propose updating Rayleigh-Ritz matrix $$ Q_x^TAQ_x = (Q_x^TQ_y)(\tilde Y \tilde X^T)(\tilde X \tilde X^T)^{-1} =: (Q_x^TQ_y) G_{y,x} G_x^{-1} $$ via updates of $Q_x, Q_y, G_{y,x}$ and $G_x$. Smaller size and special structure of $G_{y,x}$, $G_x$ make them easy to update. However, since $$\tilde Y \tilde X^T (\tilde X \tilde X^T)^{-1} = \tilde Y \tilde X^\dagger$$ this formulation increases the condition number $\kappa(\tilde X\tilde X^T) = \kappa(\tilde X)^2$ making further computation potentially unstable. Additionally, since $X$ and $Y$ only differ up to one column, maintaining and updating two separate orthogonal matrices seems redundant. Our approach is based on low-rank orthogonal decomposition of $S = Q\tilde S$ where $S = (s_1 \ s_2 \ \cdots \ s_{n+1} )$. This way we only have to store and update one orthogonal basis $Q$. We propose an alternative formulation $$\tilde Y \tilde X^{-1} = (\tilde Y\tilde Q)T^{-1} =: G_{y,q}T^{-1}$$ using LQ or RQ decomposition of $\tilde X = T\tilde Q^T$ that avoids the squaring of the condition number while maintaining the same matrix sizes and simple updates. Our numerical analysis and numerical experiments demonstrate better numerical properties that result in better forecasting skill.
Speaker: Ela Đimoti (University of Zagreb) -
376
Fast and accurate solution of the periodic differential matrix Riccati equation
We want to compute a $T$-periodic symmetric solution $X(t)$ of a $T$-periodic differential matrix Riccati equation
$-\dot{X}(t)=X(t)A(t)+A^T(t)X(t)-X(t)B(t)R^{-1}(t)B^T(t)X(t)+Q(t)$
such that all solutions of the feedback system $\dot{x} = [A(t)-B(t)R^{-1}(t)B^T(t)X(t)]x(t)$ are asymptotically stable, i.e. $\lim_{t\to\infty}x(t)=0$. The standard solvability condition for the matrix Riccati equation requires that the $T$-periodic Hamiltonian matrix
$H(t) = \begin{bmatrix}A(t) & -B(t)R^{-1}(t)B^T(t) \\ -Q(t) & - A^T(t)\end{bmatrix}$
has no characteristic multipliers on the unit circle $|z|=1$.
The transition matrices $\Phi_k=\Phi_H(t_{k+1},t_k)$ for the periodic system with the matrix $H(t)$ are calculated using the Runge-Kutta method for $k=0,1,\ldots,P-1$ on an equidistant grid $t_k=kh$ with step $h=T/P$. The monodromy matrices for $H(t)$ sampled on the grid $t_k$ are matrix products $\Psi_k=\Phi_{k+P-1}\ldots\Phi_{k+1}\Phi_k$, where the indices are taken modulo $P$. The sampled solution $X(t_k)$ is determined by the block eigenvalue problem
$\Psi_k\begin{bmatrix}I\\X(t_k)\end{bmatrix}= \begin{bmatrix}I\\X(t_k)\end{bmatrix}\Lambda_k$,
where the discrete-stable matrices $\Lambda_k$ are the monodromy matrices of the aforementioned feedback system.In paper [1], an orthogonal elimination method was proposed for accurate computation of $\Psi_k$. In paper [2], an $O(P\log P)$ algorithm was developed for computing all monodromy matrices $\Psi_k$. In this talk, I present a faster $O(P)$ method for computing the entire solution $X(t_k)$, $k=0,1,\ldots,P-1$. Alternative algorithms are discussed in paper [3].
References
[1] A.N. Malyshev, Siberian Math J., 30(4):559-567, 1989.
[2] P. Benner et al, Parallel Algorithms for LQ Optimal Control of Discrete-Time Periodic Linear Systems, J. Parallel Distrib. Comp, 62:306-325, 2002.
[3] S. Gusev et al, A numerical evaluation of solvers for the periodic Riccati differential equation, BIT Numer Math, 50:301-329, 2010.Speaker: Aleksandr Malyshev -
377
Spectral clustering and accelerated SSO for large and dense problem
Spectral clustering is a graph-based method for data partitioning, which relates a relaxed graph partitioning problem to the eigenvectors of the associated graph Laplacian; see, e.g., [Peng et al., SIAM J. Comput. 46(2):710–743, 2017]. Following the computation of these eigenvectors, a post-processing step is used to determine the final clustering. Typically, k-means is the standard choice of post-processing algorithm. Recently [Eldén, SIAM J. Matrix Anal. Appl. 45(1):112–133, 2024], an algorithm called Semi-sparse Orthogonal (SSO) Approximation, has been proposed as an alternative that improves on both clustering accuracy and stability. Given the eigenvector matrix $X \in R^{n\times p}$, SSO seeks $\min_{Q,\Psi}\ \|XQ-\Psi\|_F,$ where $Q\in R^{p\times p}$ is orthogonal and $\Psi\in R^{n\times p}$ is an indicator matrix with one nonzero per row. The method alternates between updating $\Psi$ by assigning each row to the entry of the largest magnitude of $XQ$, and updating $Q$ by the polar factor of $\Psi^T X$.
In this work, we consider clustering problems on large and dense graphs, with a large number of clusters. In such settings, standard Arnoldi type solvers become computationally unattractive. Furthermore, computing the polar factor based on singular value decomposition becomes infeasible at these scales. We present an ongoing investigation that circumvents these difficulties by replacing explicit eigenvector computations with the spectral projector of the graph Laplacian, and by performing post-processing via a large-scale variant of SSO. Our approach exploits methods from matrix functions to compute the spectral projectors and the polar factors using polynomial approximations. This avoids the explicit computation of individual eigenvectors and is well suited for large-scale, dense problems.
Speaker: Jiayu Bian (KTH Royal Institute of Technology)
-
375
-
Inverse Problems and Uncertainty Quantification through the Lens of Numerical Linear Algebra: J Torgersen Hall 1030
Torgersen Hall 1030
Virginia Tech
-
378
Multigrid-Accelerated Sparsity-Promoting Projection Methods for Inverse Problems
Hybrid projection methods are an effective iterative approach for the solution of large-scale linear inverse problems, including those promoting sparsity in the recovered solution. Priorconditioned (prior-preconditioned) hybrid methods have been proposed to improve performance, but introduce additional computational costs in each iteration related to the application of a weighted pseudoinverse to a matrix of basis vectors. We propose to use multigrid-preconditioned block Krylov methods to accelerate the application of the weighted pseudoinverses in each iteration, which are suitable for both distributed CPU and GPU computing environments. Numerical results are presented for large-scale static and dynamic inverse problems.
Speaker: Jonathan Lindbloom (Dartmouth College) -
379
Time-varying Bayesian Inverse Problems with Sparse Priors and Randomization
Sparse priors, such as the Laplace prior, are of considerable interest in Bayesian inverse problems because they promote sparsity and preserve edges in the solution, which are often more appropriate than the smooth reconstructions obtained with Gaussian priors. However, sampling from the resulting non-Gaussian posteriors is challenging, particularly in high-dimensional settings. To address this, we build on the randomize-then-optimize (RTO) framework and its extension to $\ell_1$-type priors via variable transformation. We propose a method for posterior sampling in Bayesian inverse problems with Laplace priors that converts the prior into a standard Gaussian in the transformed space. Within the RTO framework, this results in a nonlinear optimization problem at each iteration, which we solve using the Levenberg–Marquardt algorithm. To accelerate the linear system solves at each Levenberg–Marquardt iteration, we employ randomized preconditioners and investigate the effect of using single and half precision arithmetic in these solves. Performance is evaluated through numerical experiments on image deconvolution and computed tomography problems.
Speaker: Diego Arenas Mata (Virginia Tech) -
380
A Provably Convergent MM-GKS Variant for Large-Scale Image Reconstruction Problems
Reconstructing high-quality images with sharp edges requires edge-preserving regularization, often imposed using the $\ell_1$-norm of the gradient. To get a computationally tractable problem, the $\ell_1$-norm term is typically replaced with a sequence of $\ell_2$-norm weighted gradient terms with the weights determined from the current solution estimate. The majorization-minimization generalized Krylov subspace method (MM-GKS) has the advantage of combining the updating of the regularization operator with generalized Krylov subspaces (GKS). Unfortunately, the storage and the cost of repeated orthogonalization can present overwhelming memory requirements and computational costs.
We present a variant of MM-GKS that provably converges to the minimum of the smoothed functional even if the solution search space dimension remains very small. This substantially improves theoretical results for MM-GKS where the convergence proof relies on (eventually) spanning the full problem space. Using this result, we develop a new method that solves the minimization/imaging problem by alternatingly compressing and expanding the search space while maintaining strict monotonic convergence. Our method can solve large-scale problems efficiently both in terms of computational complexity and memory requirements. We further generalize our proposed method to handle streaming problems where the data is either not all available simultaneously or the size of the problem demands it be treated as such. We demonstrate the utility of our approach on several image reconstruction and restoration problems. This is joint work with Mirjeta Pasha (VT) and Eric de Sturler (VT).
Speaker: Misha Kilmer (Tufts University)
-
378
-
Code-based Cryptography: J Goodwin Hall 125
Goodwin Hall 125
Virginia Tech
-
381
Code-Based Arithmetic Circuits
Arithmetic circuits provide a versatile framework for demonstrating generic algebraic statements, expressible as a system of polynomials, in a zero-knowledge manner. While this primitive can be used to prove knowledge of solutions to NP-complete problems (graph 3-coloring, Sudoku, etc), existing implementations generally rely on discrete logarithm problem assumptions. In this talk, we introduce a novel code-based arithmetic circuit framework. Our construction permits a prover to demonstrate that committed Hamming-ball vectors satisfy certain arithmetic relationships, solely by acting on their syndromes. By translating generic circuit satisfiability to this code-based setting, our framework provides a critical stepping stone for the development of a secure code-based cryptocurrency.
Speaker: Freeman Slaughter (University of South Florida) -
382
Code distances: a new family of invariants of linear codes
We introduce code distances, a new family of invariants for linear codes. We establish some properties and prove bounds on the code distances, and show that they are not invariants of the matroid (for a linear block code) or q-polymatroid (for a rank-metric code) associated to the code. By means of examples, we show that the code distances allow us to distinguish some inequivalent MDS or MRD codes with the same parameters. We also show that no duality holds, i.e., the sequence of code distances of a code does not determine the sequence of code distances of its dual.
Speaker: Hiram López (Virginia Tech) -
383
An Elementary Analysis of Multivariate Goppa Codes
Goppa codes form a structured family of linear error-correcting codes introduced by Valery D. Goppa in 1970 and later interpreted within the framework of algebraic geometry as codes arising from algebraic curves over finite fields. Binary Goppa codes with irreducible Goppa polynomials are used in the Classic McEliece post-quantum key encapsulation mechanism (PQ-KEM), where their efficient decoding algorithms and resistance to structural attacks underpin a long-standing and well-studied code-based public-key encryption scheme.
Motivated by classical Goppa codes, multivariate Goppa codes are an analogous family of codes constructed using multivariate polynomials, originally introduced by Hiram H. López and Gretchen L. Matthews in 2021. In this talk, we generalize the original construction and show that these generalized multivariate Goppa codes exhibit similar rank and distance bounds, as well as comparable structural properties. We conclude by discussing their connection to classical Goppa codes and their suitability for cryptographic applications.
Speaker: William Mahaney (Virginia Tech)
-
381
-
Linear Algebra Foundations for Data-driven Modeling and Model Order Reduction: J Torgersen Hall 1020
Torgersen Hall 1020
Virginia Tech
-
384
Loewner linearizations of structured rational matrices
We show how to use tangential interpolation techniques to construct structured linearizations for several types of structured rational matrices. The classes studied in this paper are square rational matrices that are either Hermitian, or skew-Hermitian, or complex symmetric, or complex skew-symmetric, upon evaluation on one of the following three curves~: the real axis, the imaginary axis and the unit circle. The proposed linearizations are system matrices for these rational matrices and they preserve the structure of the rational matrices, except for the case of the unit circle. For that case, the rational matrix $R(z)$ is linearized using a palindromic or anti-palindromic system matrix for a modified rational matrix, whose eigenvalues that are not on the unit circle preserve the symmetries of the zeros and poles of $R(z)$. The basic tool used to obtain the results in this paper is tangential interpolation via the Loewner and shifted Loewner matrices. In the case of preserving symmetries with respect to the unit circle, we combine this with Mobius transforms.
Speaker: Paul Van Dooren (Universite catholique de Louvain) -
385
Tackling the curse of dimensionality through the parametric Loewner framework: recent advances and applications
We extend the Loewner framework to multivariate (static and dynamic) functions with an arbitrary number $n$ of variables [1]. We present the following facts:
(i) That $n$-variable rational functions (and realization), described in the barycentric basis, can be constructed to interpolate and/or approximate/compress any tensorized $n$-D data or $n$-variate function;
(ii) That these $n$-variable rational functions can be obtained thanks to a sequence of small-scale single-variable interpolation (performed with Loewner matrices), leading to drastically taming the curse of dimensionality (both in memory and computational effort);
(iii) That such sequence results in the variables decoupling, providing a numerically robust solution to the Kolmogorov Superposition Theorem (KST), restricted to rational functions;
(iv) That the Loewner framework bridges "Approximation theory" (both functions and tensors) with "Systems theory", and provides connections with Kolmogorov Arnold Networks (KAN).A collection of numerical examples and method comparison illustrates the effectiveness and scalability of the proposed method and its ability to tame the curse of dimensionality [2]. Issues and outlook will be specifically addressed. Some research codes are also presented [3].
[1] https://doi.org/10.1137/24M1656657
[2] https://arxiv.org/abs/2506.04791
[3] https://github.com/cpoussot/mLFThis works is carried out in collaboration with A.C. Antoulas, I-V. Gosea, C. Poussot-Vassal and P. Vuillemin.
Speaker: Ion Victor Gosea (Max Planck Institute for Dynamics of Complex Technical Systems, Magdeburg, Germany) -
386
From Structured Loewner Matrices to Balanced Mechanical Systems
Learning dynamical systems from data has emerged as a pivotal area of research, bridging the realms of mathematics, engineering, and data science. Of particular importance is the construction of models with meaningful internal structure that allows interpretability and explainability of the results. In the case of mechanical and electro-mechanical processes, dynamical systems are typically described via ordinary differential equations with second-order time derivatives
$$ M \ddot{x}(t) + D \dot{x}(t) + K x(t) = B u(t), \quad y(t) = C_{\mathrm{p}} x(t) + C_{\mathrm{v}} \dot{x}(t), $$ with the system matrices $M, D, K \in \mathbb{R}^{n \times n}$, $B \in \mathbb{R}^{n \times m}$, $C_{\mathrm{p}}, C_{\mathrm{v}} \in \mathbb{R}^{p \times n}$, the inputs $u(t) \in \mathbb{R}^{m}$, the internal states $x(t) \in \mathbb{R}^{n}$, and the outputs $y(t) \in \mathbb{R}^{p}$. A popular approach in model order reduction to construct accurate low-dimensional surrogate models that retain the second-order internal structure is the second-order position-velocity balanced truncation method. Thereby, the system is balanced with respect to structured system Gramians and states corresponding to small singular values of these Gramians are truncated. In this work, we present a new data-driven formulation of the second-order balanced truncation method. Utilizing suitable structured Loewner matrices, we can directly compute low-dimensional balanced second-order systems with generalized proportional damping, $D(s) = f(s) M + g(s) K$ where $f(s), g(s) \in \mathbb{C}$, from given frequency domain measurements. While based on classical structure-preserving model reduction, our method can be used to enforce mechanical model structure independent of the data source. Several numerical examples demonstrate the effectiveness of the proposed method.Speaker: Steffen W. R. Werner (Virginia Tech)
-
384
-
Hierarchical Low-Rank Approximations: Algorithms and Applications: J McBryde Hall 113
McBryde Hall 113
Virginia Tech
-
387
An adaptive method for constructing hierarchical approximations and its application to inverse problems
Dense matrices arise in many areas of computational science, and hierarchical approximation methods have been developed to reduce their storage and computational costs. However, many existing approaches require access to individual matrix entries, which may not be available or prohibitively expensive to compute in important applications. A prominent example is the Hessian in Bayesian inverse problems: accurate Hessian approximations can enable fast posterior sampling, yet forming Hessian entries is costly, while Hessian–vector products can often be computed efficiently.
In this talk, we present adaptive randomized algorithms for constructing hierarchical approximations to a prescribed accuracy in a matrix-free setting. In particular, we extend the peeling paradigm to an adaptive compression strategy for HODLR matrices by approximating off-diagonal blocks at each hierarchical level to meet a global tolerance. Numerical experiments show that the proposed method achieves high accuracy more efficiently than fixed-rank schemes and than global low-rank approximations of the Hessian.
Speaker: Chao Chen (NC State) -
388
Circulant Preconditioning fractional PDEs on Adaptive Meshes
A preconditioner for solving fractional partial differential equations (PDEs) is presented. In our method the fractional PDE is discretized on an adaptive grid, resulting in a Hierarchical matrix representation. The stiffness matrix has Toeplitz blocks along the diagonal and low-rank approximations off the diagonal. Our preconditioner expands on previously developed methods of conditioning Toeplitz systems with circulant matrices. We show how these methods can be applied cheaply on the adaptive mesh and prove that the spectrum of the resulting system is well-clustered. In order to prove these results, we must take special consideration of how the low-rank blocks perturb the eigenvalues of the Toeplitz block-diagonal system. We validate our results for various fractional orders and inspect any assumptions through numerical tests.
Speaker: Kate Wall (Tufts University) -
389
Adaptive coarse space for multi-level overlapping Schwarz preconditioners in FROSch
We discuss the adaptive coarse-space basis functions for the multi-level overlapping additive Schwarz preconditioners implemented in FROSch. The basis functions are formed based on the discrete Harmonic extensions of the local subdomain interface functions. The basis functions for the interface are composed of the eigenvectors corresponding to the small eigenvalues of the generalized eigenvalue problems formed by the local Schur complements on the interface. This leads to a provably-robust multi-level preconditioner for solving heterogeneous elliptic problems. We present numerical results to demonstrate that the iterations count stays relatively constant even with the increasing number of subdomains for solving problems with large coefficients jumps. Our performance results on NERSC Perlmutter supercomputer will then demonstrate the scalability of the method for solving large-scale problems.
Speaker: Ichitaro Yamazaki (Sandia Labs)
-
387
-
Topics in Randomized Numerical Linear Algebra: J Torgersen Hall 3100
Torgersen Hall 3100
Virginia Tech
-
390
Accelerating Asynchronous Iterative Methods with Residual-Biased Randomization
Asynchronous iterative methods, such as Asynchronous Jacobi, offer a promising mechanism for overcoming synchronization bottlenecks in massively parallel and heterogeneous computing environments. By allowing processing elements to update components using the latest available data without global barriers, these methods maximize computational throughput. However, asynchrony also makes performance more sensitive to the order in which unknowns (or blocks of unknowns) are updated, since processors may act on delayed information and progress unevenly. While randomized numerical linear algebra has provided bounds for asynchronous solvers using uniform sampling, this approach ignores the local error distribution, potentially wasting updates on converged components.
We consider a family of sampling strategies that range from uniform randomization to non-uniform importance sampling, and then to dynamic, time-varying distributions that adapt online using inexpensive residual-based indicators. The main theoretical focus is on convergence in expectation under randomized update rules. To make the analysis transparent, we begin with a synchronous randomized model and establish conditions under which the associated sampling schemes yield expected decrease in a suitable error measure. We then carry the same randomization patterns to an asynchronous setting, showing how mild assumptions on delays and update coverage allow the expected contraction arguments to be extended, while making explicit where asynchrony changes the requirements and the bounds.
Speaker: Evan Coleman (University of Mary Washington) -
391
Block Subset Selection based on Randomized QR with Column Pivoting
The column subset selection problem seeks to find a collection of the matrix columns that have similar spectral properties to the original matrix. Recently with the large amount of data available, many have turned to using randomization to reduce the problem's computation. While there have been many methods that motivate how to select these columns, they are just that--individual columns. However, by blocking these columns, there is less computational communication needed, which makes the process faster. In this talk we will discuss optimality conditions for selecting these block of columns using randomization. We relate the worst case of this randomized method to the deterministic block QR with column pivoting (QRCP). We then corroborate this analysis with numerical experiments to showcase the method's speed and accuracy compared to standard QRCP algorithms.
Speaker: Mitchell Scott (Emory University) -
392
Singular values and vectors of sparse random rectangular matrices at criticality
In modern machine learning applications, data matrices are always assumed to admit the signal-plus-noise structure. Typically, we assume that the spectra of signal and noise matrices are well-separated and that noise subspaces only produce a marginal influence. While these assumptions are readily verified for dense matrices via classical random matrix theory, real-world data is often sparse, posing significant theoretical challenges. We investigate the spectral properties of a matrix $X \in \mathbb{R}^{n \times m}$ with i.i.d. $\text{Bernoulli}(p)$ entries. Previous literature proved that when $np \gg \log(n)$, the singular values of $X$ almost surely remain within the compact support of the Marčenko-Pastur (MP) distribution. However, we identify a critical sparsity regime $p = b \log(n) / \sqrt{mn}$ where this classical result fails. We provide a quantitative characterization of the emergence of outlier singular values. For explicit thresholds $b_*$ and $b^*$ as functions of the aspect ratio $\gamma = n/m \ge 1$, we prove a three-phase transition: (1) for $b > b_*$, no outliers exist; (2) for $b^* < b < b_*$, outliers emerge only beyond the right edge of the MP law; and (3) for $b < b^*$, outliers appear on both sides of the bulk, all with high probability. The locations of those outliers are precisely determined by the largest and smallest degree vertices of the underlying random graph. Besides, behavior of singular vectors corresponding to bulk and edge singular values can be characterized precisely. Our approach follows the framework established by Alt, Ducatez, and Knowles (2021), which can be extended to sparse matrices with general bounded entries.
Speaker: Haixiao Wang (University of Wisconsin-Madison)
-
390
-
Spectral Interlacing, Graph Learning, and Quantum Perspectives on Signed Graphs: J Goodwin Hall 115
Goodwin Hall 115
Virginia Tech
-
393
Quantum-Classical Algorithms for Counting Triangles in a Signed Edge Stream
We discuss a quantum-classical streaming algorithm that processes signed edges to efficiently estimate the counts of triangles of diverse signed configurations in the edge stream. The approach introduces a quantum sketch register for processing the signed-edge-stream, together with measurement operators for query-pair calls in the quantum estimator, while a complementary classical estimator accounts for triangles not captured by the quantum procedure. This hybrid design yields a polynomial space advantage over purely classical approaches, extending known results from unsigned edge-stream data to the signed setting.
Speaker: Bibhas Adhikari (Fujitsu Research of America, Inc.) -
394
State transfer in discrete quantum walks: from coins to weighted graphs
A discrete quantum walk takes place on the arcs of a graph, and evolves according to a coin operator and a shift operator. One important task, given the underlying graph, is to construct quantum walks that start with a state localized at a vertex and get arbitrarily close to a state localized at another vertex. In this talk, I will show how different coin operators translate into different weightings of the underlying graph, and use the spectral properties of these weighted graphs to construct quantum walks that admit the aforementioned transport phenomenon.
Speaker: Hanmeng Zhan (Worcester Polytechnic Institute)
-
393
-
Polynomials, Krylov Methods and Applications: J McBryde Hall 129
McBryde Hall 129
Virginia Tech
-
395
Tensor Network Krylov Methods: Algorithms, Theory Gaps, and Open Problems
In recent years, tensor network methods have garnered increased attention for modeling high dimensional quantum many body systems and for representing high dimensional functions with structured correlations.
A basic unresolved question is whether standard numerical linear algebra primitives, such as Krylov based iterative solvers, can be realized in matrix product state and matrix product operator form while remaining computationally efficient.
This talk gives a pedagogical overview of the current state of tensor network Krylov methods. I will define what it means to run a Krylov iteration when vectors and operators are represented in the tensor network format, and explain why familiar steps like basis construction and reorthogonalization become nontrivial algorithmic design problems in this setting. I will then show how tensor network structure can be paired with randomized low rank approximation to speed up the tensor network analog of matrix vector products, opening a broader design space for fast high dimensional Krylov solvers. Recent approaches will be surveyed, including tensor network Lanczos variants and sketched tensor network GMRES methods.
I will close with open questions around stability, error control, and tensor network bond dimension growth that are currently shaping the development of new algorithms in this space.
Speaker: Chris Camaño (Caltech) -
396
Randomized Row Norm Estimation: Algorithms and Applications
Computing the diagonal entries of a large linear operator is a common computational primitive in numerical linear algebra, with applications in uncertainty quantification, cross-validation, perturbation analysis, electronic structure calculation and more. However, estimating the diagonals of a matrix given only implicit matrix-vector access is challenging, as randomized algorithms suffer from high variance. In many cases, these problems can be efficiently reduced to estimating the squared row norms of a suitable square root of the matrix, where randomized algorithms specialized to row norm estimation with considerably lower variance may be applied. In particular, when the matrix in question is a function of another matrix, that is, $A = f(B)$, and matrix vector products with $A$ are computed using a Krylov method, it is often the case that matrix vector products with $f^{1/2}(B)$ can be computed just as efficiently those with $f(B)$. We present algorithms for estimating squared row norms and approaches for incorporating them into other computational tasks.
Speaker: Alexander Hsu (University of Washington) -
397
The Matrix-Vector Complexity of Ax=b
Solving linear systems Ax=b is a fundamental pillar of NLA. For over 60 years, iterative methods that access A only through matrix-vector products have been the standard approach for solving large linear systems. While lower bounds exist for many special cases, prior work has not shown that methods like GMRES and MINRES achieve an asymptotically optimal matrix-vector complexity for approximately solving Ax=b. In this talk, we prove that MINRES (and CG on the normal equations) achieves an optimal complexity (up to a factor of 4) over all randomized matrix-vector methods, and GMRES achieves the best possible over all "transpose-free" methods.
Speaker: Raphael Meyer (UC Berkeley)
-
395
-
New Advancements in Tensor Decomposition and Computation: J Torgersen Hall 1060
Torgersen Hall 1060
Virginia Tech
-
398
Tensor Data for Control Strategies in Systems
We study optimal control problems arising from partial differential equations. More specifically, we look at the optimal control of Allen-Cahn Equation (ACE) with a source term. ACE is well known for modeling phase transitions and thus, has many applications, apart from physics (semiconductors), in biological systems, material science and image processing. ACE models cellular membranes which are vital in transporting vesicles, viral budding and dynamical sorting and signaling.
Through the dynamic programming principle, optimal control problem are reformulated into the Hamilton-Jacobi-Bellman (HJB) Equations. It is well known that the computational cost is prohibitively expensive when standard methods are implemented in solving the HJB Equations. Thus we obtain reduced order models, using tensor decomposition techniques for reduced order basis. Sparse optimization and flexible hybrid methods are used for low rank canonical polyadic tensor decomposition. The reduced optimal control problem leads to reduced state-dependent Riccati Equations which can be solved efficiently.
Speaker: Dr Carmeliza Navasca (University of Alabama at Birmingham) -
399
Preconditioners for Kronecker Sum Systems with Applications to Polynomial Feedback Control
Multivariate polynomial approximations to Hamilton-Jacobi-Bellman equations can be expressed using Kronecker products leading to very large, but structured, linear systems. Their structure appears as n-way generalizations of Lyapunov or generalized Lyapunov equations. For monomial terms of degree d, their dimension scales as the number of state dimensions n raised to the d. For feedback control problems with modest dimension n=1000 and relatively low degree approximations of d=3 or 4, matrix-free iterative solvers using scalable preconditioners are essential. This talk will review several options for preconditioners, then present a number of examples arising from discretizations of nonlinear partial differential equation control problems that demonstrate the performance of these linear solvers. As expected, there can be a strong connection between the discretization and the preconditioner choices. We conclude with applications to fluid flow control problems modeled through approximations to Navier-Stokes equations.
Speaker: Jeff Borggaard (Virginia Tech) -
400
A greedy approach for approximate tensor diagonalization
Symmetric tensor diagonalization has applications in statistics and signal processing. Unlike for real symmetric matrices, there is no guarantee that a real-valued symmetric tensor is diagonalizable. Therefore, one generally approaches the problem as an approximate tensor diagonalization (ATD) problem. In this talk, we show that Jacobi-type methods for ATD that naturally extend the Jacobi method for real symmetric matrix diagonalization share several useful properties with the original method, including the conservation of mass during the iteration steps. We use these properties to generalize the greedy method for the matrix case to third-order real-valued tensors, and present numerical results comparing our approach to the classical sweep method. We discuss convergence properties and place our work in the context of Jacobi-type approaches for ATD.
Speaker: Anna Konstorum
-
398
-
10:00 AM
Coffee Break Goodwin Hall Atrium & Torgersen Hall Atrium
Goodwin Hall Atrium & Torgersen Hall Atrium
Virginia Tech
-
Plenary Talks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
401
Geometric and Algebraic Methods for Constructing Hierarchically Low-Rank Matrices
Hierarchically low-rank (H-LR) matrices have been widely used to design fast solvers for integral equations, boundary element methods, discretized PDEs, and kernel matrices in statistical and machine learning. The computational bottleneck in these solvers is often the construction algorithm which converts a standard dense matrix into an H-LR format. We will present two types of algorithms for fast construction of H-LR matrices. One type of algorithm exploits geometric information that works well with high dimensional data from ML applications. Another type of algorithm is purely algebraic, which works well for building general-purpose linear solvers. In both algorithms, randomization plays a significant role.
Speaker: Sherry Li (LBNL) -
402
Nodal Statistics for Graphs and Matrices
Given a symmetric matrix with a given sign pattern, what can the sign patterns of its eigenvectors look like? This simple question is closely related to the study of discrete nodal statistics, and draws strong parallels with classical results in analysis for Laplacian eigenfunctions. In this talk, we will give an overview of the field, covering key results on nodal sets for graphs and their connection to known results and open problems in the continuous setting. In addition, we will discuss some recent progress towards a more complete understanding of the extremal properties of the nodal statistics of a matrix.
Speaker: John Urschel (MIT)
-
401
-
403
Closing Remarks McBryde Hall 100
McBryde Hall 100
Virginia Tech
-
374