Yue Shen, an ECE Ph.D. student in our lab, defended his dissertation entitled “Learning safe regions in high-dimensional dynamical systems via recurrent sets” on Friday, September 13th. Congratulations Dr Shen!
Year: 2024
1 paper accepted to Allerton
Our paper on generalized Barrier function conditions [1] has been accepted to the 60th Allerton Conference. Congrats to Yue for leading this work!
[Bibtex] [Abstract] [Download PDF]
Barrier functions constitute an effective tool for assessing and enforcing safety-critical constraints on dynamical systems. To this end, one is required to find a function $h$ that satisfies a Lyapunov-like differential condition, thereby ensuring the invariance of its zero super-level set $h_\ge 0$. This methodology, however, does not prescribe a general method for finding the function $h$ that satisfies such differential conditions, which, in general, can be a daunting task. In this paper, we seek to overcome this limitation by developing a generalized barrier condition that makes the search for $h$ easier. We do this in two steps. First, we develop integral barrier conditions that reveal equivalent asymptotic behavior to the differential ones, but without requiring differentiability of $h$. Subsequently, we further replace the stringent invariance requirement on $h≥0$ with a more flexible concept known as recurrence. A set is ($τ$-)recurrent if every trajectory that starts in the set returns to it (within $τ$ seconds) infinitely often. We show that, under mild conditions, a simple sign distance function can satisfy our relaxed condition and that the ($τ$-)recurrence of the super-level set $h_≥ 0$ is sufficient to guarantee the system’s safety.
@inproceedings{ssm2024allerton,
abstract = {Barrier functions constitute an effective tool for assessing and enforcing safety-critical constraints on dynamical systems. To this end, one is required to find a function $h$ that satisfies a Lyapunov-like differential condition, thereby ensuring the invariance of its zero super-level set $h_\ge 0$. This methodology, however, does not prescribe a general method for finding the function $h$ that satisfies such differential conditions, which, in general, can be a daunting task. In this paper, we seek to overcome this limitation by developing a generalized barrier condition that makes the search for $h$ easier. We do this in two steps. First, we develop integral barrier conditions that reveal equivalent asymptotic behavior to the differential ones, but without requiring differentiability of $h$. Subsequently, we further replace the stringent invariance requirement on $h≥0$ with a more flexible concept known as recurrence. A set is ($τ$-)recurrent if every trajectory that starts in the set returns to it (within $τ$ seconds) infinitely often. We show that, under mild conditions, a simple sign distance function can satisfy our relaxed condition and that the ($τ$-)recurrence of the super-level set $h_≥ 0$ is sufficient to guarantee the system's safety.},
author = {Shen, Yue and Sibai, Hussein and Mallada, Enrique},
booktitle = {60th Allerton Conference on Communication, Control, and Computing},
grants = {CPS-2136324, Global-Centers-2330450},
keywords = {Barrier Functions},
month = {09},
pages = {1-8},
pubstate = {presented},
record = {accepted Jul 2024, submitted Jul 2024},
title = {Generalized Barrier Functions: Integral Conditions & Recurrent Relaxations},
url = {https://mallada.ece.jhu.edu/pubs/2024-Allerton-SSM.pdf},
year = {2024}
}
Yue to join Apple
Yue will be joining Apple in Beijing after graduation. Congrats Yue!
3 papers accepted to CDC
Our papers on Dissipative Gradient Descent-Ascent algorithms for min-max problems [1], Accelerated saddle flow dynamics [2], and invertibility of sparse linear dynamical systems [3] have been accepted to the 63rd IEEE Conference on Decision and Control. Congrats to the leading students Tianqi, Yingzhu, Kyle, and to all the rest of the team!
[Bibtex] [Abstract] [Download PDF]
Gradient Descent Ascent (GDA) methods for min-max optimization problems typically produce oscillatory behavior that can lead to instability, e.g., in bilinear settings. To address this problem, we introduce a dissipation term into the GDA updates to dampen these oscillations. The proposed Dissipative GDA (DGDA) method can be seen as performing standard GDA on a state-augmented and regularized saddle function that does not strictly introduce additional convexity/concavity. We theoretically show the linear convergence of DGDA in the bilinear and strongly convex-strongly concave settings and assess its performance by comparing DGDA with other methods such as GDA, Extra-Gradient (EG), and Optimistic GDA. Our findings demonstrate that DGDA surpasses these methods, achieving superior convergence rates. We support our claims with two numerical examples that showcase DGDA’s effectiveness in solving saddle point problems.
@inproceedings{zlym2024cdc,
abstract = {Gradient Descent Ascent (GDA) methods for min-max optimization problems typically produce oscillatory behavior that can lead to instability, e.g., in bilinear settings.
To address this problem, we introduce a dissipation term into the GDA updates to dampen these oscillations. The proposed Dissipative GDA (DGDA) method can be seen as performing standard GDA on a state-augmented and regularized saddle function that does not strictly introduce additional convexity/concavity. We theoretically show the linear convergence of DGDA in the bilinear and strongly convex-strongly concave settings and assess its performance by comparing DGDA with other methods such as GDA, Extra-Gradient (EG), and Optimistic GDA.
Our findings demonstrate that DGDA surpasses these methods, achieving superior convergence rates. We support our claims with two numerical examples that showcase DGDA's effectiveness in solving saddle point problems.},
author = {Zheng, Tianqi and Loizou, Nicolas and You, Pengcheng and Mallada, Enrique},
booktitle = {63rd IEEE Conference on Decision and Control (CDC)},
grants = {CPS-2136324, Global-Centers-2330450},
month = {12},
note = {also in L-CSS},
pubstate = {presented},
record = {accepted Jul 2024, submitted Mar 2024},
title = {Dissipative Gradient Descent Ascent Method: A Control Theory Inspired Algorithm for Min-max Optimization},
url = {https://mallada.ece.jhu.edu/pubs/2024-CDC-ZLYM.pdf},
year = {2024}
}
[Bibtex] [Abstract] [Download PDF]
Minimax problems have attracted much attention due to various applications in constrained optimization problems and zero-sum games. Identifying saddle points within these problems is crucial, and saddle flow dynamics offer a straightforward yet useful approach. This study focuses on a class of bilinearly coupled minimax problems and designs an accelerated algorithm based on saddle flow dynamics that achieves a convergence rate beyond stereotype limits. The algorithm is derived based on a sequential two-step transformation of the objective function. First, a change of variable is aimed at a better-conditioned saddle function. Second, a proximal regularization, when staggered with the first step, guarantees strong convexity-strong concavity of the objective function that can be tuned for accelerated exponential convergence. Besides, such an approach can be extended to a class of weakly convex-weakly concave functions and still achieve exponential convergence to one stationary point. The theory is verified by a numerical test on an affine equality-constrained convex optimization problem.
@inproceedings{lmly2024cdc,
abstract = {Minimax problems have attracted much attention due to various applications in constrained optimization problems and zero-sum games. Identifying saddle points within these problems is crucial, and saddle flow dynamics offer a straightforward yet useful approach. This study focuses on a class of bilinearly coupled minimax problems and designs an accelerated algorithm based on saddle flow dynamics that achieves a convergence rate beyond stereotype limits. The algorithm is derived based on a sequential two-step transformation of the objective function. First, a change of variable is aimed at a better-conditioned saddle function. Second, a proximal regularization, when staggered with the first step, guarantees strong convexity-strong concavity of the objective function that can be tuned for accelerated exponential convergence.
Besides, such an approach can be extended to a class of weakly convex-weakly concave functions and still achieve exponential convergence to one stationary point. The theory is verified by a numerical test on an affine equality-constrained convex optimization problem.},
author = {Liu, Yingzhu and Mallada, Enrique and Li, Zhongkui and You, Pengcheng},
booktitle = {63rd IEEE Conference on Decision and Control (CDC)},
grants = {CPS-2136324, Global-Centers-2330450},
month = {07},
pubstate = {accepted, submitted Mar 2024},
title = {Accelerated Saddle Flow Dynamics for Bilinearly Coupled Minimax Problems},
url = {https://mallada.ece.jhu.edu/pubs/2024-Preprint-LMLY.pdf},
year = {2024}
}
[Bibtex] [Abstract] [Download PDF]
One of the fundamental problems of interest for discrete-time linear systems is whether its input sequence may be recovered given its output sequence, a.k.a. the left inversion problem. Many conditions on the state space geometry, dynamics, and spectral structure of a system have been used to characterize the well-posedness of this problem, without assumptions on the inputs. However, certain structural assumptions, such as input sparsity, have been shown to translate to practical gains in the performance of inversion algorithms, surpassing classical guarantees. Establishing necessary and sufficient conditions for left invertibility of systems with sparse inputs is therefore a crucial step toward understanding the performance limits of system inversion under structured input assumptions. In this work, we provide the first necessary and sufficient characterizations of left invertibility for linear systems with sparse inputs, echoing classic characterizations for standard linear systems. The key insight in deriving these results is in establishing the existence of two novel geometric invariants unique to the sparse-input setting, the weakly unobservable and strongly reachable subspace arrangements. By means of a concrete example, we demonstrate the utility of these characterizations. We conclude by discussing extensions and applications of this framework to several related problems in sparse control.
@inproceedings{pmv2024cdc,
abstract = {One of the fundamental problems of interest for discrete-time linear systems is whether its input sequence may be recovered given its output sequence, a.k.a. the left inversion problem. Many conditions on the state space geometry, dynamics, and spectral structure of a system have been used to characterize the well-posedness of this problem, without assumptions on the inputs. However, certain structural assumptions, such as input sparsity, have been shown to translate to practical gains in the performance of inversion algorithms, surpassing classical guarantees. Establishing necessary and sufficient conditions for left invertibility of systems with sparse inputs is therefore a crucial step toward understanding the performance limits of system inversion under structured input assumptions. In this work, we provide the first necessary and sufficient characterizations of left invertibility for linear systems with sparse inputs, echoing classic characterizations for standard linear systems. The key insight in deriving these results is in establishing the existence of two novel geometric invariants unique to the sparse-input setting, the weakly unobservable and strongly reachable subspace arrangements. By means of a concrete example, we demonstrate the utility of these characterizations. We conclude by discussing extensions and applications of this framework to several related problems in sparse control.},
author = {Poe, Kyle and Mallada, Enrique and Vidal, Rene},
booktitle = {63rd IEEE Conference on Decision and Control (CDC)},
grants = {CPS-2136324, Global-Centers-2330450},
month = {12},
pubstate = {presented},
record = {accepted Jul 2024, submitted Mar 2024},
title = {Invertibility of Discrete-Time Linear Systems with Sparse Inputs},
url = {https://mallada.ece.jhu.edu/pubs/2024-CDC-PMV.pdf},
year = {2024}
}
1 paper accepted to L-CSS
Our paper on Dissipative Gradient Descent-Ascent algorithms for min-max problems [1] has been accepted to the IEEE Control Systems Letters. Congrats Tianqi!
[Bibtex] [Abstract] [Download PDF]
Gradient Descent Ascent (GDA) methods for min-max optimization problems typically produce oscillatory behavior that can lead to instability, e.g., in bilinear settings. To address this problem, we introduce a dissipation term into the GDA updates to dampen these oscillations. The proposed Dissipative GDA (DGDA) method can be seen as performing standard GDA on a state-augmented and regularized saddle function that does not strictly introduce additional convexity/concavity. We theoretically show the linear convergence of DGDA in the bilinear and strongly convex-strongly concave settings and assess its performance by comparing DGDA with other methods such as GDA, Extra-Gradient (EG), and Optimistic GDA. Our findings demonstrate that DGDA surpasses these methods, achieving superior convergence rates. We support our claims with two numerical examples that showcase DGDA’s effectiveness in solving saddle point problems.
@article{zlym2024lcss,
abstract = {Gradient Descent Ascent (GDA) methods for min-max optimization problems typically produce oscillatory behavior that can lead to instability, e.g., in bilinear settings.
To address this problem, we introduce a dissipation term into the GDA updates to dampen these oscillations. The proposed Dissipative GDA (DGDA) method can be seen as performing standard GDA on a state-augmented and regularized saddle function that does not strictly introduce additional convexity/concavity. We theoretically show the linear convergence of DGDA in the bilinear and strongly convex-strongly concave settings and assess its performance by comparing DGDA with other methods such as GDA, Extra-Gradient (EG), and Optimistic GDA.
Our findings demonstrate that DGDA surpasses these methods, achieving superior convergence rates. We support our claims with two numerical examples that showcase DGDA's effectiveness in solving saddle point problems.},
author = {Zheng, Tianqi and Loizou, Nicolas and You, Pengcheng and Mallada, Enrique},
doi = {10.1109/LCSYS.2024.3413004},
grants = {CPS-2136324, Global-Centers-2330450},
journal = {IEEE Control Systems Letters (L-CSS)},
month = {06},
pages = {2009-2014},
record = {published, accepted May 2024, submitted Mar 2024},
title = {Dissipative Gradient Descent Ascent Method: A Control Theory Inspired Algorithm for Min-max Optimization},
url = {https://mallada.ece.jhu.edu/pubs/2024-LCSS-ZLYM.pdf},
volume = {8},
year = {2024}
}
Roy to do an internship at Amazon
This summer, Roy will be doing an internship with the LLMs team at Amazon. Congrats Roy!
2 papers accepted to PES General Meeting
Our papers on decentralized stability analysis for grid forming control [1] and on grid forming based grid shaping [2] have been accepted to PES General Meeting!
[Bibtex] [Abstract] [Download PDF]
This paper presents a decentralized stability analysis of power systems comprising grid-forming (GFM) inverters. We leverage a decentralized stability framework capable of ensuring the stability of the entire interconnection through individual assessments at each bus. The key novelty lies in incorporating voltage dynamics and their coupling with reactive power, in addition to the angle dynamics and their coupling with active power. We perform loop transformation to address the challenge posed by the non-Laplacian nature of the network Jacobian matrix in this case. This methodology is applied to characterize conditions on the droop gains of GFM controllers that can preserve system-wide stability. Our proposed stability criteria exhibit scalability and robustness, and can be extended to accommodate delays, variations in network conditions, and plug-and-play of new components in the network.
@inproceedings{smg2024pesgm,
abstract = {This paper presents a decentralized stability analysis of power systems comprising grid-forming (GFM) inverters. We leverage a decentralized stability framework capable of ensuring the stability of the entire interconnection through individual assessments at each bus. The key novelty lies in incorporating voltage dynamics and their coupling with reactive power, in addition to the angle dynamics and their coupling with active power. We perform loop transformation to address the challenge posed by the non-Laplacian nature of the network Jacobian matrix in this case. This methodology is applied to characterize conditions on the droop gains of GFM controllers that can preserve system-wide stability. Our proposed stability criteria exhibit scalability and robustness, and can be extended to accommodate delays, variations in network conditions, and plug-and-play of new components in the network.},
author = {Siahaan, Zudika and Mallada, Enrique and Geng, Sijia},
booktitle = {PES General Meeting},
doi = {10.1109/PESGM51994.2024.10689037},
grants = {CPS-2136324, CAREER-1752362, Global Centers-2330450},
month = {06},
pages = {1-5},
record = {presented Jun. 2024, accepted Mar. 2024, submitted Nov. 2023},
title = {Decentralized Stability Criteria for Grid-Forming Control in Inverter-Based Power Systems},
url = {https://mallada.ece.jhu.edu/pubs/2024-PESGM-SMG.pdf},
year = {2024}
}
[Bibtex] [Abstract] [Download PDF]
We consider the problem of controlling the frequency response of weakly-coupled multi-machine multi-inverter low-inertia power systems via grid-forming inverter-based resources (IBRs). In contrast to existing methods, our approach relies on dividing the larger system into multiple strongly-coupled subsystems, without ignoring either the underlying network or approximating the subsystem response as an aggregate harmonic mean model. Rather, through a structured clustering and recursive dynamic shaping approach, the frequency response of the overall system to load perturbations is shaped appropriately. We demonstrate the proposed approach for a three-node triangular configuration and a small-scale radial network. Furthermore, previous synchronization analysis for heterogeneous systems requires the machines to satisfy certain proportionality property. In our approach, the effective transfer functions for each cluster can be tuned by the IBRs to satisfy such property, enabling us to apply the shaping control to systems with a wider range of heterogeneous machines.
@inproceedings{plbmg2024pesgm,
abstract = {We consider the problem of controlling the frequency response of weakly-coupled multi-machine multi-inverter low-inertia power systems via grid-forming inverter-based resources (IBRs). In contrast to existing methods, our approach relies on dividing the larger system into multiple strongly-coupled subsystems, without ignoring either the underlying network or approximating the subsystem response as an aggregate harmonic mean model. Rather, through a structured clustering and recursive dynamic shaping approach, the frequency response of the overall system to load perturbations is shaped appropriately. We demonstrate the proposed approach for a three-node triangular configuration and a small-scale radial network. Furthermore, previous synchronization analysis for heterogeneous systems requires the machines to satisfy certain proportionality property. In our approach, the effective transfer functions for each cluster can be tuned by the IBRs to satisfy such property, enabling us to apply the shaping control to systems with a wider range of heterogeneous machines.},
author = {Poolla, Bala Kameshwar and Lin, Yashen and Bernstein, Andrey and Mallada, Enrique and Groß, Dominic},
booktitle = {PES General Meeting},
doi = {10.1109/PESGM51994.2024.10688717},
grants = {CPS-2136324, CAREER-1752362, Global Centers-2330450},
month = {06},
pages = {1-5},
record = {presented Jun. 2024, accepted Mar. 2024, submitted Nov. 2023},
title = {Dynamic Shaping of Grid Response of Multi-Machine Multi-Inverter Systems Through Grid-Forming IBRs},
url = {https://mallada.ece.jhu.edu/pubs/2024-PESGM-PLBMG.pdf},
year = {2024}
}
1 paper accepted to ICLR
Our paper on Early Neuron Alignment in Two-layer ReLU Networks [1] has been accepted to the International Conference on Representation Learning. Congrats Hancheng!
[Bibtex] [Abstract] [Download PDF]
This paper studies the problem of training a two-layer ReLU network for binary classification using gradient flow with small initialization. We consider a training dataset with well-separated input vectors: Any pair of input data with the same label are positively correlated, and any pair with different labels are negatively correlated. Our analysis shows that, during the early phase of training, neurons in the first layer try to align with either the positive data or the negative data, depending on its corresponding weight on the second layer. A careful analysis of the neurons’ directional dynamics allows us to provide an $\mathcalO(\fracłog n\sqrtμ)$ upper bound on the time it takes for all neurons to achieve good alignment with the input data, where $n$ is the number of data points and $μ$ measures how well the data are separated. After the early alignment phase, the loss converges to zero at a $\mathcalO(\frac1t)$ rate, and the weight matrix on the first layer is approximately low-rank. Numerical experiments on the MNIST dataset illustrate our theoretical findings.
@inproceedings{mvm2024iclr,
abstract = {This paper studies the problem of training a two-layer ReLU network for binary classification using gradient flow with small initialization. We consider a training dataset with well-separated input vectors: Any pair of input data with the same label are positively correlated, and any pair with different labels are negatively correlated. Our analysis shows that, during the early phase of training, neurons in the first layer try to align with either the positive data or the negative data, depending on its corresponding weight on the second layer. A careful analysis of the neurons' directional dynamics allows us to provide an $\mathcalO(\fracłog n\sqrtμ)$ upper bound on the time it takes for all neurons to achieve good alignment with the input data, where $n$ is the number of data points and $μ$ measures how well the data are separated. After the early alignment phase, the loss converges to zero at a $\mathcalO(\frac1t)$ rate, and the weight matrix on the first layer is approximately low-rank. Numerical experiments on the MNIST dataset illustrate our theoretical findings.},
author = {Min, Hancheng and Vidal, Rene and Mallada, Enrique},
booktitle = {International Conference on Representation Learning (ICLR)},
grants = {CAREER-1752362},
month = {05},
record = {published, accepted Jan 2024, submitted Sep 2023},
title = {Early Neuron Alignment in Two-layer ReLU Networks with Small Initialization},
url = {https://mallada.ece.jhu.edu/pubs/2024-ICLR-MVM.pdf},
year = {2024}
}