Slide 1

Slide 1 text

Spectral Sparsification of Hypergraphs Tasuku Soma (Institute of Statistical Mathematics; Tokyo) Joint work with Kam Chuen Tung (Univerisity of Waterloo) Yuichi Yoshida (Natinal Institute of Informatics; Tokyo) 1 / 20

Slide 2

Slide 2 text

1 Overview 2 Algorithm 3 Summary 2 / 20

Slide 3

Slide 3 text

Graph Laplacian G = (V, E, w): (undirected) graph with edge weight w ∈ RE + Graph Laplacian Matrix LG = e={i,j}∈E we (ei − ej )(ei − ej )⊤ 3 / 20

Slide 4

Slide 4 text

Graph Laplacian G = (V, E, w): (undirected) graph with edge weight w ∈ RE + Graph Laplacian Matrix LG = e={i,j}∈E we (ei − ej )(ei − ej )⊤ Quadratic form x⊤LG x = e={i,j}∈E we (xi − xj )2 Energy of G (as circuit) with potential x 3 / 20

Slide 5

Slide 5 text

Graph Laplacian G = (V, E, w): (undirected) graph with edge weight w ∈ RE + Quadratic form x⊤LG x = e={i,j}∈E we (xi − xj )2 Cut function For a vertex set X ⊆ V , κG (X) = {i,j}∈E:|{i,j}∩X|=1 we = 1⊤ X LG 1X . X The cut function is the quadratic form of Laplacian. 3 / 20

Slide 6

Slide 6 text

Spectral Graph Sparsification [Spielman, Teng 2011] A weighted subgraph ˜ G = (V, ˜ E , ˜ w ) of G is an ε-spectral sparsifier △ ⇐⇒ (1 − ε)x⊤LG x ≤ x⊤L ˜ G x ≤ (1 + ε)x⊤LG x (∀x ∈ RV ) edge weight can be modified edge subset ˜ E ⊆ E G ˜ G • Spectral spasifier =⇒ all cuts are preserved (i.e., cut sparsifier) • Current best: O(n/ε2) edges [Batson, Spielman, Srivastava 2014; Y. T. Lee, Sun 2015] 4 / 20

Slide 7

Slide 7 text

Hypergraph Cut H = (V, E): hypergraph with edge weight w ∈ RE + Hypergraph cut function For vertex subset X ⊆ V , κH (X) = e∈E:0<|e∩X|<|e| we Hypergraph cut functions are NOT quadratic forms! 5 / 20

Slide 8

Slide 8 text

Spectral Hypergraph Sparsification [Soma, Yoshida 2019] Energy function QH (x) = e we max i,j∈e (xi − xj )2 Especially, QH (1X ) = κH (X) for X ⊆ V . 6 / 20

Slide 9

Slide 9 text

Spectral Hypergraph Sparsification [Soma, Yoshida 2019] Energy function QH (x) = e we max i,j∈e (xi − xj )2 Especially, QH (1X ) = κH (X) for X ⊆ V . A weighted subhypergraph ˜ H of H is an ε-spectral sparsifier △ ⇐⇒ (1 − ε)QH (x) ≤ Q ˜ H (x) ≤ (1 + ε)QH (x) (∀x ∈ RV ) spectral sparsifier =⇒ cut sparsifier [Kogan, Krauthgamer 2015] 6 / 20

Slide 10

Slide 10 text

Our Result Our result (Soma, Yoshida SODA’19) For any hypergraph H and ε ∈ (0, 1), we can construct an ε-spectral sparsifier ˜ H with O n3 log n ε2 many hyperedges in polynomial time w.h.p. (n = |V |) Applications • Speeding up algorithms for cuts/flows in hypergraphs; • Semi-supervised learning on hypergraphs [Zhang, Hu, Tang, Chan 2017]. • Agnostic learning of hypernetwork type submodular functions 7 / 20

Slide 11

Slide 11 text

Known bounds for offline sparsification reference cut spectral Kogan, Krauthgamer (2015) O(n(r+log n) ε2 ) Soma, Yoshida (2019) O(n3 ε2 log n) Bansal, Svensson, Trevisan (2019) O(nr3 ε2 log n) Chen, Khanna, Nagda (2020) O( n ε2 log n) Kapralov, Krauthgamer, Tardos, Yoshida (2021) ˜ O( nr εO(1) ) Kapralov, Krauthgamer, Tardos, Yoshida (2022) O( n ε4 log3 n) J. R. Lee (2023) Jambulapati, Liu, Sidford (2023) O( n ε2 log n log r) n = |V |, r = maxe∈E |e| 8 / 20

Slide 12

Slide 12 text

Drawbacks of offline sparsification algorithms To store an input hypergraph, we need space of size exponential in n in the worst case. On the other hand, output sparsifiers have only O(ε−2n log n log r) hyperedges, which is nearly linear! Q. Is it possible to construct a spectral sparsifier of hypergraphs using only polynomial space? 9 / 20

Slide 13

Slide 13 text

Online hypergraph spectral sparsification • Hyperedges and weights (e1 , w1 ), (e2 , w2 ), . . . , (em , wm ) arrive in stream • When (ei , wi ) arrives, we must irrevocably decide whether to include ei as well as its weight in ˜ H • Only poly(n) working space is available 10 / 20

Slide 14

Slide 14 text

Online hypergraph spectral sparsification • Hyperedges and weights (e1 , w1 ), (e2 , w2 ), . . . , (em , wm ) arrive in stream • When (ei , wi ) arrives, we must irrevocably decide whether to include ei as well as its weight in ˜ H • Only poly(n) working space is available A weighted subhypergraph ˜ H of H is an (ε, δ)-spectral sparsifier △ ⇐⇒ (1 − ε)QH (x) − δ∥x∥2 2 ≤ Q ˜ H (x) ≤ (1 + ε)QH (x) + δ∥x∥2 2 (∀x ∈ RV ) 10 / 20

Slide 15

Slide 15 text

Online hypergraph spectral sparsification • Hyperedges and weights (e1 , w1 ), (e2 , w2 ), . . . , (em , wm ) arrive in stream • When (ei , wi ) arrives, we must irrevocably decide whether to include ei as well as its weight in ˜ H • Only poly(n) working space is available A weighted subhypergraph ˜ H of H is an (ε, δ)-spectral sparsifier △ ⇐⇒ (1 − ε)QH (x) − δ∥x∥2 2 ≤ Q ˜ H (x) ≤ (1 + ε)QH (x) + δ∥x∥2 2 (∀x ∈ RV ) Theorem (Soma, Tung, and Yoshida IPCO’24) There is an online algorithm that outputs an (ε, δ)-spectral sparsifier with O(n log n log r ε2 log(εW/δn)) many hyperedges using O(n2) space. (Here, W = i wi) 10 / 20

Slide 16

Slide 16 text

1 Overview 2 Algorithm 3 Summary 11 / 20

Slide 17

Slide 17 text

Edge sampling algorithm Algorithm 1: Compute edge sampling probability pe ∈ [0, 1] for every hyperedge e ∈ E. 2: for each hyperedge e : 3: Add a copy of e to ˜ H with weight we /pe w.p. pe . Properties • E ˜ H [Q ˜ H (x)] = QH (x) (i.e., unbiased) • Expected total # of hyperedges = e pe 12 / 20

Slide 18

Slide 18 text

Challenge small pe large pe # edges sparse dense variance large small Need to carefully design pe that balances both! 13 / 20

Slide 19

Slide 19 text

Reweighting [Kapralov, Krauthgamer, Tardos, Yoshida 2022; J. R. Lee 2023; Jambulapati, Liu, Sidford 2023] Clique expansion: G = (V, F), each hyperedge being replaced with clique 14 / 20

Slide 20

Slide 20 text

Reweighting [Kapralov, Krauthgamer, Tardos, Yoshida 2022; J. R. Lee 2023; Jambulapati, Liu, Sidford 2023] Clique expansion: G = (V, F), each hyperedge being replaced with clique Lemma ([Kapralov, Krauthgamer, Tardos, Yoshida 2022; J. R. Lee 2023]) There is an edge weight ce,u,v ≥ 0 on F such that • u,v∈e ce,u,v = we (e ∈ E) • ce,u,v > 0 =⇒ ru,v = maxu′,v′∈e ru′,v′ , where ru,v is the effective resistance between u, v. Effective resistance: ru,v = (eu − ev)⊤L+ G (eu − ev) 14 / 20

Slide 21

Slide 21 text

Reweighting [Kapralov, Krauthgamer, Tardos, Yoshida 2022; J. R. Lee 2023; Jambulapati, Liu, Sidford 2023] Clique expansion: G = (V, F), each hyperedge being replaced with clique Lemma ([Kapralov, Krauthgamer, Tardos, Yoshida 2022; J. R. Lee 2023]) There is an edge weight ce,u,v ≥ 0 on F such that • u,v∈e ce,u,v = we (e ∈ E) • ce,u,v > 0 =⇒ ru,v = maxu′,v′∈e ru′,v′ , where ru,v is the effective resistance between u, v. Effective resistance: ru,v = (eu − ev)⊤L+ G (eu − ev) Given {ce,u,v }, setting pe ∝ we maxu,v∈e ru,v yields ε-spectral sparsifier with O(ε−2n log n log r) hyperedges. [J. R. Lee 2023; Jambulapati, Liu, Sidford 2023] Graph case: effective registance sampling [Spielman, Srivastava 2011] 14 / 20

Slide 22

Slide 22 text

Convex optimization for reweighting [J. R. Lee 2023] The edge weight {ce,u,v } can be found by convex optimization: max log det   e∈E u,v∈V ce,u,v Lu,v + J   sub. to u,v∈e ce,u,v = we (e ∈ E) ce,u,v ≥ 0 (e ∈ E, u, v ∈ e) Lu,v := (eu −ev )(eu −ev )⊤ J: all-one matrix KKT condition =⇒ the required conditions for reweighting. 15 / 20

Slide 23

Slide 23 text

Our online algorithm (cf. online spectral sparsification of graphs [Cohen, Musco, Pachocki 2020]) Let η = δ/ε > 0. (regularization parameter) Define matrix Li (i = 0, 1, . . . ) as follows: • L0 = On , • Li = Li−1 + u,v∈ei wi ci,u,v Lu,v , where ci,u,v is the solution of the following convex optimization: max ci∈∆ei log det Li−1 + u,v∈ei wi ci,u,v Lu,v + ηIn Laplacian matrix up to i − 1 ridge regularizer ∆e : probability simplex in R (e 2 ) 16 / 20

Slide 24

Slide 24 text

Our online algorithm Let η = δ/ε > 0. 1: ˜ H0 := ∅, L0 := On 2: for i = 1, 2, . . . : 3: Solve the following convex optimization: max ci∈∆ei log det Li−1 + u,v∈ei wi ci,u,v Lu,v + ηIn 4: Li ← Li−1 + u,v∈ei wi ci,u,v Lu,v 5: pe ∝ wi maxu,v∈ei ∥(Li + ηI)−1/2(eu − ev )∥2 2 . 6: Add ei with weight wi /pe to ˜ Hi−1 with probability pe and obtain ˜ Hi . Laplacian matrix for reweighted clique expansion ridge regularizer ridged effective resistance ∆e : prob. simplex in R (e 2 ) Lu,v := (eu −ev )(eu −ev )⊤ 17 / 20

Slide 25

Slide 25 text

Analysis Hi : input hypergraph up to time i Our algorithm satisfies: • Li depends only on Hi , not on algorithm’s random choice. −→ Sampling of each ei is independent • ˜ Hi is an (ε, δ)-spectral sparsifier of Hi w.h.p. The expected # of hyperedges = O(n log n log r ε2 log(εW/δn)) (application of chaining analysis by [J. R. Lee 2023] ) • maintains only sparsifier ˜ Hi and Laplacian matrix Li −→ O(n2) space 18 / 20

Slide 26

Slide 26 text

1 Overview 2 Algorithm 3 Summary 19 / 20

Slide 27

Slide 27 text

Summary • Introduced the concept of spectral sparsification of hypergraphs. [Soma, Yoshida 2019] • Current best: O n log n log r ε2 hyperedges. [J. R. Lee 2023; Jambulapati, Liu, Sidford 2023] • Online algorithm: O(n log n log r ε2 log(εW/δn)) hyperedges with O(n2) space • We also show the number of hyperedges is tight for online setting 20 / 20

Slide 28

Slide 28 text

Summary • Introduced the concept of spectral sparsification of hypergraphs. [Soma, Yoshida 2019] • Current best: O n log n log r ε2 hyperedges. [J. R. Lee 2023; Jambulapati, Liu, Sidford 2023] • Online algorithm: O(n log n log r ε2 log(εW/δn)) hyperedges with O(n2) space • We also show the number of hyperedges is tight for online setting Thanks! Questions? 20 / 20