Professional Documents
Culture Documents
Abstract—Signal processing over graphs has recently attracted example of GSP applications is in processing data from sensor
significant attentions for dealing with structured data. Normal networks [5]. Based on graph models directly built over
graphs, however, only model pairwise relationships between network structures, a graph Fourier space could be defined
nodes and are not effective in representing and capturing some
according to the eigenspace of a representing graph matrix
arXiv:1907.09203v3 [eess.SP] 3 Oct 2019
(a) Example of a trait and genes: (b) Example of pairwise relation- (c) Example of multi-lateral
the trait t1 is triggered by three ships: arrow links represent the in- relationships: the solid circu-
genes v1 , v2 , v3 and the genes fluences from genes, whereas the lar line represents a quadrilat-
may also influence each other. potential interactions among genes eral relationship among four
cannot be captured. nodes, while the purple dash
lines represent the potential
inter-node interactions in this
quadrilateral relationship.
(a) A 3-uniform hypergraph H with 7 nodes, 3 (b) A general hypergraph H with 7 nodes, 3
hyperedges and m.c.e(H) = 3, where V = hyperedges and m.c.e(H) = 3, where V =
{v1 , · · · , v7 } and E = {e1 , e2 , e3 }. Three hyper- {v1 , · · · , v7 } and E = {e1 , e2 , e3 }. Three hyper-
edges are e1 = {v1 , v4 , v6 }, e2 = {v2 , v3 , v7 } edges are e1 = {v1 , v4 , v6 }, e2 = {v2 , v3 } and
and e3 = {v5 , v6 , v7 }. e3 = {v5 , v6 , v7 }.
Definition 9 (Hypergraph frequency). Hypergraph frequency Since λ is real and nonnegative, we have
describes how oscillatory the signal component is with respect
λj − λi
to the given hypergraph. A frequency component fr is asso- TV(f i ) − TV(f j ) = . (37)
λmax
ciated with a higher frequency if the total variation of this
frequency component is larger. Obviously, TV(fi ) > TV(fj ) iff λi < λj .
Theorem 1 shows that the supporting matrix Ps can help
Note that, the physical meaning of graph frequency was
us apply the total variation more efficiently in some real
stated in GSP [2]. Generally, the graph frequency is highly
applications. Moreover, it provides the order of frequency
related to the total variation of the corresponding frequency
according to the coefficients λi ’s with the following property.
component. Similarly, the hypergraph frequency also relates
to the corresponding total variation. We will discuss more Property 4. A smaller λ is related to a higher frequency in the
about the interpretation of the hypergraph frequency and its hypergraph Fourier space, where its corresponding spectrum
relationships with DSP and GSP later in Section IV-A, to basis is called a high frequency component.
further solidate our hypergraph frequency definition.
Based on the definition of total variation, we describe one
F. Signals with Limited Spectrum Support
important property of TV(f r ) in the following theorem.
With the order of frequency, we define the bandlimited
Theorem 1. Define a supporting matrix signals as follows.
f1T
λ1
1 .. Definition 10 (Bandlimited signal). Order the coefficients
Ps =
f1 · · · fN .. . . (33)
λmax . as λ = [λ1 · · · λN ] where λ1 ≥ · · · ≥ λN ≥ 0,
T together with their corresponding fr ’s. A hypergraph signal
λN fN
1
s[M −1] is defined as K-bandlimited if the HGFT transformed
With the normalized representing tensor Fnorm = λmax F, the signal ŝ = [ŝ1 , · · · , ŝN ]T has ŝi = 0 for all i ≥ K
total variation of hypergraph spectrum fr is calculated as where K ∈ {1, 2, · · · , N }. The smallest K is defined as
norm
TV(fr ) = ||fr − fr(1) ||1 , (34a) the bandwidth and the corresponding boundary is defined as
W = λK .
= ||fr − Ps fr ||1 , (34b)
λr Note that, a larger λi corresponds to a lower frequency as
= |1 − |. (34c) we mentioned in Property 4. Then, the frequency are ordered
λmax
from low to high in the definition above. Moreover, we use
Moreover, TV(fi ) > TV(fj ) iff λi < λj . the index K instead of the coefficient value λ to define the
Proof: For hypergraph signals, the output of one-time bandwidth for the following reasons:
shifting of fr is calculated as • Identical λ’s in two diferent hypergraphs do not refer to
N
X the same frequency. Since each hypergraph has its own
fr(1) = λi fi (fiT fr )M −1 = λr fr . (35) adjacency tensor and spectrum space, the comparison
i=1 of multiple spectrum pairs (λi , fi )’s is only meaningful
λr within the same hypergraph. Moreover, there exists a
Based on the normalized Fnorm , we have fr(1)
norm
= λmax fr . It
normalization issue in the decomposition of different
is therefore easy to obtain Eq. (34c) from Eq. (34a). To obtain
adjacency tensors. Thus, it is not meaningful to compare
Eq. (34b), we have
λk ’s across two different hypergraphs.
N • Since λk values are not continuous over k, different
X λr
Ps f r = λi fi (fiT fr ) = fr . (36) frequency cutoffs of λ may lead to the same bandlim-
i=1
λmax
ited space. For example, suppose that λk = 0.5 and
It is clear that Eq. (34b) is the same as Eq. (34c). λk+1 = 0.8. Then, λ = 0.6 and λ0 = 0.7 would lead
11
to the same cutoff in the frequency space, which makes IV. D ISCUSSIONS AND I NTERPRETATIONS
bandwidth definition non-unique. In this section, we focus on the insights and physical mean-
As we discussed in Section III-D, the hypergraph signal is ing of frequency to help interpret the hypergraph spectrum
the Hadamard product of the original signal in the frequency space. We also consider the relationships between HGSP
domain. Then, we have the following property of bandwidth. and other existing works to better understand the HGSP
framework.
Property 5. The bandwidth K is the same based on the HGFT
of the hypergraph signals ŝ and that of the original signals s̃.
A. Interpretation of Hypergraph Spectrum Space
This property allows us to analyze the spectrum support of We are interested in an intuitive interpretation of the hy-
the hypergraph signal by looking into the original signal with pergraph frequency and its relations with the DSP and GSP
lower complexity. Recall that we can add fi by using zero frequencies. We start with the frequency and the total variation
coefficients λi when R < N as mentioned in Section III-C. in DSP. In DSP, the discrete Fourier transform (DFT) of a
The added basis should not affect the HGFT signals in Fourier sequence sn is given by ŝk =
PN −1 −j 2πkn
n=0 sn e and the
N
space. According to the structure of bandlimited signal, we n
frequency is defined as νn = N , n = 0, 1, · · · , N − 1. From
need the added fi could meet the following conditions: (1) [39], we can easily summarize the following conclusions:
fi ⊥ fp for p 6= i; (2) fiT · s → 0; and (3) |fi | = 1. N
• νn : 1 < n < 2 − 1 corresponds to a continuous time
n
signal frequency N fs ;
N
• νn : 2 + 1 < n < N − 1 corresponds to a continuous
G. Implementation and Complexity n
time signal frequency −(1 − N )fs ;
We now consider the implementation and complexity issues • ν N corresponds to fs /2;
2
of HGFT. Similar to GFT, the process of HGFT consists of • n = 0 corresponds to frequency 0.
two steps: decomposition and execution. The decomposition is Here, fs is the critical sampling frequency. In tradi-
to calculate the hypergraph spectrum basis, and the execution tional we generate the Fourier transform fˆ(ω) =
transforms signals from the hypergraph vertex domain into the R ∞ DFT,−2πjxω n
−∞
f (x)e dx at each discrete frequency N fs , n =
spectrum domain. N N N N
− 2 + 1, − 2 + 2, · · · , 2 − 1, 2 . The highest and lowest
• The calculation of spectrum basis by the orthogonal- frequencies correspond to n = N/2 and n = 0, respectively.
n
CP decomposition is an important preparation step for Note that n varies from − N2 + 1 to N2 here. Since e−j2πk N =
HGFT. A straightforward algorithm would decompose the n+N
e−j2πk N , we can let n vary from 0 to N − 1 and cover
representing tensor F with the spectrum basis fi ’s and the complete period. Now, n varies in exact correspondence
coefficients λi ’s as in Eq. (24). Efficient tensor decom- to νn , and the aforementioned conclusions are drawn. The
position is an active topic in both fields of mathematics highest frequency occurs at n = N2 .
and engineering. There are a number of methods for CP The total variation in DSP is defined as the differences
decomposition in the literature. In [56], [60], motivated among the signals over time [61], i.e.,
by the spectral theorem for real symmetric matrices,
N −1
orthogonal-CP decomposition algorithms for symmetric X
TV(s) = |sn − s(n−1 mod N ) | (38a)
tensors are developed based polynomial equations. In
n=0
[26], Afshar et al. proposed a more general decompo-
= ||s − CN s||1 , (38b)
sition algorithm for spatio-temporal data. Other works,
including [57]–[59], tried to develop faster decomposition where
methods for signal processing and big data applications.
0 0 ··· 0 1
The rapid development of tensor decomposition and 1 0 ··· 0 0
the advancement of computation ability will benefit the
.. .. .. ..
..
efficient derivation of hypergraph spectrum. CN =
. . . . .. (39)
• The execution of HGFT with a known spectrum basis is ..
0 0 . 0 0
defined in Eq. (26b). According to Eq. (29), the HGFT of
0 0 ··· 1 0
hypergraph signal is an M − 1 times Hadamard product
of the original signal in the hypergraph spectrum space. When we perform the eigen-decomposition of CN , we see
2πn
This relationship can help execute HGFT and iHGFT of that the eigenvalues are λn = e−j N with eigenvector fn ,
hypergraph signals more efficiently by applying matrix 0 ≤ n ≤ N − 1. More specifically, the total variation of the
operations on the original signals. Clearly, the complexity frequency component fn is calculated as
of calculating the original signals in the frequency domain 2πn
TV(f n ) = |1 − ej N |, (40)
s̃ = Vs is O(N 2 ). In addition, since the computation
complexity of the power function x(M −1) could be which increases with n for n ≤ N2 before decreasing with n
O(log(M − 1)) and each vector has N entries, the for N2 < n ≤ N − 1.
complexity of calculating the M − 1 times Hadamard Obviously, the total variations of frequency components
product is O(N log(M − 1)). Thus, the complexity of have a one-to-one correspondence to frequencies in the order
general HGFT implementation is O(N 2 +N log(M −1)). of their values. If the total variation of a frequency component
12
general model for multi-lateral relationships while normal high-dimensional vertex domain, while HOS focuses on the
graphs are only able to model bilateral relationship. For statistical domain. In addition, the forms of signal combination
example, a 3-uniform hypergraph is able to model the are also different, where HGSP signals are based on the
trilateral interaction among users in a social network hypergraph shifting defined as in Eq. (21), whereas HOS
[63]. As hypergraph is a more general model for high- cumulants are based on the statistical average of shifted signal
dimensional interactions, HGSP is also more powerful for products.
high-dimensional signals. 3) Learning over Hypergraphs: Hypergraph learning is
• Algebraic models: HGSP relies on tensors while GSP another tool to handle structured data and sometimes uses
relies on matrices, which are second-order tensors. Ben- similar techniques to HGSP. For example, the authors of
efiting from the generality of tensor, HGSP is broadly [72] proposed an alternative definition of hypergraph total
applicable in high-dimensional data analysis. variation and design algorithms in accordance for classification
• Signals and signal shifting: In HGSP, we define the and clustering problems. In addition, hypergraph learning also
hypergraph signal as M − 1 times tensor outer product has its own definition of the hypergraph spectrum space. For
of the original signal. More specifically, the hypergraph example, [40], [41] represented the hypergraphs using a graph-
signal is the original signal if M = 2. Basically, the like similarity matrix and defined a spectrum space as the
hypergraph signal is the same as the graph signal if each eigenspace of this similarity matrix. Other works considered
hyperedge has exactly two nodes. Also shown in Fig. different aspects of hypergraph, including the hypergraph
10(b) of Section III-C, graph shifting is a special case of Laplacian [73] and hypergraph lifting [74].
hypergraph shifting when M = 2. The HGSP framework exhibits features different from hy-
• Spectrum properties: In HGSP, the spectrum space is pergraph learning: 1) HGSP defines a framework that general-
defined over the orthogonal-CP decomposition in terms of izes the classical digital signal processing and traditional graph
the basis and coefficients, which are also the E-eigenpairs signal processing; 2) HGSP applies different definitions of
of the representing tensor [64], shown in Eq. (27). In GSP, hypergraph characteristics such as the total variation, spectrum
the spectrum space is defined as the matrix eigenspace. space, and Laplacian; 3) HGSP cares more about the spectrum
Since the tensor algebra is an extension of matrix, the space while learning focuses more on data; 4) As HGSP is
HGSP spectrum is also an extension of the GSP spectrum. an extension of DSP and GSP, it is more suitable to handle
For example, as discussed in Section III, GFT is the same detailed tasks such as compression, denoising, and detection.
as HGFT when M = 2. All these features make HGSP a different technical concept
Overall, HGSP is an extension of GSP, which is both more from hypergraph learning.
general and novel. The purpose of developing the HGSP
V. T OOLS FOR H YPERGRPH S IGNAL P ROCESSING
framework is to facilitate more interesting signal processing
tasks that involve high-dimensional signal interactions. In this section, we introduce several useful tools built within
2) Higher-Order Statistics: Higher-order statistics (HOS) the framework of HGSP.
has been effectively applied in signal processing [65], [66],
A. Sampling Theory
which can analyze the multi-lateral interactions of signal
samples and have found successes in many applications, such Sampling is an important tool in data analysis, which selects
as blind feature detection [67], decision [68], and signal a subset of individual data points to estimate the characteristics
classifications [69]. In HOS, the kth-order cumulant of random of the whole population [93]. Sampling plays an important
variables x = [x1 , · · · , xk ]T is defined [70] based on the co- role in applications such as compression [27] and storage [94].
efficients of v = [v1 , · · · , vk ]T in the Talyor series expansion Similar to sampling signals in time, the HGSP sampling theory
of cumulant-gernerating function, i.e., can be developed to sample signals over the vertex domain.
We now introduce the basics of HGSP sampling theory for
K(v) = ln E{exp(jvT x)}. (43) lossless signal dimension reduction.
To reduce the size of a hypergraph signal s[M −1] , there
It is easy to see that HGSP and HOS are related in high- are two main approaches: 1) to reduce the dimension of
dimensional signal processing. They can be both represented each order; and 2) to reduce the number of orders. Since
by tensor. For example, in the multi-channel problems of [71], the reduction of order breaks the structure of hypergraph
the 3rd-order cumulant C = {Cyi ,yj ,yz (t, t1 , t2 )} of zero- and cannot always guarantee perfect recovery, we adopt the
mean signals can be represented as a multilinear array, e.g., dimension reduction of each order. To change the dimension
Cyi ,yj ,yz (t, t1 , t2 ) = E{yi (t)yj (t + t1 )yz (t + t2 )}, (44) of a certain order, we can use the n-Mode product. Since each
order of the complex signal is equivalent, the n-Mode product
which is essentially a third-order tensor. More specifically, if operators of each order are the same. Then, the sampling
there are k samples, the cumulant C can be represented as an operation of the hypergraph signal is defined as follows:
pk -element vector, which is the flattened signal tensor similar
Definition 11 (Sampling and Interpolation). Suppose that Q is
to the n-mode flattening of HGSP signals.
the dimension of each sampled order. The sampling operation
Although both HOS and HGSP are high-dimensional signal
is defined as
processing tools, they focus on complementary aspects of the
[M−1]
signals. Specifically, HGSP aims to analyze signals over the sQ = s[M−1] ×1 U ×2 U · · · ×M −1 U, (45)
14
where the sampling operator is U ∈ RQ×N to be defined later, the conditions for perfect recovery. For the original signal, we
Q×Q×...×Q
| {z } have the following property.
[M−1]
and the sampled signal is sQ ∈ R M −1 times .
The interpolation operation is defined by Lemma 1. Suppose that s ∈ RN is a K-bandlimited signal.
Then, we have
[M−1]
s[M−1] = sQ ×1 T ×2 T · · · ×M −1 T, (46) T
s = F[K] s̃[K] , (52)
where the interpolation operator is T ∈ RN ×Q to be defined T
where F[K] = [f1 , · · · , fK ] and s̃[K] ∈ RK consists of the first
later.
K elements of the original signal in the frequency domain,
As presented in Section III, the hypergraph signal and i.e., s̃.
original signal are different forms of the same data. They may
Proof: Since s is K-bandlimited, s̃i = fiT s = 0 when
have similar properties in structures. To derive the sampling
i > K. Then, according to Eq.(30), we have
theory for perfect signal recovery efficiently, we first consider
the sampling operations of the original signal. K
X N
X K
X
s = VT Vs = fi fiT s + fi fiT s = fi fiT s + 0 = F[K]
T
s̃[K] ,
Definition 12 (Sampling original signal). Suppose an original i=1 i=K+1 i=1
K-bandlimited signal s ∈ RN is to be sampled into sQ ∈ RQ , (53)
where q = {q1 , · · · , qQ } denotes the sequence of sampled
indices and qi ∈ {1, 2, · · · , N }. The sampling operator U ∈ where V = [f1 , · · · , fN ]T .
RQ×N is a linearing mappling from RN to RQ , defined by This lemma implies that the first K frequency components
( carry all the information of the original signal. Since the
1, j = qi ;
Uij = (47) hypergraph signal and the original signal share the same
0, otherwise, sampling operators, we can reach a similar conclusion for
perfect recovery as [27], [28], given in the following theorem.
and the interpolation operator T ∈ RN ×Q is a linear mapping
from RQ to RN . Then, the sampling operation is defined by Theorem 3. Define the sampling operator U ∈ RQ×N
sQ = U · s, (48) according to Uji = δ[i − qj ] where 1 ≤ qi ≤ N, i =
1, . . . , Q. By choosing Q ≥ K and the interpolation
and the interpolation operation is defined by operator T = F[K] T
Z ∈ RN ×Q with ZUF[K] T
= IK and
T
s0 = T · sQ . (49) F[K] = [f1 , · · · , fK ], we can achieve a perfect recovery, i.e.,
s = TUs for all K-bandlimited original signal s and the
Analyzing the structure of the sampling operations, we have corresponding hypergraph signal s[M −1] .
the following properties.
Proof: To prove the theorem, we show that TU is a
Theorem 2. The hypergraph signal s[M−1] shares the same projection operator and T spans the space of the first K
sampling operator U ∈ RQ×N and interpolation operator eigenvectors. From Lemma 1 and s = TsQ , we have
T ∈ RN ×Q with the original signal s.
T T
s = F[K] s̃[K] = F[K] ZsQ . (54)
Proof: We first examine one of the orders in n-Mode
product of hypergraph signal, i.e., nth-order of s[M−1] , 1 ≤ As a result, rank(ZsQ ) = rank(s̃[K] ) = K. Hence, we
n ≤ N , as conclude that K ≤ Q.
N
X Next, we show that TU is a projection by proving that
(s[M−1] ×n U)i1 ...in−1 jin+1 ...iM −1 = si1 si2 ...siM −1 Ujin . TU · TU = TU. Since we have Q ≥ K and
in =1
T
(50) ZUF[K] = IK , (55)
[M−1]
Since all elements in sQ should also be the elements of
s[M−1] after sampling, only one Ujin = 1 exists for each j We have
according to Eq. (50), i.e., only one term in the summation T
TU · TU = F[K] T
ZUF[K] ZU (56a)
exists for each j in the right part of Eq. (50). Moreover, since T
U samples over all the order, Upin = 1 and Ujin = 1 cannot = F[K] ZU = TU. (56b)
[M−1]
exist at the same time so that all the entries in sQ are Hence, TU is a projection operator. For the spanning part, the
[M−1]
also in s . Suppose q = {q1 , q2 , · · · , qQ } is the places of proof is the same as that in [27].
non-zero Ujqj ’s, we have Theorem 3 shows that a perfect recovery is possible for
[M−1] a bandlimited hypergraph signal. We now examine some
sQ (i1 , i2 , · · · , iQ ) = siq1 siq2 · · · siqQ . (51)
interesting properties of the sampled signal.
As a result, we have Uji = δ[i − qj ], which is the same as the From the previous discussion, we have s̃[K] = ZsQ , which
sampling operator for the original signal. For the interpolation has a similar form to HGFT, where Z can be treated as the
operator, the proof is similar and hence omitted. Fourier transform operator. Suppose that Q = K and Z =
Given Theorem 2, we only need to analyze the operations [z1 · · · zK ]T . We have the following first-order difference
of the original signal in the sampling theory. Next, we discuss property.
15
PK
Theorem 4. Define a new hypergraph by FK = i=1 λi · used for different signal processing tasks by selecting specific
zi ◦ · · · ◦ zi . Then, for all K-bandlimited signal s[M −1] ∈ parameters a and {αi }.
N ×N ×...×N
| {z } In addition to the general polynomial filter based on hyper-
R M times , it holds that graph signals, we provide another specific form of polynomial
[M −1] filter based on the original signals. As mentioned in Section
s[K] − FK s[K] = U(s − Fs[M −1] ). (57)
III-E, the supporting matrix Ps in Eq. (33) captures all
Proof: Let the diagonal matrix Σ[K] consist of the first the information of the frequency space. For example, the
T
K coefficients {λ1 , . . . , λK }. Since ZUF[K] = IK , we have unnormalized supporting matrix P = λmax Ps is calculated
as T
[M −1] λ1 f1
FK s[K] = Z−1 Σ[K] ŝ[K] (58a) ..
P = f1 · · · fN
.. . . (61)
= T
UF[K] Σ[K] ŝ[K] (58b) .
T
λN fN
K
X
= U[( λi · fi ◦ ... ◦ fi )(s| ◦ {z
... ◦ s}) + 0] (58c) Obviously, the hypergraph spectrum pair (λr , fr ) is an eigen-
| {z }
i=1 M times M-1 times pair of the supporting matrix P. Moreover, Theorem 1 shows
= UFs[M −1] . (58d) that the total variation of frequency component equals to a
function of P, i.e.,
[M −1]
Since s[K] = Us, it therefore holds that s[K] − FK s[K] =
1
U(s − Fs[M −1] ). TV(f r ) = ||fr − Pf r ||1 . (62)
λmax
Theorem 4 shows that the sampled signals form a new
hypergraph that preserves the information of the one-time From Eq. (62), P can be interpreted as a shifting matrix for
shifting filter over the original hypergraph. For example, the the original signal. Accordingly, we can design a polynomial
left-hand side of Eq. (57) represent the difference between the filter for the original signal based on the supporting matrix P
sampled signal and the one-time shifted version in the new whose kth-order term is defined as
hypergraph. The right-hand side of Eq. (57) is the difference
between a signal and its one-time shifted version in the s<k> = Pk s. (63)
original hypergraph, together with the sampling operator. That
is, the sampled result of the one-time shifting differences The a-th order polynomial filter is simply given as
in the original hypergraph is equal to the one-time shifting a
X
differences in the new sampled hypergraph. s0 = αk Pk s. (64)
k=1
B. Filter Desgin A polynomial filter over the original signal can be determined
Filter is an important tool in signal processing applica- with specific choices of a and α.
tions such as denoising, feature enhancement, smoothing, Let us consider some interesting properties of the polyno-
and classification. In GSP, the basic filtering is defined as mial filter for the original signal. Given the kth-order term,
s0 = FM s where FM is the representing matrix [2]. In HGSP, we have the following properties.
the basic hypergraph filtering is defined in Section III-C as
Lemma 2.
s(1) = Fs[M −1] , which is designed according to the tensor N
X
contraction. The HGSP filter is a multilinear mapping [75]. s<k> = λkr (frT s)fr . (65)
The high-dimensionality of tensors provides more flexibility r=1
in designing the HGSP filter.
1) Polynomial Filter based on Representing Tensor: Poly- Proof: Let VT = [f1 , · · · , fN ] and Σ =
nomial filter is one basic form of HGSP filters, with which diag([λ1 , · · · , λN ]). Since VT V = I, we have
signals are shifted several times over the hypergraph. An
example of polynomial filter is given as Fig. 8 in Section III-B. Pk = V T T T
{z · · · V ΣV}
| ΣVV ΣV (66a)
A k-time shifting filter is defined as k times
[M −1] = VT Σk V. (66b)
s(k) = Fs(k−1) (59a)
= F(F(...(Fs[M −1] )[M −1] )[M −1] )[M −1] . (59b) Therefore, the kth-order term is given as
| {z }
k times s<k> = VT Σk Vs (67a)
More generally, a polynomial filter is designed as N
X
a
X = λkr (frT s)fr . (67b)
0
s = αk s(k) , (60) r=1
k=1
where {αk } are the filter coefficients. Such HGSP filters From Lemma 2, we obtain the following property of the
are based on multilinear tensor contraction, which could be polynomial filter for the original signal.
16
TABLE I
C OMPRESSION R ATION OF D IFFERENT M ETHODS
(a) Variance in the same cluster (b) Average Silhouette in the hypergraph
The basic idea adopted for the classification filter design Algorithm 2 LP-HGSP Classification
is label propagation (LP), where the main steps are to first 1: Input: Dataset s consisting of labeled training data and
construct a transmission matrix and then propagate the label unlabeled test data.
based on the transmission matrix [36]. The label will converge 2: Establish a hypergraph by similarities and set unlabeled
after a sufficient number of shifting steps. Let W be the data as {0} in the signal s.
propagation matrix. Then the label could be determined by 3: Train the coefficients α of the LP-HGSP filter in Eq.
k
the distribution s0 = W s. We see that s0 is in the form (70) by minimizing the errors of signs of training data
of filtered graph signal. Recall that in Section V-B, the in s0 = Hs.
supporting matrix P has been shown to capture the properties 4: Implement LP-HGSP filter. If s0 i > 0, the ith data is
of hypergraph shifting and total variation. Here, we propose a labeled as 1; otherwise, it is labeled as −1.
HGSP classifier based on the supporting matrix P defined in 5: Output: Labels of test signals.
Eq. (61) to generate matrix
(a) Over datasets with a fixed datasize and different ratios of (b) Over datasets with different datasizes and a fixed ratio of
training data. training data.
method with the RBF kernel and the label propagation GSP could be measured by the total variation. Assume that the
(LP-GSP) method [2]. In the experiment, we model the dataset original signal is smooth. We formulate signal denoising as
as hypergraph or graph based on the distance of data. The an optimization problem. Suppose that y = s + n is a noisy
threshold of determining the existence of edges is designed signal with noise n, and s0 = h(F, y) is the denoised data
to ensure the absence of isolated nodes in the graph. For by the HGSP filter h(·). The denoising problem could be
the label propagation method, we set k = 15. The result is formulated as an optimization problem:
shown in Fig. 18(a). From the result, we see that the label norm
propagation HGSP method (LP-HGSP) is moderately better min ||s0 − y||22 + γ · ||s0 − s0 <1> ||22 , (71)
h
than LP-GSP. The graph-based methods, i.e., LP-GSP and LP- where the second term is the weighted quadratic total variation
HGSP, both perform better than SVM. The performance of of the filtered signal s0 based on the supporting matrix.
SVM appears less satisfactory, likely because the dataset is The denoising problem of Eq. (71) aims to smooth the signal
rather small. Model-based graph and hypergraph methods are based on the original noisy data y. The first term keeps the
rather robust when applied to such small datasets. To illustrate denoised signal close to the original noisy signal, whereas the
this effect more clearly, we tested the SVM and hypergraph second term tries to smooth the recovered signal. Clearly, the
performance with new configurations by the increasing dataset optimized solution of filter design is
size and the fixing ratio of training data in Fig. 18(b). In
the experiment, we first pick different sizes of data subsets s0 = h(F, y) = [I + γ(I − Ps )T (I − Ps )]−1 y, (72)
from the original zoo dataset randomly as the new datasets. PN λi
Then, with each size of the new dataset, 40% data points where Ps = i=1 λmax fi fiT describes a hypergraph Fourier
are randomly picked as the training data, and the remaining decomposition. From Eq. (72), we see that the solution is in
data points are used as the test data. We average 10000 the form of s0 = Hy for denoising, which adopts a hypergraph
times to smooth the curve. We can see from Fig. 18(b) that filter h(·) as
the performance of SVM shows significant improvement as H = [I + γ(I − Ps )T (I − Ps )]−1 . (73)
the dataset size grows larger. This comparison indicates that
SVM may require more data to achieve better performance, as The HGSP-based filter follows a similar idea to GSP-based
shown in the comparative results of Fig. 18(a). Generally, the denoising filter [33]. However, different definitions of the total
HGSP-based method exhibits better overall performance and variation and signal shifting result in different designs of
shows significant advantages with small datasets. Although HGSP vs. GSP filters. To test the performance, we compare
GSP and HGSP classifiers are both model-based, hypergraph- our method with the basic Wiener filter, Median filter, and
based ones usually perform better than graph-based ones, since GSP-based filter [33] using the image datasets of Fig. 14.
hypergraphs provide a better description of the structured data We apply different types of noises. To quantify the filter
in most applications. performance, we use the mean square error (MSE) between
each true signal and the corresponding signal after filtering.
D. Denoising The results are given in Table II. From these results, we can
see that, for each type of noise and picking optimized γ for all
Signals collected in the real world often contain noises. the methods, our HGSP-based filter out-performs other filters.
Signal denoising is thus an important application in signal
processing. Here, we design a hypergraph filter to implement
signal denoising. E. Other Potential Applications
As mentioned in Section III, the smoothness of a graph In addition to the application algorithms discussed above,
signal, which describes the variance of hypergraph signals, there could be many other potential applications for HGSP.
20
TABLE II
MSE OF F ILTERED S IGNAL
In this subsection, we suggest several potential applicable practicality of the proposed HGSP framework. All the features
datasets and systems for HGSP. of HGSP make it a powerful tool for IoT applications in the
• IoT: With the development of IoT techniques, the system future.
structures become increasingly complex, which makes Future Directions: With the development of tensor algebra
traditional graph-based tools inefficient to handle the and hypergraph spectra, more opportunities are emerging to
high-dimensional interactions. On the other hand, the explore HGSP and its applications. One interesting topic is
hypergraph-based HGSP is powerful in dealing with high- how to construct the hypergraph efficiently, where distance-
dimensional analysis in the IoT system: for example, data based and model-based methods have achieved significant
intelligence over sensor networks, where hypergraph- successes in specific areas, such as image processing [97] and
based analysis has already attracted significant attentions natural language processing [98]. Another promising direction
[96], and HGSP could be used to handle tasks like is to apply HGSP in analyzing and optimizing multi-layer net-
clustering, classification, and sampling. works. As we discussed in the introduction, hypergraph is an
• Social Network: Another promising application is the alternative model to present the multi-layer network [13], and
analysis of social network datasets. As discussed earlier, HGSP becomes a useful tool when dealing with multi-layer
a hyperedge is an efficient representation for the multi- structures. Other future directions include the development of
lateral relationship in social networks [83], [84]; HGSP fast operations such as the fast hypergraph Fourier transform,
can then be effective in analyzing multi-lateral node and applications over high-dimensional datasets [99].
interactions.
• Nature Language Processing: Furthermore, natural lan-
guage processing is an area that can benefit from HGSP. R EFERENCES
Modeling the sentence and language by hypergraphs [85],
[86], HGSP can be a tool for language classification and [1] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on
graphs,” IEEE Transactions on Signal Processing, vol. 61, no. 7, pp.
clustering tasks. 1644-1656, Apr. 2013.
Overall, due to its systematic and structural approach, HGSP [2] A. Ortega, P. Frossard, J. Kovaevi, J. M. F. Moura, and P. Vandergheynst,
“Graph signal processing: overview, challenges, and applications,” Pro-
is expected to become an important tool in handling high- ceedings of the IEEE, vol. 106, no. 5, pp. 808-828, Apr. 2018.
dimensional signal processing tasks that are traditionally ad- [3] M. Newman, D. J. Watts, and S. H. Strogatz, “Random graph models of
dressed by DSP or GSP based methods. social networks,” Proceedings of the National Academy of Sciences, vol.
99, no. 1, pp. 2566-2572, Feb. 2002.
[4] S. Barbarossa, and M. Tsitsvero, “An introduction to hypergraph sig-
VII. C ONCLUSIONS nal processing,” in Proc of Acoustics, Speech and Signal Processing
(ICASSP), Shanghai, China, Mar. 2016, pp. 6425-6429.
In this work, we proposed a novel tensor-based framework [5] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on
of Hypergraph Signal Processing (HGSP) that generalizes the graphs: frequency analysis,” IEEE Transactions on Signal Processing,
vol. 62, no. 12, pp. 3042-3054, Apr. 2014.
traditional GSP to high-dimensional hypergraphs. Our work [6] J. Shi, and J. Malik, “Normalized cuts and image segmentation, IEEE
provided important definitions in HGSP, including hyerpgraph Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no.
signals, hypergraph shifting, HGSP filters, frequency, and 8, pp. 888905, Aug. 2000.
[7] R. Wagner, V. Delouille, and R. G. Baraniuk, “Distributed wavelet de-
bandlimited signals. We presented basic HGSP concepts such noising for sensor networks, in Proceedings of the 45th IEEE Conference
as the sampling theory and filtering design. We show that on Decision and Control, San Diego, USA, Dec. 2006, pp. 373379.
hypergraph can serve as an efficient model for many complex [8] S. K. Narang, and A. Ortega, “Local two-channel critically sampled filter-
datasets. We also illustrate multiple practical applications banks on graphs,” in Proc. of 17th IEEE International Conference on
Image Processing (ICIP), Hong Kong, China, Sept. 2010, pp. 333-336.
for HGSP in signal processing and data analysis, where we [9] X. Zhu and M. Rabbat, “Approximating signals supported on graphs, in
provided numerical results to validate the advantages and the Proc. ICASSP, Kyoto, Japan, Mar. 2012, pp. 39213924.
21
[10] S. Zhang, H. Zhang, H. Li and S. Cui, “Tensor-based spectral analysis [34] D. Dua, and E. K. Taniskidou, UCI Machine Learning Repository, 2017.
of cascading failures over multilayer complex systems,” in Proceedings [Online]. Avaiable: http://archive.ics.uci.edu/ml. [Irvine, CA: University
of 56th Annual Allerton Conference on Communication, Control, and of California, School of Information and Computer Science].
Computing (Allerton), Monticello, USA, Oct. 2018, pp. 997-1004. [35] D. L. Medin, and M. Schaffer, “Context theory of classification learn-
[11] Z. Huang, C. Wang, M. Stojmenovic, and A. Nayak, “Characterization ing,” Psychological Review, vol. 85, no. 3, pp. 207-238, 1978.
of cascading failures in interdependent cyberphysical systems,” IEEE [36] X. Zhu, and Z. Ghahramani, “Learning from labeled and unlabeled data
Transactions on Computers, vol. 64, no. 8, pp. 2158-2168, Aug. 2015. with label propagation,” CMU CALD Tech Report, vol. 02, no. 107, 2002.
[12] T. Michoel, and B. Nachtergaele, “Alignment and integration of complex [37] D. Zhou, J. Huang, and B. Schlkopf, “Learning with hypergraphs: clus-
networks by hypergraph-based spectral clustering,” Physical Review E, tering, classification, and embedding,” Advances in Neural Information
vol. 86, no. 5, p. 056111, Nov. 2012. Processing Systems. pp. 1601-1608, Sept. 2007.
[13] B. Oselio, A. Kulesza, and A. Hero, “Information extraction from large [38] J. Shi and J. Malik, “Normalized cuts and image segmentation, IEEE
multi-layer social networks,” in 2015 IEEE International Conference on Trans. Pattern Anal. Mach. Intell., vol. 22, no. 8, pp. 888905, Aug. 2000.
Acoustics, Speech and Signal Processing, Brisbane, Australia, Apr. 2015, [39] E. O. Brigham and E. O. Brigham, The Fast Fourier Transform and its
pp. 5451-5455. Applications, New Jersey: Prentice Hall, 1988.
[14] M. De Domenico, A. Sole-Ribalta, E. Cozzo, M. Kivel, Y. Moreno, [40] X. Li, W. Hu, C. Shen, A. Dick, and Z. Zhang, “Context-aware hyper-
M. A. Porter, S. Gomez, and A. Arenas, “Mathematical formulation of graph construction for robust spectral clustering,” IEEE Transactions on
multilayer networks, Physical Review X, vol. 3, no. 4, p. 041022, Dec. Knowledge and Data Engineering, vol. 26, no. 10, pp. 2588-2597, Jul.
2013. 2014.
[15] G. Ghoshal, V. Zlati, G. Caldarelli, and M. E. J. Newman, “Random [41] K. Ahn, K. Lee, and C. Suh, “Hypergraph spectral clustering in the
hypergraphs and their applications,” Physical Review E, vol. 79, no. 6, p. weighted stochastic block model,” IEEE Journal of Selected Topics in
066118, Jun. 2009. Signal Processing, vol. 12, no. 5, pp. 959-974, May 2018.
[16] L. J. Grady and J. Polimeni, Discrete Calculus: Applied Analysis on [42] P. J. Rousseeuw, “Silhouettes: a graphical aid to the interpretation and
Graphs for Computational Science, USA: Springer Science and Business validation of cluster analysis,” Journal of Computational and Applied
Media, 2010. Mathematics, vol. 20, pp. 53-65, Nov. 1987.
[17] C. G. Drake, S. J. Rozzo, H. F. Hirschfeld, N. P. Smarnworawong, [43] G. Li, L. Qi, and G. Yu, “The Zeigenvalues of a symmetric tensor and
E. D. Palmer, and B. L. Kotzin, “Analysis of the New Zealand black its application to spectral hypergraph theory,” Numerical Linear Algebra
contribution to lupus-like renal disease. Multiple genes that operate in with Applications, vol. 20, no. 6, pp. 1001-1029, Mar. 2013.
a threshold manner,” The Journal of Immunology, vol. 154, no. 5, pp. [44] R. B. Bapat, Graphs and Matrices, New York: Springer, 2010.
2441-2447, Mar. 1995. [45] K. C. Chang, K. Pearson, and T. Zhang, “On eigenvalue problems of real
[18] V. I. Voloshin, Introduction to Graph and Hypergraph Theory. New symmetric tensors,” Journal of Mathematical Analysis and Applications,
York, USA: Nova Science Publishers, 2009. vol. 305, no. 1 pp. 416-422, Oct. 2008.
[19] C. Berge, and E. Minieka, Graphs and Hypergraphs. Amsterdam,
[46] D. Domenico, A. Sol-Ribalta, E. Cozzo, M. Kivel, Y. Moreno, M. A.
Nederland: North-Holland Pub. Co., 1973.
Porter, S. Gmez, and A. Arenas, “Mathematical formulation of multilayer
[20] A. Banerjee, C. Arnab, and M. Bibhash, “Spectra of general hyper-
networks, Physical Review X, vol. 3, no. 4, p. 041022, Dec. 2013.
graphs,” Linear Algebra and its Applications, vol. 518, pp. 14-30, Dec.
[47] P. Comon and B. Mourrain, “Decomposition of quantics in sums of
2016.
powers of linear forms,” Signal Processing, vol.53, no.2-3, pp. 93107,
[21] S. Kok, and P. Domingos, “Learning Markov logic network structure
Sept. 1996.
via hypergraph lifting,” in Proceedings of the 26th Annual International
Conference on Machine Learning, New York, USA, Jun. 2009, pp. 505- [48] P. Comon, G. Golub, L. H. Lim, and B. Mourrain, “Symmetric tensors
512. and symmetric tensor rank,” SIAM J. Matrix Anal. Appl., vol. 30, no. 3,
[22] N. Yadati, M. Nimishakavi, P. Yadav, A. Louis, and P. Talukdar, pp. 12541279, Sept. 2008.
“HyperGCN: hypergraph convolutional networks for semi-supervised [49] H. A. L. Kiers, “Towards a standardized notation and terminology
classification,” arXiv:1809.02589, Sept. 2018. in multiway analysis,” Journal of Chemometrics: A Journal of the
[23] A. Mithani, G. M. Preston, and J. Hein, “Rahnuma: hypergraph- Chemometrics Society, vol. 14, no. 3, pp. 105-122, Jan. 2000.
based tool for metabolic pathway prediction and network comparison,” [50] R. A. Horn, “The hadamard product,” Proc. Symp. Appl. Math, Vol. 40,
Bioinformatics, vol. 25, no. 14, pp. 1831-1832, Jul. 2009. 1990.
[24] T. G. Kolda and B. W. Bader, “Tensor decompositions and applications,” [51] P. Symeonidis, A. Nanopoulos, and Y. Manolopoulos, “Tag recommen-
SIAM Review, vol. 51, no. 3, pp. 455-500, Aug. 2009. dations based on tensor dimensionality reduction,” in Proceedings of the
[25] L. Qi, “Eigenvalues of a real supersymmetric tensor, Journal of Symbolic 2008 ACM Conference on Recommender Systems, Lousanne, Switzerland,
Computation, vol. 40, no. 6, pp. 1302-1324, Jun. 2005. Oct. 2018, pp. 43-50.
[26] A. Afshar, J. C. Ho, B. Dilkina, I. Perros, E. B. Khalil, L. Xiong, and V. [52] S. Liu, and G. Trenkler, “Hadamard, Khatri-Rao, Kronecker and other
Sunderam, “Cp-ortho: an orthogonal tensor factorization framework for matrix products,” International Journal of Information and Systems
spatio-temporal data,” in Proceedings of the 25th ACM SIGSPATIAL In- Sciences, vol. 4, no. 1, pp. 160-177, 2008.
ternational Conference on Advances in Geographic Information Systems, [53] K. J. Pearson, and T. Zhang, “On spectral hypergraph theory of the
Redondo Beach, CA, USA, Jan. 2017, p. 67. adjacency tensor,” Graphs and Combinatorics, vol. 30, no. 5, pp. 1233-
[27] S. Chen, A. Sandryhaila, and J. Kovaevi, “Sampling theory for graph 1248, Sept. 2014.
signals,” in Proc. of Acoustics, Speech and Signal Processing (ICASSP), [54] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on
Brisbane, Australia, Apr. 2015, pp. 3392-3396. graphs: graph Fourier transform,” in Proceedings of 2013 IEEE In-
[28] S. Chen, R. Varma, A. Sandryhaila, and J. Kovaevi, “Discrete signal ternational Conference on Acoustics, Speech and Signal Processing,
processing on graphs: sampling theory,” IEEE Transactions on Signal Vancouver, BC, Canada, May 2013, pp. 26-31.
Processing, vol. 63, no. 24, pp. 6510-6523, Dec. 2015. [55] M. A. Henning, and Y. Anders, “2-Colorings in k-regular k-uniform
[29] B. W. Bader, and T. G. Kolda, “MATLAB Tensor Toolbox Version 2.6,” hypergraphs,” European Journal of Combinatorics, vol. 34, no. 7, pp.
[Online]. Available: http://www.sandia.gov/∼tgkolda/TensorToolbox/. 1192-1202, Oct. 2013.
[Feb. 2015]. [56] E. Robeva, “Orthogonal decomposition of symmetric tensors,” SIAM
[30] A. Bretto, and L. Gillibert, “Hypergraph-based image representation,” Journal on Matrix Analysis and Applications, vol. 37, no. 1, pp. 86-102,
in International Workshop on Graph-Based Representations in Pattern Jan. 2016.
Recognition, Berlin, Heidelberg, Apr. 2005, pp. 1-11. [57] A. Cichocki, D. Mandic, L. Lathauwer, G. Zhou, Q. Zhao, C. Caiafa, and
[31] A. Morar, F. Moldoveanu, and E. Grller, “Image segmentation based on H. A. Phan, “Tensor decompositions for signal processing applications:
active contours without edges,” in Proc. of 2012 IEEE 8th International from two-way to multiway component analysis,” IEEE Signal Processing
Conference on Intelligent Computer Communication and Processing, Magazine, vol. 32, no. 2, pp. 145-163, Feb. 2015.
Cluj-Napoca, Romania, Sept. 2012, pp. 213-220. [58] D. Silva, and A. Pereira, ‘Tensor techniques for signal processing:
[32] K. Fukunaga, S. Yamada, H. S. Stone, and T. Kasai, “A representation algorithms for Canonical Polyadic decomposition,’ University Grenoble
of hypergraphs in the euclidean space,” IEEE Transactions on Computers, Alpes, 2016.
vol. 33, no. 4, pp. 364-367, Apr. 1984. [59] V. Nguyen, K. Abed-Meraim, and N. Linh-Trung, “Fast tensor decom-
[33] S. Chen, A. Sandryhaila, J. M. F. Moura, and J. Kovacevic, “Signal de- positions for big data processing,” in Proceedings of 2016 International
noising on graphs via graph filtering,” in Proc. of Signal and Information Conference on Advanced Technologies for Communications, Hanoi, Viet-
Processing (GlobalSIP), Atlanta, GA, USA, Dec. 2014, pp. 872-876. nam, Oct, 2016, pp. 215-221.
22
[60] T. Kolda, “Symmetric orthogonal tensor decomposition is trivial,” Human Language Technologies, Montreal, Canada, Jun. 2012, pp. 752-
arXiv:1503.01375, Mar. 2015. 761.
[61] S. Mallat, A Wavelet Tour of Signal Processing, 3rd ed. New York, NY, [86] H. Boley, “Directed recursive labelnode hypergraphs: a new
USA: Academic, 2008. representation-language,” Artificial Intelligence, vol. 9, no. 1, pp. 49-85,
[62] V. Rdl, and J. Skokan, “Regularity lemma for kuniform hypergraphs,” Aug. 1977.
Random Structures and Algorithms, vol. 25, no. 1, pp. 1-42, Jun. 2004. [87] F. Maxwell Harper and Joseph A. Konstan, “The movieLens
[63] Z. K. Zhang, and C. Liu, “A hypergraph model of social tagging datasets: history and context,” ACM Transactions on Interactive
networks,” Journal of Statistical Mechanics: Theory and Experiment, Intelligent Systems (TiiS), vol. 5, no. 19, Jan. 2016. [Online]
vol.2010, P10005, Oct. 2010. DOI=http://dx.doi.org/10.1145/2827872
[64] K. Pearson, and T. Zhang, “Eigenvalues on the adjacency tensor of [88] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on
products of hypergraphs,” Int. J. Contemp. Math. Sci, vol. 8, pp. 151- graphs: graph filters,” in Proceedings of 2013 IEEE International Con-
158, Jan. 2013. ference on Acoustics, Speech and Signal Processing, Vancouver, Canada,
[65] J. M. Mendel, “Use of higher-order statistics in signal processing and May 2013, pp. 6163-6166.
system theory: an update,” Advanced Algorithms and Architectures for [89] L. Qi, “Eigenvalues and invariants of tensors,” Journal of Mathematical
Signal Processing, vol. 975, pp. 126-145, Feb. 1988. Analysis and Applications, vol. 325, no. 2, pp. 1363-1377, Jan. 2007.
[66] J. M. Mendel, “Tutorial on higher-order statistics (spectra) in signal [90] L. H. Lim, “Singular values and eigenvalues of tensors: a variational ap-
processing and system theory: theoretical results and some applications,” proach,” in 1st IEEE International Workshop on Computational Advances
Proceedings of the IEEE, vol. 79, no. 3, pp. 278-305, Mar. 1991. in Multi-Sensor Adaptive Processing, Puerto Vallarta, Mexico, Dec. 2005,
[67] A. K. Nandi, Blind estimation using higher-order statistics. New York, pp. 129-132.
NY, USA: Springer Science and Business Media, 2013. [91] M. Ng, L. Qi, and G. Zhou, “Finding the largest eigenvalue of a
[68] M. S. Choudhury, S. L. Shah, and N. F. Thornhill, “Diagnosis of poor nonnegative tensor,” SIAM Journal on Matrix Analysis and Applications,
control-loop performance using higher-order statistics,” Automatica, vol. vol.31, no.3, pp. 1090-1099, Feb. 2009.
40, no. 10, pp. 1719-1728, Oct. 2004. [92] D. Cartwright, and B. Sturmfels, “The number of eigenvalues of a
[69] O. A. Dobre, Y. Bar-Ness, and W. Su, “Higher-order cyclic cumulants tensor,” Linear Algebra and its Applications, vol. 438, no. 2, pp. 942-
for high order modulation classification,” in Proceedings of IEEE Military 952, Jan. 2013.
Communications Conference, Boston, MA, USA, Oct. 2003, pp. 112-117. [93] M. N. Marshall, “Sampling for qualitative research,” Family practice,
[70] M. B. Priestley, Spectral Analysis and Time Series. London, UK: vol. 13, no. 6, pp. 522-526, Dec. 1996.
Academic Press, 1981. [94] G. E. Batley, and D. Gardner, “Sampling and storage of natural waters
[71] A. N. A. N. T. H. R. A. M. Swami, and J. M. Mendel, “Time and for trace metal analysis,” Water Research, vol. 11, no. 9, pp. 745-756,
lag recursive computation of cumulants from a state-space model,” IEEE 1977.
Transactions on Automatic Control, vol. 35, no. 1, pp. 4-17, Jan. 1990. [95] U. V. Luxburg, “A tutorial on spectral clustering,” Statistics and Com-
[72] M. Hein, S. Setzer, L. Jost, and S. S. Rangapuram, “The total variation puting, vol. 17, no.4, pp. 395-416, Dec. 2007.
on hypergraphs-learning on hypergraphs revisited,” in Advances in Neural [96] J. Jung, S. Chun, and K. Lee, “Hypergraph-based overlay network model
Information Processing Systems, Lake Tahoe, Nevada, Dec. 2013, pp. for the Internet of Things,” 2015 IEEE 2nd World Forum on Internet of
2427-2435. Things (WF-IoT), Milan, Italy, Dec. 2015, pp. 104-109.
[73] G. Chen, J. Zhang, F. Wang, C. Zhang, and Y. Gao, “Efficient multi-label [97] J. Yu, D. Tao, and M. Wang, “Adaptive hypergraph learning and
classification with hypergraph regularization,” in 2009 IEEE Conference its application in image classification,” IEEE Transactions on Image
on Computer Vision and Pattern Recognition, Miami, FL, USA, Jun. Processing, vol. 21, no. 7, pp. 3262-3272, Mar. 2012.
2009, pp. 1658-1665. [98] S. Rital, H. Cherifi, and S. Miguet, “Weighted adaptive neighborhood
[74] S. Kok, and P. Domingos, “Learning Markov logic network structure hypergraph partitioning for image segmentation,” in International Con-
via hypergraph lifting,” in Proceedings of the 26th Annual International ference on Pattern Recognition and Image Analysis, Berlin, Heidelberg,
Conference on Machine Learning, Montreal, Quebec, Canada, Jun. 2009, Aug. 2005, pp. 522-531.
pp. 505-512. [99] R. B. Rusu, and S. Cousins, “3d is here: Point cloud library (pcl),”
[75] V. I. Paulsen, and R. R. Smith, “Multilinear maps and tensor norms in 2011 IEEE International Conference on Robotics and Automation,
on operator systems,” Journal of Functional Analysis, vol. 73, no. 2, pp. Shanghai, China, May 2011, pp. 1-4.
258-276, Aug. 1987.
[76] S. Chen, A. Sandryhaila, J. M. F. Moura, and J. Kovaevi, “Adaptive
graph filtering: multiresolution classification on graphs,” in 2013 IEEE
Global Conference on Signal and Information Processing, Austin, TX,
USA, Feb. 2014, pp. 427-430.
[77] S. Chen, F. Cerda, P. Rizzo, J. Bielak, J. H. Garrett, and J. Kovaevi,
“Semi-supervised multiresolution classification using adaptive graph fil-
tering with application to indirect bridge structural health monitoring,”
IEEE Transactions on Signal Processing, vol. 62, no. 11, pp. 2879-2893,
Mar. 2014.
[78] S. McCanne, and M. J. Demmer, “Content-based segmentation scheme
for data compression in storage and transmission including hierarchical
segment representation,” U.S. Patent, 6667700, Dec., 2003.
[79] K. Sayood, Introduction to Data Compression, Burlington, Mas-
sachusetts, USA: Morgan Kaufmann, 2017.
[80] D. Salomon, Data Compression: the Complete Reference. New York,
NY, USA: Springer Science and Business Media, 2004.
[81] M. Yan, K. Lam, S. Han, E. Chan, Q. Chen, P. Fan, D. Chen, and M.
Nixon, “Hypergraph-based data link layer scheduling for reliable packet
delivery in wireless sensing and control networks with end-to-end delay
constraints,” Information Sciences, vol. 278, pp. 34-55, Sept. 2014.
[82] C. Adjih, S. Y. Cho, and P. Jacquet, “Near optimal broadcast with
network coding in large sensor networks,” arXiv:0708.0975, Aug. 2007.
[83] V. Zlati, G. Ghoshal, and G. Caldarelli, “Hypergraph topological quan-
tities for tagged social networks,” Physical Review E, vol. 80, no. 3, p.
036118, Sept. 2009.
[84] G. Ghoshal, V. Zlati, G. Caldarelli, and M. E. J. Newman, “Random
hypergraphs and their applications,” Physical Review E, vol. 79, no. 6, p.
066118, Jun. 2009.
[85] I. Konstas, and M. Lapata, “Unsupervised concept-to-text generation
with hypergraphs,” in Proceedings of the 2012 Conference of the North
American Chapter of the Association for Computational Linguistics: