Results 111 to 120 of about 121,870 (266)
Oversampled Graph Laplacian Matrix For Graph Signals
Publication in the conference proceedings of EUSIPCO, Lisbon, Portugal ...
Sakiyama, Akie, Tanaka, Yuichi
openaire +1 more source
Laplacian matrix and distance in trees
Laplacian matrix and distance in trees.
Vukičević, Damir, Gutman, Ivan
openaire +1 more source
Abstract Graph neural networks (GNNs) have revolutionised the processing of information by facilitating the transmission of messages between graph nodes. Graph neural networks operate on graph‐structured data, which makes them suitable for a wide variety of computer vision problems, such as link prediction, node classification, and graph classification.
Amit Sharma +4 more
wiley +1 more source
Taylor Expansion Proof of the Matrix Tree Theorem - Part I
The Matrix-Tree Theorem states that the number of spanning trees of a graph is given by the absolute value of any cofactor of the Laplacian matrix of the graph.
Zernik, Amitai
core
Enhancing generalized spectral clustering with embedding Laplacian graph regularization
Abstract An enhanced generalised spectral clustering framework that addresses the limitations of existing methods by incorporating the Laplacian graph and group effect into a regularisation term is presented. By doing so, the framework significantly enhances discrimination power and proves highly effective in handling noisy data.
Hengmin Zhang +5 more
wiley +1 more source
Boosted unsupervised feature selection for tumor gene expression profiles
Abstract In an unsupervised scenario, it is challenging but essential to eliminate noise and redundant features for tumour gene expression profiles. However, the current unsupervised feature selection methods treat all samples equally, which tend to learn discriminative features from simple samples.
Yifan Shi +5 more
wiley +1 more source
ABSTRACT Accurately predicting line loss rates is crucial for effective management in distribution networks, particularly for short‐term multihorizon forecasts ranging from 1 hour to 1 week. In this study, we propose attention‐GCN–LSTM, a novel method that integrates graph convolutional networks (GCN), long short‐term memory (LSTM) and a three‐level ...
Jie Liu +4 more
wiley +1 more source
Robust Partial Multi‐Label Learning Under Dual Noise via Joint Subspace Learning
ABSTRACT Partial Multi‐label Learning (PML) deals with the ambiguity where each instance is annotated with a set of candidate labels, and only a subset of which is valid. While existing PML methods focus primarily on label disambiguation, they often rely on the assumption of a clean feature space.
Yuanjian Zhang +4 more
wiley +1 more source
What is a proper graph Laplacian? An operator-theoretic framework for graph diffusion
We introduce an operator-theoretic definition of a proper graph Laplacian as any matrix associated with a given graph that can be expressed as the composition of a divergence and a gradient operator, with the gradient acting between graph-related spaces ...
Estrada Ernesto
doaj +1 more source
Abstract We estimate the price impact of very nearby concurrently listed properties in the Sydney housing market and assess their competition effects. We apply a hedonic model with spatiotemporal effects regularized via a graph Laplacian prior at the month‐by‐SA2 regional level to seven SA4 subregions of metropolitan Sydney. The model structure enables
Willem P. Sijp, Mengheng Li
wiley +1 more source

