Skip to content

Commit 9f5a7aa

Browse files
committed
updated publications and news
1 parent 835dfa8 commit 9f5a7aa

2 files changed

Lines changed: 8 additions & 21 deletions

File tree

_data/news.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,5 @@
1+
- date: 2023/08
2+
text: Our paper <a href="https://jmlr.org/papers/v24/22-1154.html">Sparse Graph Learning from Spatiotemporal Time Series</a> (Cini et al.) has been published in <strong>JMLR</strong>!
13
- date: 2023/05
24
text: 'We uploaded <strong>8 new preprints</strong> about our latest research! Check them out:
35
learning graph structures from data [<a href="https://arxiv.org/abs/2305.19183">1</a>,<a href="https://arxiv.org/abs/2304.05099">2</a>,<a href="https://arxiv.org/abs/2205.13492">3</a>],

_data/publications.yaml

Lines changed: 6 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -247,35 +247,20 @@
247247
the practical value of the test on both synthetic and real-world problems, and
248248
show how the test can be employed to assess the quality of spatio-temporal forecasting
249249
models by analyzing the prediction residuals appended to the graphs stream.
250-
- title: Sparse Graph Learning for Spatiotemporal Time Series
250+
- title: Sparse Graph Learning from Spatiotemporal Time Series
251251
links:
252-
paper: https://arxiv.org/abs/2205.13492
253-
venue: Preprint
254-
year: 2022
252+
paper: https://jmlr.org/papers/v24/22-1154.html
253+
venue: Journal of Machine Learning Research
254+
year: 2023
255255
authors:
256256
- id:acini
257257
- id:dzambon
258258
- id:calippi
259259
keywords:
260260
- spatiotemporal graphs
261261
- graph learning
262-
abstract: Outstanding achievements of graph neural networks for spatiotemporal time
263-
series prediction show that relational constraints introduce a positive inductive
264-
bias into neural forecasting architectures. Often, however, the relational information
265-
characterizing the underlying data generating process is unavailable; the practitioner
266-
is then left with the problem of inferring from data which relational graph to
267-
use in the subsequent processing stages. We propose novel, principled -- yet practical
268-
-- probabilistic methods that learn the relational dependencies by modeling distributions
269-
over graphs while maximizing, at the same time, end-to-end the forecasting accuracy.
270-
Our novel graph learning approach, based on consolidated variance reduction techniques
271-
for Monte Carlo score-based gradient estimation, is theoretically grounded and
272-
effective. We show that tailoring the gradient estimators to the graph learning
273-
problem allows us also for achieving state-of-the-art forecasting performance
274-
while controlling, at the same time, both the sparsity of the learned graph and
275-
the computational burden. We empirically assess the effectiveness of the proposed
276-
method on synthetic and real-world benchmarks, showing that the proposed solution
277-
can be used as a stand-alone graph identification procedure as well as a learned
278-
component of an end-to-end forecasting architecture.
262+
- forecasting
263+
abstract: Outstanding achievements of graph neural networks for spatiotemporal time series analysis show that relational constraints introduce an effective inductive bias into neural forecasting architectures. Often, however, the relational information characterizing the underlying data-generating process is unavailable and the practitioner is left with the problem of inferring from data which relational graph to use in the subsequent processing stages. We propose novel, principled - yet practical - probabilistic score-based methods that learn the relational dependencies as distributions over graphs while maximizing end-to-end the performance at task. The proposed graph learning framework is based on consolidated variance reduction techniques for Monte Carlo score-based gradient estimation, is theoretically grounded, and, as we show, effective in practice. In this paper, we focus on the time series forecasting problem and show that, by tailoring the gradient estimators to the graph learning problem, we are able to achieve state-of-the-art performance while controlling the sparsity of the learned graph and the computational scalability. We empirically assess the effectiveness of the proposed method on synthetic and real-world benchmarks, showing that the proposed solution can be used as a stand-alone graph identification procedure as well as a graph learning component of an end-to-end forecasting architecture.
279264
- title: Deep learning for graphs (special session at ESANN 2022)
280265
links:
281266
paper: https://doi.org/10.14428/esann/2022.ES2022-7

0 commit comments

Comments
 (0)