Skip to content

Commit

Permalink
Merge pull request #22 from loreloc/aaai25-and-faster
Browse files Browse the repository at this point in the history
AAAI25 and faster marginalization arxiv
  • Loading branch information
arranger1044 authored Jan 3, 2025
2 parents 02c1fbc + 0ed99d5 commit d950b2b
Show file tree
Hide file tree
Showing 4 changed files with 37 additions and 7 deletions.
7 changes: 7 additions & 0 deletions _news/aaai2025.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
---
title: "One paper accepted at AAAI 2025"
collection: news
permalink: /news/aaai-2025
date: 2024-12-10
---
Our paper on <a href="https://arxiv.org/abs/2408.11778"><b>sum of squares circuits</b></a> is accepted at <b><i>AAAI 2025</i></b>.
24 changes: 24 additions & 0 deletions _publications/loconte2024faster.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,24 @@
---
collection: publications
ref: "loconte2024faster"
permalink: "publications/loconte2024faster"
title: "On Faster Marginalization with Squared Circuits via Orthonormalization"
date: 2024-12-10 00:00
tags: circuits probml tensor-networks
image: "/images/papers/loconte2024faster/mar-squared-circuit.png"
authors: "Lorenzo Loconte, Antonio Vergari"
paperurl: "https://arxiv.org/abs/2412.07883"
pdf: "https://arxiv.org/abs/2412.07883"
venue: "arXiv 2024"
excerpt: " Inspired by canonical forms in tensor networks, we devise sufficient conditions to ensure squared circuits are already normalized and then devise a more efficient marginalization algorithm."
abstract: "Squared tensor networks (TNs) and their generalization as parameterized computational graphs -- squared circuits -- have been recently used as expressive distribution estimators in high dimensions. However, the squaring operation introduces additional complexity when marginalizing variables or computing the partition function, which hinders their usage in machine learning applications. Canonical forms of popular TNs are parameterized via unitary matrices as to simplify the computation of particular marginals, but cannot be mapped to general circuits since these might not correspond to a known TN. Inspired by TN canonical forms, we show how to parameterize squared circuits to ensure they encode already normalized distributions. We then use this parameterization to devise an algorithm to compute any marginal of squared circuits that is more efficient than a previously known one. We conclude by formally showing the proposed parameterization comes with no expressiveness loss for many circuit classes."
supplemental:
bibtex: "@misc{loconte2024faster,<br/>
title={On Faster Marginalization with Squared Circuits via Orthonormalization},<br/>
author={Lorenzo Loconte and Antonio Vergari},<br/>
year={2024},<br/>
eprint={2412.07883},<br/>
archivePrefix={arXiv},<br/>
primaryClass={cs.LG},<br/>
url={https://arxiv.org/abs/2412.07883}}"
---
13 changes: 6 additions & 7 deletions _publications/loconte2024sos.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,22 +3,21 @@ collection: publications
ref: "loconte2024sos"
permalink: "publications/loconte2024sos"
title: "Sum of Squares Circuits"
date: 2024-08-21 00:00
date: 2024-12-10 00:00
tags: circuits probml
image: "/images/papers/loconte2024sos/sos-hierarchy.png"
spotlight: "/images/papers/loconte2024sos/sos-spotlight.png"
authors: "Lorenzo Loconte, Stefan Mengel, Antonio Vergari"
paperurl: "https://arxiv.org/abs/2408.11778"
pdf: "https://arxiv.org/abs/2408.11778"
venue: "arXiv 2024"
venue: "AAAI 2025"
code: "https://github.com/april-tools/sos-npcs"
excerpt: "We theoretically prove an expressiveness limitation of deep subtractive mixture models learned by squaring circuits. To overcome this limitation, we propose sum of squares circuits and build an expressiveness hierarchy around them, allowing us to unify and separate many tractable probabilistic models."
abstract: "Designing expressive generative models that support exact and efficient inference is a core question in probabilistic ML. Probabilistic circuits (PCs) offer a framework where this tractability-vs-expressiveness trade-off can be analyzed theoretically. Recently, squared PCs encoding subtractive mixtures via negative parameters have emerged as tractable models that can be exponentially more expressive than monotonic PCs, i.e., PCs with positive parameters only. In this paper, we provide a more precise theoretical characterization of the expressiveness relationships among these models. First, we prove that squared PCs can be less expressive than monotonic ones. Second, we formalize a novel class of PCs -- sum of squares PCs -- that can be exponentially more expressive than both squared and monotonic PCs. Around sum of squares PCs, we build an expressiveness hierarchy that allows us to precisely unify and separate different tractable model classes such as Born Machines and PSD models, and other recently introduced tractable probabilistic models by using complex parameters. Finally, we empirically show the effectiveness of sum of squares circuits in performing distribution estimation."
supplemental:
bibtex: "@misc{loconte2024sumsquarescircuits,<br/>
bibtex: "@inproceedings{loconte2024sos,<br/>
title={Sum of Squares Circuits},<br/>
author={Lorenzo Loconte and Stefan Mengel and Antonio Vergari},<br/>
year={2024},<br/>
eprint={2408.11778},<br/>
archivePrefix={arXiv},<br/>
primaryClass={cs.LG}}"
year={2025},<br/>
booktitle={The 39th Annual AAAI Conference on Artificial Intelligence}}"
---
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

0 comments on commit d950b2b

Please sign in to comment.