kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
NeuralSolver: Learning Algorithms For Consistent and Efficient Extrapolation Across General Tasks
INESC-ID, Instituto Superior Técnico, Universidade de Lisboa, INESC-ID, Instituto Superior Técnico, Universidade de Lisboa.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems.ORCID iD: 0000-0002-5761-4105
INESC-ID, Instituto Superior Técnico, Universidade de Lisboa, INESC-ID, Instituto Superior Técnico, Universidade de Lisboa.
2024 (English)In: Advances in Neural Information Processing Systems 37 - 38th Conference on Neural Information Processing Systems, NeurIPS 2024, Neural information processing systems foundation , 2024, Vol. 37Conference paper, Published paper (Refereed)
Abstract [en]

We contribute NeuralSolver, a novel recurrent solver that can efficiently and consistently extrapolate, i.e., learn algorithms from smaller problems (in terms of observation size) and execute those algorithms in large problems. Contrary to previous recurrent solvers, NeuralSolver can be naturally applied in both same-size problems, where the input and output sizes are the same, and in different-size problems, where the size of the input and output differ. To allow for this versatility, we design NeuralSolver with three main components: a recurrent module, that iteratively processes input information at different scales, a processing module, responsible for aggregating the previously processed information, and a curriculum-based training scheme, that improves the extrapolation performance of the method. To evaluate our method we introduce a set of novel different-size tasks and we show that NeuralSolver consistently outperforms the prior state-of-the-art recurrent solvers in extrapolating to larger problems, considering smaller training problems and requiring less parameters than other approaches. Code available at https://github.com/esteveste/NeuralSolver.

Place, publisher, year, edition, pages
Neural information processing systems foundation , 2024. Vol. 37
National Category
Computational Mathematics Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-361954Scopus ID: 2-s2.0-105000478336OAI: oai:DiVA.org:kth-361954DiVA, id: diva2:1949627
Conference
38th Conference on Neural Information Processing Systems, NeurIPS 2024, Vancouver, Canada, Dec 9 2024 - Dec 15 2024
Note

QC 20250409

Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-09Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Vasco, Miguel

Search in DiVA

By author/editor
Vasco, Miguel
By organisation
Intelligent systems
Computational MathematicsComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 19 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf