We introduce a Transformer-based approach to computational enumerative geometry, specifically targeting the computation of
To capture the recursive and hierarchical nature inherent in these intersection numbers, we propose the Dynamic Range Activator (DRA), a new activation function that enhances the Transformer's ability to model recursive patterns and handle severe heteroscedasticity. Given precision requirements for computing
Beyond simply computing intersection numbers, we explore the enumerative "world-model" of Transformers. Our interpretability analysis reveals that the network is implicitly modeling the Virasoro constraints in a purely data-driven manner. Moreover, through abductive hypothesis testing, probing, and causal inference, we uncover evidence of an emergent internal representation of the large-genus asymptotic of
This paper has been published at ICLR 2025. You can read it at:
link
A learnable activation function, Dynamic Range Activator (DRA), designed for recursive and periodic data modalities.
pip install torch-dra
If you use this work, please cite it as:
@inproceedings{
hashemi2025can,
title={Can Transformers Do Enumerative Geometry?},
author={Baran Hashemi and Roderic Guigo Corominas and Alessandro Giacchetto},
booktitle={The Thirteenth International Conference on Learning Representations},
year={2025},
url={https://openreview.net/forum?id=4X9RpKH4Ls}
}