Computational complexity results in Neural Nets

January 17, 2025 — February 3, 2025

compsci
machine learning
making things
neural nets
Figure 1

We can build automata from neural nets. And they seem to do weird things, like learn languages, in a predictable way, which is wildly at odds with our traditional understanding of the difficulty of the task (Paging Doctor Chomsky). How can we analyse NNs in terms of computational complexity? What are the useful results in this domain?

Related: grammatical inference, memory machines, overparameterization, NN compression, learning automata, NN at scale, explainability

1 Computational cost of explaining complex reasoning

Intuitively, something might be, should be difficult, computationally about explaining the reasoning of a large neural network. Surely simpler things cannot explain more complex things?

As an extreme example, we might run into incompleteness results. (Prove the consistency of the following system of axioms…) I can believe such extreme esoterica would not be relevant in practice, but could problems close to those arise in practice? TBC.

2 Incoming

3 References

Blazek, and Lin. 2020. A Neural Network Model of Perception and Reasoning.” arXiv:2002.11319 [Cs, q-Bio].
———. 2021. Explainable Neural Networks That Simulate Reasoning.” Nature Computational Science.
Blazek, Venkatesh, and Lin. 2021. Deep Distilling: Automated Code Generation Using Explainable Deep Learning.” arXiv:2111.08275 [Cs].
Bölcskei, Grohs, Kutyniok, et al. 2019. Optimal Approximation with Sparsely Connected Deep Neural Networks.” SIAM Journal on Mathematics of Data Science.
Bubeck, Chandrasekaran, Eldan, et al. 2023. Sparks of Artificial General Intelligence: Early Experiments with GPT-4.”
Clark, Tafjord, and Richardson. 2020. Transformers as Soft Reasoners over Language.” In IJCAI 2020.
Dziri, Lu, Sclar, et al. 2023. Faith and Fate: Limits of Transformers on Compositionality.”
Frankle, and Carbin. 2019. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks.” arXiv:1803.03635 [Cs].
Garcez, and Lamb. 2020. Neurosymbolic AI: The 3rd Wave.”
Kleinberg, and Mullainathan. 2024. Language Generation in the Limit.”
Minsky, and Papert. 1972. Perceptrons: an introduction to computational geometry.
Olazaran. 1993. A Sociological History of the Neural Network Controversy.” In Advances in Computers.
Peng, Narayanan, and Papadimitriou. 2024. On Limitations of the Transformer Architecture.”
Pereira. 2002. Formal Grammar and Information Theory: Together Again? In Current Issues in Linguistic Theory.
Russell, and Norvig. 2021. Artificial Intelligence: A Modern Approach. Pearson Series in Artificial Intelligence.
Shi, Feng, and ZhifanZhu. 2016. Functional Hashing for Compressing Neural Networks.” arXiv:1605.06560 [Cs].
van Gelder, Wortsman, and Ehsani. 2020. Deconstructing the Structure of Sparse Neural Networks.” In.
Wu, Tan, Wang, et al. 2024. Beyond Language Models: Byte Models Are Digital World Simulators.”
Zhang, Backurs, Bubeck, et al. 2022. Unveiling Transformers with LEGO: A Synthetic Reasoning Task.”