Skip to content

Commit 7cc7ec1

Browse files
committed
begin
1 parent 74c1a8f commit 7cc7ec1

File tree

1 file changed

+38
-3
lines changed

1 file changed

+38
-3
lines changed

docs/source/intro/reading-list.rst

Lines changed: 38 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,42 @@
11
Reading list
22
=============
33

4-
.. admonition:: Warning
5-
:class: warning
4+
There are two main academic papers for understanding Modula. The first is called *"Scalable optimization in the modular norm"*. In this paper, we construct a recursive procedure for assigning a norm to the weight space of general neural architectures. Neural networks are automatically Lipschitz and (when possible) Lipschitz smooth in this norm with respect to their weights. The construction also provides means to track input-output Lipschitz properties. The paper is available here:
65

7-
This page is still under construction.
6+
| 📘 `Scalable optimization in the modular norm <https://arxiv.org/abs/2405.14813>`_
7+
| Tim Large, Yang Liu, Minyoung Huh, Hyojin Bahng, Phillip Isola & Jeremy Bernstein
8+
| NeurIPS 2024
9+
10+
The second paper builds on the first and is called *"Modular duality in deep learning"*. In this paper, we take the modular norm and use it to derive optimizers via a procedure called "modular dualization". Modular dualization chooses a weight update :math:`\Delta w` to minimize the linearization of the loss :math:`\mathcal{L}(w)` subject to a constraint on the modular norm :math:`\|\Delta w\|_{M}` of the weight update. In symbols, we solve:
11+
12+
.. math::
13+
14+
\Delta w = \operatorname{arg min}_{\Delta w : \|\Delta w\|_{M} \leq \eta} \;\langle \Delta w, \nabla \mathcal{L}(w) \rangle,
15+
16+
where :math:`\eta` sets the learning rate. Due to the structure of the modular norm, this duality procedure can be solved recursively leveraging the modular structure of the neural architecture. This procedure leads to modular optimization algorithms, where different layer types can have different optimization rules depending on which norm is assigned to that layer. The paper is available here:
17+
18+
| 📗 `Modular duality in deep learning <https://arxiv.org/abs/2410.21265>`_
19+
| Jeremy Bernstein & Laker Newhouse
20+
| arXiv 2024
21+
22+
23+
There are many other papers by myself and other authors that I feel contain important ideas on this topic. Here are some of them:
24+
25+
Optimization
26+
-------------
27+
28+
- `Preconditioned spectral descent for deep learning <https://papers.nips.cc/paper_files/paper/2015/hash/f50a6c02a3fc5a3a5d4d9391f05f3efc-Abstract.html>`_
29+
- `The duality structure gradient descent algorithm: analysis and applications to neural networks <https://arxiv.org/abs/1708.00523>`_
30+
- `On the distance between two neural networks and the stability of learning <https://arxiv.org/abs/2002.03432>`_
31+
- `Automatic gradient descent: Deep learning without hyperparameters <https://arxiv.org/abs/2304.05187>`_
32+
- `A spectral condition for feature learning <https://arxiv.org/abs/2310.17813>`_
33+
- `Universal majorization-minimization algorithms <https://arxiv.org/abs/2308.00190>`_
34+
- `Old optimizer, new norm: An anthology <https://arxiv.org/abs/2409.20325>`_
35+
- `Muon: An optimizer for hidden layers in neural networks <https://kellerjordan.github.io/posts/muon/>`_
36+
37+
Generalization
38+
---------------
39+
40+
- `Spectrally-normalized margin bounds for neural networks <https://arxiv.org/abs/1706.08498>`_
41+
- `A PAC-Bayesian approach to spectrally-normalized margin bounds for neural networks <https://arxiv.org/abs/1707.09564>`_
42+
- `Investigating generalization by controlling normalized margin <https://arxiv.org/abs/2205.03940>`_

0 commit comments

Comments
 (0)