4 - FAU MoD Mini-workshop: AI for maths and maths for AI [ID:58175]
50 von 237 angezeigt

So I thought I would present to you something I didn't have time to present yesterday.

This is not finished, there's no paper, this is work in progress, this is what keeps me

up at night.

Don't scoop me, I haven't published this anywhere.

So you see yesterday at the end of the talk we left after having been able to calculate

the GCD with a fairly good accuracy.

So we showed that Transformer can learn relatively complex arithmetic functions, more complex

than multiplication or something like that.

You need to work for it by maybe adjusting the training distribution and all that.

And you can usually understand what the model is doing just by looking at the model prediction.

So it's an interesting approach to interpretability because usually when people want to do interpretability

they look at the weights.

The problem is that even in a very small model you have hundreds of thousands if not millions

of weights.

So how do you make sense of a million of stuff, most of them random?

It's extremely difficult.

Looking at model prediction provides an avenue for explanation of a simpler nature.

But this was a fairly simple case, this was a classification problem mostly because you

just have a small number of possible outcomes.

GCD 1 to 10, you know, account for 99% of all examples or all the GCDs you can see,

you will see in your training set.

So the task of the model there is to take the inputs, classify them into a small number

of classes corresponding to single outcomes and then assigning an outcome to every class.

What if we worked on a much harder problem?

So I want to talk here about the infamous collapse problem which I will show is a regression

problem.

So it's typical in machine learning to distinguish between pattern recognition or classification

where you have just a few, you want to distinguish two kinds of or three or any kinds of elements

and regression when you want to estimate a function and admittedly regression is a much

more difficult thing.

So collapse, you probably know of the collapse sequence, you know, it's this funny mathematical

object, you take an integer, a positive integer, if it's odd you multiply it by 3 and add 1

and if it's even you divide by 2 and then you get a sequence, you know, of number of

integers that go one after the other.

So the collapse sequence, you know, there's a very famous conjecture that says that all

sequence end up with a sub loop which is 1, 4, 2, 1, 1, 4, 2, 1 and everybody, so nobody

has been able to prove that, people are advised not to work on it except if you're a Terrain

star but apart from that you shouldn't work on it because it's too difficult, Erdos said

that mathematics is not ready for this, etc.

So I'm not interested in the collapse conjecture, I'm interested in the sequence and the sequence

has an amusing properties.

Suppose you start with an odd number of the form 2 to the k m minus 1 where m is odd,

so k is the largest power of 2 that divides, is the number of zeros at the end of the binary

representation of m plus 1 or the number of ones at the end of the binary representation

of the number, right?

If you have such a number after exactly 2k collapse steps, the number 2 to the k m minus

1 is transformed into 3 to the k m minus 1 which is an even number because m is odd and

then you have a number of k prime, it depends on m obviously, down steps, okay?

So I'm defining on the collapse sequence something I call long collapse steps which are these

transforms, I start with an odd number, I multiply it by 3n, I do n to 3n plus 1 then

Presenters

Dr. François Charton Dr. François Charton

Zugänglich über

Offener Zugang

Dauer

00:25:17 Min

Aufnahmedatum

2025-06-24

Hochgeladen am

2025-06-25 07:17:24

Sprache

en-US

Date: Mon.-Tue. June 23 - 24, 2025
Event: FAU MoD Lecture & Workshop
Organized by: FAU MoD, the Research Center for Mathematics of Data at Friedrich-Alexander-Universität Erlangen-Nürnberg (Germany)
 
FAU MoD Lecture: Mon. June 23, 2025 at 16:00H
AI for maths and maths for AI
Speaker: Dr. François Charton, Meta | FAIR | École Nationale des Ponts et Chaussées
 
Mini-workshop: Tue. June 24, 2025 (AM/PM sessions)
FAU room: H11
 
AM session (09:45H to 11:30H)
• 10:00H The Turnpike Phenomenon for Optimal Control Problems under Uncertainty. Dr. Michael Schuster, FAU DCN-AvH Chair for Dynamics, Control, Machine Learning and Numerics – Alexander von Humboldt Professorship
• 10:30H AI in Mechanics Dr.-Ing. Hagen Holthusen, FAU MoD, Research Center for Mathematics of Data | Institute of Applied Mechanics
• 11:00H Contribution evaluation in Federated Learning Daniel Kuznetsov, Visiting Student at FAU DCN-AvH from ENS Paris-Saclay
 
PM session (14:15H to 16:00H)
• 14:15H AI for maths and maths for AI Dr.-Ing. François Charton, Meta | FAIR | ENPC
• 14:30H Exact sequence prediction with transformers Giovanni Fantuzzi, FAU MoD, Research Center for Mathematics of Data | FAU DCN-AvH at Friedrich-Alexander-Universität Erlangen-Nürnberg
• 15:00H Discovering the most suitable material model for cardiac tissue with constitutive neural networks Dr. Denisa Martonová, FAU MoD, Research Center for Mathematics of Data | Institute of Applied Mechanics
• 15:30H Stability of Hyperbolic Systems with Non-Symmetric Relaxation Dr. Lorenzo Liverani, FAU MoD, Research Center for Mathematics of Data | FAU DCN-AvH at Friedrich-Alexander-Universität Erlangen-Nürnberg  
 
AUDIENCE. This is a hybrid event (On-site/online) open to: Public, Students, Postdocs, Professors, Faculty, Alumni and the scientific community all around the world.
 
WHEN
• Lecture: Mon. June 23, 2025 at 16:00H (Berlin time)
• Workshop: Tue. June 24, 2025 (AM/PM sessions) at 09:45H and 14:15H (Berlin time)
 
WHERE. On-site / Online

Tags

Mathematik Erlangen mathematics AI FAU MoD FAU MoD Lecture Series FAU MoD workshop Maths FAU
Einbetten
Wordpress FAU Plugin
iFrame
Teilen