NeurIPS 2023
Trends in AI
Nikolaos Vasiloglou
VP of Research-ML
1
1
A few words
2
2
In a nutshell (1)
3
3
In a nutshell (2)
4
4
AI futurism
5
5
Other NeurIPS reviews
6
6
Understanding LLMs
7
7
Optimizing attention
8
8
Test of time (Word2vec)
Distributed Representations of Words and Phrases and their Compositionality
Tomas Mikolov* · Ilya Sutskever* · Kai Chen · Greg Corrado · Jeff Dean
9
*Absent in the ceremony
9
What we have learned so far
10
10
An ontology of word embeddings (Before GPT)
11
Something is missing
11
The revolution of context
12
12
The Language Model is the Embedding
10 years later
“Efficiently Tuned Parameters are Task Embedding,
“Task Arithmetic in the Tangent Space: Improved Editing of Pre-Trained Models,
“TASK2VEC: Task Embedding for Meta-Learning”
13
Jump here if you want more!
13
The unreasonable behavior of LLMs
What is going on as we scale from thousand to million to billion to trillion parameters
14
14
Do LLMs have emergent properties?
15
15
Are we running
out of oil?
Sustainability analysis of data
16
16
Managing data repetitions and parameter budget
17
17
Training data trend
18
18
Training LLMs on data budget
19
19
LLMs as a World Model
20
20
The rise of simulators
21
21
Train LLMs with Simulation data!
22
22
LLMs not good enough for Common Sense
23
23
Literature
24
24
LLM not good enough for Social Tasks
25
25
Literature
26
26
Literature
27
27
A deep dive into the physics simulators
Recent Advances
28
28
MuJoCo: A physics engine for model-based control
Multi-Joint dynamics with Contact
29
29
iGibson 2.0
Object-Centric Simulation for Robot Learning of Everyday Household Tasks
30
30
iGibson 2.0
31
31
Habitat 2.0
Training Home Assistants to Rearrange their Habitat
32
32
AI2-THOR
An Interactive 3D Environment for Visual AI
33
33
AI2-THOR
34
34
A big library
35
35
Habitat 2.0
36
36
ThreeDWorld
A Platform for Interactive Multi-Modal Physical Simulation
37
37
ThreeDWorld
38
38
ScenseScript
39
39
Imitating Shortest Paths in Simulation
Effective Navigation and Manipulation in the Real World
40
40
LEARNING RIGID DYNAMICS WITH FACE INTERACTION GRAPH NETWORKS
41
41
Modeling faces not just nodes
42
42
Graph Networks as Learnable Physics Engines
43
43
Follow Kelsey Allen
44
44
Follow Alvaro Sanchez Gonzalez
45
45
Unisim: Probably the LLM equivalent for simulators
Learning Interactive Real-World Simulators
What is the difference with Sora?
46
46
Unisim
47
47
Other resources for Physics Simulation @NeurIPS 2023
48
48
Climate simulations with AI
49
49
A taxonomy of simulation research papers
50
Molecular/DNA/Proteins
Newtonian based
Rigid body
particles
Astronomical Scale
Human scale
Thermal physics
GNN based
Neural ODE/PDE
Generative based
Logic Rules
Videos Images
Quantum particles
Fluid Dynamics
Other simulations
50
Other NeurIPS 2023 resources relevant to physics
AI for Science: from Theory to Practice
51
51
Other NeurIPS 2023 resources relevant to physics
Machine Learning and the Physical Sciences
52
52
19 Main track + 11 Benchmark papers (simulation)
Before 2023
53
53
Nick’s favorite Simulation framework
54
54
Back to Reasoning with simulators
55
55
Probabilistic Programs as the language of simulation
56
56
Detailed info
57
57
Language models as goal/reward
58
58
Guiding simulators with a structured language
59
59
Limitations
60
60
Multi Modal Theory of Mind
61
61
Language Models for social reasoning
62
62
Games as simulators
63
63
64
64
65
65
66
66
67
67
Social Simulator
68
68
Imagining and Verbalizing
69
69
70
70
An example
71
71
Takeaways
72
72
Distributed LLMs
Composing Big models from smaller ones
73
73
The LLM as a giant vector
74
74
The LLM as a giant vector
75
75
Would you ever write your product as a huge C++ file?
Let’s see what the software 3.0 might look like
76
76
Software 3.0
77
77
Building a bigger Language Model
78
78
Distribution of Heterogeneous LLM
79
79
Collaborating on LLM training (Git)
80
80
A taxonomy of LLM merging
81
81
Incremental maintenance of LLMs
82
82
Building an LLM from scratch
83
83
Democratizing LLM building
84
84
All the necessary plumbing for building LLMs
Infrastructure is not easy
85
85
Hardware cannot keep up with model growth
86
86
Startups will be able to afford building GPT-3 in 2 years
87
87
Colossal-AI contribution
88
88
It takes a village to build one of them
89
burns
89
Colossal-AI offers a framework for building a LLM
90
90
Looks like there is a lot of demand!
91
91
Congratulations!
You build your first really Large LM
Can you tune, maintain, grow it without redoing everything from scratch?
92
92
Attribution as a dimension for optimizing LLM costs
93
93
Pillar I: Data attribution
94
94
Pillar II: Model Attribution
95
95
Pillar III: Algorithm Attribution
96
96
DataInf
97
97
Tiny LMs might be the solution
98
98
Mathematics/Reasoning and LLMs
99
99
The most exciting topic
100
100
Mathematical Reasoning
101
101
Deepmind’s recent buzz
102
102
What is formal theory proving?
103
103
Code vs. math symbols
104
104
LLMs are doing ok on high school level
105
105
LLema vs. Minerva
106
106
Informal vs. Formal Mathematical Reasoning
107
107
Checking Mathematical Proofs is Hard for Humans
108
108
Proof Assistants (Interactive Theorem Provers)
109
109
Examples of Proof Assistants
110
110
Generating Proof Steps (Tactics)
111
111
Searching for Proofs
112
112
Best First Search
113
113
Hyper Tree Proof Search
114
114
Is Proof Search Really Necessary?
115
115
Premise Selection
116
116
Magnushammer
117
117
Reprover: Retrieval-Augmented Prover
118
118
Premise Retrieval Improves Theorem Proving
119
119
LeanDojo
120
120
From Informal to Formal Proofs
121
121
The ecosystem
122
122
The process
123
123
An example
124
124
Theorem provers for code verification
With the help of LLMs
125
125
How can we verify code produced by LLMs?
126
126
Theorem Proving for Verified Code Generation
127
127
Formal Software Verification
128
128
Software verification on the wild
129
129
Current tools
130
130
Proof Synthesis SoTA
131
131
Proofster
132
132
The maintenance problem
133
133
How about an LLM?
134
134
The Clover Paradigm
135
135
Theorem Proving and LLMs: Takeaways
136
136
MATH-AI: The 3rd Workshop on Mathematical Reasoning and AI
137
137
138
138
Generating Reasoning data with LLMs for finetuning LLMs
139
139
Analogical prompting
140
140
A math example (1)
141
141
A math example (2)
142
142
Benchmarks
143
143
More Benchmarks
144
144
Discussion: abstraction is key in analogical reasoning
145
145
Takeaways
146
146
Literature
147
147
The trend continues
148
148
A programming Language for Transformers RASP-L
149
149
The original Language RASP
150
150
RASP-L extension
151
151
The paradox of Learning to reason from data
152
152
What can BERT learn?
153
153
BERT prefers Statistical to Logical Thinking
154
154
How should we interpret this phenomenon?
155
155
Injecting Logic to GenAI models
156
156
More examples
157
157
Literature
158
158
Injecting Logic in the transformer
159
159
Symbol processing required for implication rule
160
160
Symbol processing required for implication rule
161
161
Symbol processing required for implication rule
162
162
Transformer Production Framework
163
163
Symbol processing required for implication rule
164
164
What we have learned
165
165
Biases for a new generation of deep-reasoning LMs
166
166
LLMs for Tabular Data
167
167
Invited talk: Advances in In-Context Learning for Tabular Datasets
Invited talk: Next-Generation Data Management with Large Language Models
168
168
Tables are everywhere
169
169
Goals of the TRL workshop
170
170
Last year’s best paper
171
171
Very Interesting Results
172
172
TabPFN 1 year ago
173
173
TabPFN 2.0
174
174
GPT-4 as Data Science Assistant
175
175
An example
176
176
Benchmarks
177
177
Feature Engineering
178
178
Takeaways from CAAFE
179
179
Can LLMs learn how to do Gradient Descent?
180
180
In-Context Learning
181
181
A surprising Experiment
182
182
Literature
183
183
Prior-Fitted Networks (PFNs) Visualized
184
184
Quantitative result (87 numerical datasets without missing values)
185
185
Performance with many objectives
186
186
MotherNet
187
187
Generate a model instead of predictions
188
188
Architecture
189
189
Literature
190
190
Conclusions
191
191
Takeaways on TabPFN
192
192
The LLM version of Kumo
193
193
Another approach to the same problem
194
194
Text to SQL
What we have learned so far
195
195
Lessons learned from Natural Language to SQL
196
196
How Documentation improves GPT’s Text-to-SQL
197
197
Another Text-To-SQL paper
198
198
Benchmarks
199
199
New frontiers in Graph Learning
200
200
Foundational Models for graphs and relational data
201
201
The new Web
202
202
RAG to the rescue
203
203
Application Development with LLMs
204
204
Elementary and advanced
205
205
Wrapping up
206
206
Where do we go next?
207
207
208
208