-
Notifications
You must be signed in to change notification settings - Fork 4
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
26b8d5d
commit 0297891
Showing
488 changed files
with
865 additions
and
752 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,55 @@ | ||
# IA | ||
|
||
## training | ||
|
||
[Yann LeCun about IA training on LinkedIn](https://www.linkedin.com/feed/update/urn:li:activity:7133567569684238336/) | ||
|
||
```text | ||
Animals and humans get very smart very quickly with vastly smaller amounts of training data than current AI systems. | ||
Current LLMs are trained on text data that would take 20,000 years for a human to read. | ||
And still, they haven't learned that if A is the same as B, then B is the same as A. | ||
Humans get a lot smarter than that with comparatively little training data. | ||
Even corvids, parrots, dogs, and octopuses get smarter than that very, very quickly, with only 2 billion neurons and a few trillion "parameters." | ||
My money is on new architectures that would learn as efficiently as animals and humans. | ||
Using more text data (synthetic or not) is a temporary stopgap made necessary by the limitations of our current approaches. | ||
The salvation is in using sensory data, e.g. video, which has higher bandwidth and more internal structure. | ||
The total amount of visual data seen by a 2 year-old is larger than the amount of data used to train LLMs, but still pretty reasonable. | ||
2 years = 2x365x12x3600 or roughly 32 million seconds. | ||
We have 2 million optical nerve fibers, carrying roughly ten bytes per second each. | ||
That's a total of 6E14 bytes. The volume of data for LLM training is typically 1E13 tokens, which is about 2E13 bytes. | ||
It's a factor of 30. | ||
Importantly, there is more to learn from video than from text because it is more redundant. | ||
It tells you a lot about the structure of the world. | ||
``` | ||
|
||
TLDR : Next gen IA needs to use video instead of text. | ||
|
||
To compare, see [this Jean-Baptiste Kempf (VLC) interview about how video works](https://www.youtube.com/watch?v=Kv4FzAdxclA). | ||
|
||
- an image is an array of pixel, each pixel is a color | ||
- a video is a collection of images (something between 24 to 60 images per second) | ||
- CODEC = compression decompression algorithm to send video. | ||
- Video pixel by pixel is around 10 to 40 Gb/s | ||
- the goal of CODEC is to divide 100, 200, ... 1K the bandwith used. | ||
- dividing bandwith is destroying information | ||
- the tech behind is based on how the human eyes behave, some colors are better seen then others, so we can delete some colors without downgrading the image seen. | ||
|
||
Each CODEC behave the same way, they delete data not seen by eyes, and they seek data blocks that are redundant image by image or between images. | ||
|
||
```text | ||
MPEG-1 (1993) ---> MPEG-2 (1995) = DVD ---> DIVX (1999) (=MPEG-4) ---> H.264 (2003) ---> HEVC (2013) ---> VP9 (2013) | ||
``` | ||
|
||
- H.264 is the most common CODEC used in the world, around 80% of usage. | ||
- HEVC is crippled by royalties, it remains unused on the web instead of television, around 5%. | ||
- VP9 created by Google, royalty free, opensource, Youtube and Facebook uses it. | ||
- AV1 then AV2 created by the Open Media Alliance initiated by Google. | ||
- AV1 is implemented by [Dav1d](https://github.com/videolan/dav1d), a VLC project, around 210K assembly LoC + 30K C LoC. This impl is widely used by GAFAM. | ||
|
||
## misc | ||
|
||
[Guide ChatGPT pour développeurs](https://gen-ai.fr/outils/generation-code/chatgpt-pour-developpeurs/) |
Large diffs are not rendered by default.
Oops, something went wrong.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Oops, something went wrong.
2 changes: 1 addition & 1 deletion
2
docs/assets/js/100.44086804.js → docs/assets/js/101.c2c7bbe7.js
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Oops, something went wrong.
Oops, something went wrong.