Amit Mate 2nd
Founder and CEO | Multi-AI, Multi-stream Connected-Intelligent Applications
3h
Very interesting to see this precise articulation of why vision is central to unlocking more intelligence on machines. GMAC Intelligence remains committed to enabling Intelligent-Digital-Assistants that can see, hear, understand, talk, move-around and operate at the fastest possible speed and lowest possible energy. #edgeai #artificialintelligence #visionai #llm #northstar
Yann LeCun 3rd
VP & Chief AI Scientist at Meta
1d Edited
Follow
* Language is low bandwidth: less than 12 bytes/second. A person can read 270 words/minutes, or 4.5 words/second, which is 12 bytes/s (assuming 2 bytes per token and 0.75 words per token). A modern LLM is typically trained with 1x10^13 two-byte tokens, which is 2x10^13 bytes. This would take about 100,000 years for a person to read (at 12 hours a day). * Vision is much higher bandwidth: about 20MB/s. Each of the two optical nerves has 1 million nerve fibers, each carrying about 10 bytes per second. A 4 year-old child has been awake a total 16,000 hours, which translates into 1x10^15 bytes. In other words: - The data bandwidth of visual perception is roughly 1.6 million times higher than the data bandwidth of written (or spoken) language. - In a mere 4 years, a child has seen 50 times more data than the biggest LLMs trained on all the text publicly available on the internet. This tells us three things: 1. Yes, text is redundant, and visual signals in the optical nerves are even more redundant (despite being 100x compressed versions of the photoreceptor outputs in the retina). But redundancy in data is *precisely* what we need for Self-Supervised Learning to capture the structure of the data. The more redundancy, the better for SSL. 2. Most of human knowledge (and almost all of animal knowledge) comes from our sensory experience of the physical world. Language is the icing on the cake. We need the cake to support the icing. 3. There is *absolutely no way in hell* we will ever reach human-level AI without getting machines to learn from high-bandwidth sensory inputs, such as vision. Yes, humans can get smart without vision, even pretty smart without vision and audition. But not without touch. Touch is pretty high bandwidth, too.
- Forums
- ASX - By Stock
- BRN
- 2024 BrainChip Discussion
2024 BrainChip Discussion, page-3477
-
- There are more pages in this discussion • 7,532 more messages in this thread...
You’re viewing a single post only. To view the entire thread just sign in or Join Now (FREE)
Featured News
Add BRN (ASX) to my watchlist
(20min delay)
|
|||||
Last
27.0¢ |
Change
0.035(14.9%) |
Mkt cap ! $532.5M |
Open | High | Low | Value | Volume |
24.5¢ | 27.5¢ | 24.5¢ | $5.262M | 20.02M |
Buyers (Bids)
No. | Vol. | Price($) |
---|---|---|
11 | 627278 | 27.0¢ |
Sellers (Offers)
Price($) | Vol. | No. |
---|---|---|
27.5¢ | 1931568 | 51 |
View Market Depth
No. | Vol. | Price($) |
---|---|---|
4 | 374692 | 0.270 |
13 | 1445743 | 0.265 |
14 | 601071 | 0.260 |
25 | 1008114 | 0.255 |
32 | 1128854 | 0.250 |
Price($) | Vol. | No. |
---|---|---|
0.275 | 1825711 | 47 |
0.280 | 1274792 | 38 |
0.285 | 628151 | 21 |
0.290 | 964686 | 20 |
0.295 | 422758 | 16 |
Last trade - 14.05pm 06/11/2024 (20 minute delay) ? |
Featured News
BRN (ASX) Chart |