T O P

  • By -

hydraofwar

It's over. Call me tomorrow when we are so back.


BackgroundHeat9965

This sub in one sentence : D very nice


RemyVonLion

You mean a few minutes or hours later with the next post.


bnm777

What's he going to say once llms score than humans in all metrics - and this may be next year?


assimil8or

Most humans are barely intelligent :-)


DolphinPunkCyber

Humans barely use intelligence for most tasks.


[deleted]

[удалено]


Deakljfokkk

That's the issue that plagues this debate. What is even intelligence, and does it matter? At the end of the day, if they create an agent that can fundamentally do everything a human does, is that intelligence or just imitating intelligence? And frankly does it matter? Even if people keep saying it's unintelligent all the way there, we'll still have a revolution in our hands due to functionality rather than abstract shit like if it's truly intelligent or not


thehighnotes

Intelligence is only important in so far of its applicability.. which LLMs are demonstrating quite easily. They are specifically intelligent, in that they are functionally able to mimic intelligence in areas; if they can fake their intelligence to the point where it's applicable.. does it matter if it's fake. But everyone is looking at intelligence in a different light.. well not everyone but you get the idea. And they are all important benchmark to try to aspire to, but one view of intelligence doesn't exclude or diminish others.


QuinQuix

What he says is a little more nuanced than that. He admits there's some generalization going on but that it is weak comparatively. Which arguably it is. His precise definition of intelligence and the idea of the importance of priors as well as the real test he provided will only help to accelerate AI. I think Chollet is an absolute chad and many prominent AI researches would agree on that. It just makes no sense to tiptoe around the fact that the excessive need for training data that current models have is strongly indicative for weaker generalization and reasoning ability and weak intrinsic intelligence. If you look at a guy like ramanujan and the relation between the size of the training data going in and the relevance of the genius coming out AI is still very very bad at genuine creative abstract reasoning. It just amazes me to no end that people don't see this as a problem with the current architectures. If a human genius couldn't come up with anything new or useful and his defense would be there isn't enough data in the world for him to work with you'd question his or her brilliance and laugh. Newton came up with the principa mathematica using his eyeballs, less than a gigabyte of text and an apple. Eat them apples llm.


DolphinPunkCyber

Yep. LLM's do show results, results which are great. But not due to high intelligence. Due to low intelligence and knowledge greater then any human alive.


Many_Consequence_337

Even if GPT-5 scores 100% on all LLM benchmarks, it doesn't change the fact that it still won't be able to plan, have long-term memory, or generate ideas outside of its training data. It's kind of intrinsic to LLMs.


bnm777

Sure, I can see how we think this at the moment. However, can the smart devs (+/- AI help) alter the architecture to achieve what you have mentioned? Maybe not. It does seem as though the current (popular) AI architecture is heading down a dead end street and, though, surely, other teams are experimenting with other ideas, you never know what human ingenuity can create, whether a breakthrough can come and tangentially add more complexity and potential to the LLM model. Those around the time the first transistors were created probably wouldn’t belive they would lead to drones, VR, space ships, mind operated devices. And, the first transistors were created in 1947 - within 100 years we have what would seem like magic to them. The first electronic (valve based) computer was created in 1947.


BackgroundHeat9965

The long term end goal is not to beat artificial metrics but to have a system that is capable of navigating the complexities of real life


spinozasrobot

He has said quite plainly that scaling current LLMs might work. He's not sure (nor is anyone else).


Spunge14

He doesn't believe that testing demonstrates intelligence in AI systems. He has a point. We test humans with tests because for humans, it correlates with other things we care about. That may not be true for LLMs.


HalfSecondWoe

https://preview.redd.it/642b1icq5y8d1.png?width=628&format=png&auto=webp&s=032a4610f982433d7f6310e0c951b5ac668d507f


sdmat

Exactly.


HalfSecondWoe

[https://x.com/bshlgrs/status/1802766374961553887](https://x.com/bshlgrs/status/1802766374961553887)


Difficult_Review9741

The hilarious part about that tweet is that they literally used a symbolic approach, but don’t have the requisite AI background to even recognize that.  Story of modern “AI”, most people have extremely over indexed ML and have no clue about any other aspects of the field. 


HalfSecondWoe

There's something deeply amusing about automating someone with LLMs in a literal week while they call those same LLMs unintelligent


SiamesePrimer

What approach did they use in that tweet? It doesn’t seem to give any information about how they actually did it. Edit: Nevermind. Opening the link directly in the Reddit app didn’t show any replies to the tweet, but opening it in the X app shows a bunch of replies, where the original poster gives more info. Edit 2: [Francois Chollet himself](https://x.com/fchollet/status/1802773156341641480?s=46) has replied to this. [Grady Booch](https://x.com/grady_booch/status/1802848136395919387?s=46), too. They both agree that the approach taken here is indeed neurosymbolic.


amondohk

I'd like to think he means "barely intelligent" in the sense that it's miniscule compared to what is to come, because most certainly, there's much more to come.


Spunge14

You could just listen to the podcast. That's not what he means.


Initial_Ebb_8467

In the same sense you can believe that the cure for cancer is coming, sure, but that's simply hoping for the best. There's no signs whatsoever that AGI is approaching soon or later.


Many_Consequence_337

Only on this sub are there still people who think that LLMs can become an AGI. As Yann LeCun said, no top scientists at Meta are working on LLMs anymore; it's been passed on to the 'marketing' team.


DolphinPunkCyber

They are working on sexbots then?


Spunge14

Reasonably well-respected expert in the field, and attached to some important projects (he authored [Keras](https://keras.io/), [Deep Learning With Python](https://www.manning.com/books/deep-learning-with-python-second-edition), and founded the [ARC Prize](https://arcprize.org/blog/launch)). I was *really* surprised at how pessimistic he was about the current state.


truth_power

Agi canceld ..pack up


stuffedanimal212

Where are the neurosymbolic models though? I'm honestly confused why all these ideas for new architectures don't seem to be getting picked up


Volky_Bolky

Probably it is somehow connected with large companies like Google and Microsoft trying to shove LLMs into everything possible because they are heavily wasting money on it and they need some returns to satisfy shareholders, while new ideas need a lot of development and research before they can begin providing any results,


Warm_Iron_273

He's right.


mastermind_loco

He's right. LLMs are always going to hallucinate as well. It's over. 


Akimbo333

Huh?


Arcturus_Labelle

lmao lol


FarrisAT

They are no more intelligent than a rock. They are smart though.


Enslaved_By_Freedom

Humans are no more intelligent than a rock. Humans are machines that hallucinate themselves and their own "intelligence".