Over for Google and OpenAI

loyolaxavvierretard

loyolaxavvierretard

𝕯𝖝𝕯 π–ˆπ–—π–Šπ–œ . Alonso
Joined
Mar 1, 2025
Posts
18,651
Reputation
46,149

Apparently, the researchers say that the reasoning models have 0 accuracy as the logical reasoning tests go up in complexity

The ques for the established benchmarks might already have answers baked into the training set of the models so they were inaccurate when assessing a model's accuracy

@Jason Voorhees career extended by 20 years πŸ’€πŸ’€

Link to paper
 
Last edited:
  • +1
  • JFL
Reactions: itzyaboyJJ, Bitchwhipper2, Basedman420 and 13 others
@Snicket
 
  • +1
Reactions: Snicket and 2025cel
@Chimera
 
  • +1
Reactions: Chimera
chatgpt couldn't understand my 1 sentence prompt today, can confirm
 
  • JFL
  • +1
Reactions: cute and loyolaxavvierretard
chatgpt couldn't understand my 1 sentence prompt today, can confirm
Its for purely logical reasoning tasks though.

Specifically highly complex tasks
 
  • Woah
  • +1
Reactions: cute and DirtyBlonde
exactly, AI is extremely gay and over-hyped to the point where it cringes you
 
  • JFL
Reactions: loyolaxavvierretard
  • JFL
Reactions: 2025cel
  • Love it
Reactions: loyolaxavvierretard
Can’t be bothered with this tech stuff.

Just going outside and doing anything mogs.

1
 
Last edited:
  • +1
Reactions: loyolaxavvierretard
Lol, Apple has been left behind in the AI department, that's why they desperately try to downplay the importance of AI...
 
  • +1
  • JFL
Reactions: 2025cel and loyolaxavvierretard
Lol, Apple has been left behind in the AI department, that's why they desperately try to downplay the importance of AI...

The problems in the paper were not very tough though. The river problem was simple enough for humans to do it
 
  • +1
Reactions: Gonthar
I haven't been following the story very closely.
What's the bottleneck of AI in personal phone usage?
Surely some kind of GPT based model could do an adequate job? But obviously it's more complex than that.
 
  • +1
Reactions: loyolaxavvierretard
this is why grok mogs
 
  • JFL
Reactions: loyolaxavvierretard and Lord Shadow
obviously lol for anyone thinking AI wasnt going to be any different from just a smarter google search than its over @lifeless
 
  • JFL
Reactions: loyolaxavvierretard and lifeless
I haven't been following the story very closely.
What's the bottleneck of AI in personal phone usage?
Surely some kind of GPT based model could do an adequate job? But obviously it's more complex than that.
Its not personal usage to be specific.

The reasoning models just fail when given deterministic algorithmic problems.

It should be doable for them since its within the context window but they just "give up" before attempting as more variables are stacked on
 
  • +1
Reactions: Snicket
because Ai is stupid they cant think
 
  • JFL
Reactions: piec and loyolaxavvierretard
Chatgpt is a genuine retard tbh
 
  • JFL
Reactions: loyolaxavvierretard
Chatgpt is a genuine retard tbh
Still has a good knowledge base though. 2030 will be either complete bust and start an AI winter or a complete boom
 
  • +1
Reactions: Bitchwhipper2
Tell that to the soldiers who get their faces blasted off by machine learning drones
 
  • +1
Reactions: loyolaxavvierretard
Tell that to the soldiers who get their faces blasted off by machine learning drones
You dont need high level reasoning for that.

A bqsic Coordinate system and tracer technology can do that
 
Still has a good knowledge base though. 2030 will be either complete bust and start an AI winter or a complete boom
Yea, it has a good bit of knowledge to draw upon.

Good data analytics. But asking for its own take on philosophical ponderings is just shooting yourself in the foot
 
  • +1
Reactions: loyolaxavvierretard
I haven't been following the story very closely.
What's the bottleneck of AI in personal phone usage?
Surely some kind of GPT based model could do an adequate job? But obviously it's more complex than that.
They can't run the top LLM models locally, you would need a very powerful computer and lots of RAM for that, mostly they are run in the cloud on special servers.
 
  • +1
Reactions: Snicket
They can't run the top LLM models locally, you would need a very powerful computer and lots of RAM for that, mostly they are run in the cloud on special servers.
Good point. Hadn't considered this.
Why can’t phones have cloud-based AI instead of on-device?
 
Last edited:
Its not personal usage to be specific.

The reasoning models just fail when given deterministic algorithmic problems.

It should be doable for them since its within the context window but they just "give up" before attempting as more variables are stacked on
Interesting.

Is Apple limited by trying to develop on-device AI versus using something server-based like Chat GPT or Gemini instead?
 
Last edited:
Good point. Hadn't considered this.
Why can’t phones have cloud-based AI instead of on-device?
It still takes a few seconds until you get a response from ChatGPT, cloud performance can fluctuate a lot depending on how many users are online and using that service, or the Internet speed, etc., a phone would simply lag too much if you would have to wait for seconds for a response to your various requests.
 
  • +1
Reactions: Snicket
It still takes a few seconds until you get a response from ChatGPT, cloud performance can fluctuate a lot depending on how many users are online and using that service, or the Internet speed, etc., a phone would simply lag too much if you would have to wait for seconds for a response to your various requests.
Yeah, makes sense. And with over 1 billion iPhone users worldwide, the cloud infrastructure costs would be astronomical.
 
  • +1
Reactions: Gonthar

Similar threads

Rigged
Replies
143
Views
7K
Brad_
Brad_
D
Replies
26
Views
3K
Deleted member 86409
D
Deleted member 6403
Replies
14
Views
2K
oldcelloser
oldcelloser

Users who are viewing this thread

Back
Top