NHacker Next
  • new
  • past
  • show
  • ask
  • show
  • jobs
  • submit
Why AI systems don't learn – On autonomous learning from cognitive science (arxiv.org)
Animats 47 minutes ago [-]
Not learning from new input may be a feature. Back in 2016 Microsoft launched one that did, and after one day of talking on Twitter it sounded like 4chan.[1] If all input is believed equally, there's a problem.

Today's locked-down pre-trained models at least have some consistency.

[1] https://www.bbc.com/news/technology-35890188

Earw0rm 33 minutes ago [-]
Incredible to accomplish that in a day - it took the rest of the world another decade to make Twitter sound like 4chan, but thanks to Elon we got there in the end.
zhangchen 5 hours ago [-]
Has anyone tried implementing something like System M's meta-control switching in practice? Curious how you'd handle the reward signal for deciding when to switch between observation and active exploration without it collapsing into one mode.
robot-wrangler 4 hours ago [-]
> Curious how you'd handle the reward signal for deciding when to switch between observation and active exploration without it collapsing into one mode.

If you like biomimetic approaches to computer science, there's evidence that we want something besides neural networks. Whether we call such secondary systems emotions, hormones, or whatnot doesn't really matter much if the dynamics are useful. It seems at least possible that studying alignment-related topics is going to get us closer than any perspective that's purely focused on learning. Coincidentally quanta is on some related topics today: https://www.quantamagazine.org/once-thought-to-support-neuro...

fallous 3 hours ago [-]
The question is does this eventually lead us back to genetic programming and can we adequately avoid the problems of over-fitting to specific hardware that tended to crop up in the past?
t-writescode 3 hours ago [-]
Or possibly “in addition to”, yeah. I think this is where it needs to go. We can’t keep training HUGE neural networks every 3 months and throw out all the work we did and the billions of dollars in gear and training just to use another model a few months.

That loops is unsustainable. Active learning needs to be discovered / created.

aanet 9 hours ago [-]
by Emmanuel Dupoux, Yann LeCun, Jitendra Malik

"he proposed framework integrates learning from observation (System A) and learning from active behavior (System B) while flexibly switching between these learning modes as a function of internally generated meta-control signals (System M). We discuss how this could be built by taking inspiration on how organisms adapt to real-world, dynamic environments across evolutionary and developmental timescales. "

iFire 6 hours ago [-]
https://github.com/plastic-labs/honcho has the idea of one sided observations for RAG.
dasil003 8 hours ago [-]
If this was done well in a way that was productive for corporate work, I suspect the AI would engage in Machievelian maneuvering and deception that would make typical sociopathic CEOs look like Mister Rogers in comparison. And I'm not sure our legal and social structures have the capacity to absorb that without very very bad things happening.
gotwaz 4 hours ago [-]
Not just CEOs, Legal and social structures will also be run by AI. Chimps with 3 inch brains cant handle the level of complexity global systems are currently producing.
AdieuToLogic 2 hours ago [-]
> If this was done well in a way that was productive for corporate work, I suspect the AI would engage in Machievelian maneuvering and deception that would make typical sociopathic CEOs look like Mister Rogers in comparison.

Algorithms do not possess ethics nor morality[0] and therefore cannot engage in Machiavellianism[1]. At best, algorithms can simulate same as pioneered by ELIZA[2], from which the ELIZA effect[3] could be argued as being one of the best known forms of anthropomorphism.

0 - https://www.psychologytoday.com/us/basics/ethics-and-moralit...

1 - https://en.wikipedia.org/wiki/Machiavellianism_(psychology)

2 - https://en.wikipedia.org/wiki/ELIZA

3 - https://en.wikipedia.org/wiki/ELIZA_effect

qsera 2 hours ago [-]
https://en.wikipedia.org/wiki/ELIZA_effect

>As Weizenbaum later wrote, "I had not realized ... that extremely short exposures to a relatively simple computer program could induce powerful delusional thinking in quite normal people."...

That pretty much explain the AI Hysteria that we observe today.

reverius42 1 hours ago [-]
ELIZA couldn't write working code from an English-language prompt though.

I think the "AI Hysteria" comes more from current LLMs being actually good at replacing a lot of activity that coders are used to doing regularly. I wonder what Weizenbaum would think of Claude or ChatGPT.

qsera 43 minutes ago [-]
>ELIZA couldn't write working code from an English-language prompt though.

Yea, that is kind of the point. Even such a system could trick people into delusional thinking.

> actually good at replacing a lot of activity that coders are used to...

I think even that is unrealistic. But that is not what I was thinking. I was thinking when people say that current LLMs will go on improving and reach some kind of real human like intelligence. And ELIZA effect provides a prefect explanation for this.

It is very curious that this effect is the perfect thing for scamming investors who are typically bought into such claims, but under ELIZA effect with this, they will do 10x or 100x investment....

marsten 7 hours ago [-]
Agents playing the iterated prisoner's dilemma learn to cooperate. It's usually not a dominant strategy to be entirely sociopathic when other players are involved.
ehnto 6 hours ago [-]
You don't get that many iterations in the real world though, and if one of your first iterations is particularly bad you don't get any more iterations.
cortesoft 4 hours ago [-]
But AI will train in the artificial world
ehnto 4 hours ago [-]
They still fail in the real world, where a single failure can be highly consequential. AI coding is lucky it has early failure modes, pretty low consequence. But I don't see how that looks for an autonomous management agent with arbitrary metrics as goals.

Anyone doing AI coding can tell you once an agent gets on the wrong path, it can get very confused and is usually irrecoverable. What does that look like in other contexts? Is restarting the process from scratch even possible in other types of work, or is that unique to only some kinds of work?

est 41 minutes ago [-]
"don't learn" might be a good feature from a business point of view

Imagine if AI learns all your source code and apply them to your competitor /facepalm

beernet 9 hours ago [-]
The paper's critique of the 'data wall' and language-centrism is spot on. We’ve been treating AI training like an assembly line where the machine is passive, and then we wonder why it fails in non-stationary environments. It’s the ultimate 'padded room' architecture: the model is isolated from reality and relies on human-curated data to even function.

The proposed System M (Meta-control) is a nice theoretical fix, but the implementation is where the wheels usually come off. Integrating observation (A) and action (B) sounds great until the agent starts hallucinating its own feedback loops. Unless we can move away from this 'outsourced learning' where humans have to fix every domain mismatch, we're just building increasingly expensive parrots. I’m skeptical if 'bilevel optimization' is enough to bridge that gap or if we’re just adding another layer of complexity to a fundamentally limited transformer architecture.

jdkee 7 hours ago [-]
LeCun has been talking about his JEPA models for awhile.

https://ai.meta.com/blog/yann-lecun-ai-model-i-jepa/

Xunjin 3 hours ago [-]
In this podcast episode[0] he does talk about this kind of model and how it "learns about physics" through experience instead of just ingesting theorical material.

It's quite eye opening.

0. https://youtu.be/qvNCVYkHKfg

tranchms 4 hours ago [-]
We are rediscovering Cybernetics
walterbell 1 hours ago [-]
Biological Computer Laboratory (1958-1976), https://web.archive.org/web/20190829234412/http://bcl.ece.il...
QuesnayJr 1 hours ago [-]
It's striking how cybernetics has gone from dated to timely.
Frannky 3 hours ago [-]
Can I run it?
lock-locku 7 hours ago [-]
[dead]
g_a_a_s 6 hours ago [-]
[dead]
theLewisLu 3 hours ago [-]
[dead]
Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact
Rendered at 06:59:17 GMT+0000 (Coordinated Universal Time) with Vercel.