Ronan Farrow here. Andrew Marantz and I spent 18 months on this investigation. Happy to answer questions about the reporting.
philip1209 24 minutes ago [-]
We talk about Sam Altman a lot. At this point he has a Hollywood movie in post-production, a book ("The Optimist"), and a seemingly endless stream of profiles. It feels intellectually lazy to keep researching the same guy when the industry is moving beyond him.
All evidence today suggests Anthropic is passing OpenAI in relative and absolute growth. So where's the critical reporting? The DOD coverage was framed around the Pentagon's decisions, not Anthropic's. And nobody seems interested in examining whether the company that branded itself as the ethical AI lab actually is one. That seems like a story worth writing.
solenoid0937 11 minutes ago [-]
> whether the company that branded itself as the ethical AI lab actually is one
FWIW I have two(!!) close friends working for Anthropic, one for nearly two years and one for about 4 months.
Both of them tell me that this is not just marketing, that the company actually is ethical and safety conscious everywhere, and that this was the most surprising part about joining Anthropic for them. They insist the culture is actually genuine which is practically unicorn rarity in corporate America.
We have worked for FAANG so I know where they're coming from; this got me to drop my cynicism for once and I plan on interviewing with them soon. Hopefully I can answer this question for myself.
giwook 12 minutes ago [-]
There may be a reason why Altman is talked about a lot. This article in particular surfaces real information and new perspectives we've not heard in this level of detail before on some pretty significant topics that will be impacting you, me, and pretty much everyone we know not only today but well into the future.
You have a point in that Anthropic deserves some coverage too and that there are interesting perspectives that we've not heard of on that front either.
But just because that's true doesn't mean this article isn't very much relevant and needed.
Because it is.
ronanfarrow 16 minutes ago [-]
For what it’s worth, the story, while focused on OpenAI, is not uncritical of Anthropic. It explores whether there is a wider race to the bottom in terms of safety, and erosion of even some of Anthropic’s commitments.
basisword 7 minutes ago [-]
OP says they’ve been working on this for 18 months. Most of what you’ve said wasn’t the case until much more recently.
xvector 20 minutes ago [-]
Normies don't know what an "Anthropic" is. They use ChatGPT. Particularly sharp normies might know that ChatGPT is made by OpenAI, and the sharpest might know that Sam Altman is the CEO.
Now, they may have heard the word "Anthropic" due to recent media coverage. But they don't know what it is and don't remember what it makes. The fact that all businesses use "Anthropic" is about as relevant to them as knowing the overseas shipping company for all the shit they buy off Amazon.
So articles about OAI will always produce more revenue for the media, because it's related to what normies actually use day to day.
aragonite 31 minutes ago [-]
I had a question about reporting conventions. In the paragraph where Altman is said to have told Murati that his allies were "going all out" to damage her reputation, the claim is attributed to "someone with knowledge of the conversation" but the attribution is tucked inconspicuously into the middle of the sentence (rather than say leading upfront ("According to someone with knowledge of the conversation, Altman...")) and Altman's non-recollection appears only parenthetically.
As a reader, am I supposed to infer anything about evidentiary weight from these stylistic choices? When a single anonymous source's testimony is presented in a "declarative" narrative style like here (with the attribution in a less prominent position), should we read that as reflecting high confidence on your end (perhaps from additional corroboration not fully spelled out)? And does the fact that Altman’s non-recollection appears in parentheses carry any epistemic signal (e.g. that you assign it less evidentiary weight)? Or is that mostly a matter of (say) prose rhythm?
replytofarrow 19 minutes ago [-]
it's not complicated. most of what Ronan Farrow knows about the attempted ouster came from Helen Toner, and some of it is directly attributed to her, and some of it she didn't want them to attribute to her, which is totally valid.
she feels very frustrated that she was dragged through the mud, with the WSJ and Satya Nadella saying she wasn't qualified for being on the board, when in fact she was totally right, along with the other PhDs, that Sam Altman - and for that matter Greg Brockman - had to go.
everybody else has the wisdom to not talk to Ronan Farrow.
jzymbaluk 2 hours ago [-]
Hi Ronan, thanks for the article and for answering questions.
My question is, how do you know when an enormous project like this, conducted over an 18-month time span is "done"? I assume you get a lot of leeway from editors and publishers on this matter. How do you make the decision to finally pull the trigger on publishing?
cs702 12 hours ago [-]
Thank you for coming on HN and offering to answer questions.[a]
This is a fantastic piece, very timely, evidently well-researched, and also well-written. Judging by the little that I know, it's accurate. Thank you for doing the work and sharing it with the world.
OpenAI may be in a more tenuous competitive position than many people realize. Recent anecdotal evidence suggests the company has lost its lead in the AI race to Anthropic.[b]
Many people here, on HN, who develop software prefer Claude, because they think it's a better product.[c]
Is your understanding of OpenAI's current competitive position similar?
Thank you for this, very much appreciate the thoughtful response.
The piece captures some of the anxieties within OpenAI right now about their competitive position. This obviously ebbs and flows but of late there has been much focus on Anthropic's relative position. We of course mention the allegations of "circular deals" and concerns about partners taking on debt.
cs702 6 hours ago [-]
Thank you. Yes, I saw that. The company's always been surrounded by endless talk about insane hype, speculative bubbles, and financial engineering. I wasn't asking so much about that.
I was asking more about your informed view on how OpenAI's technology, products, and roadmap are perceived, particularly by customers and partners, in comparison to those of competitors.
If you have an opinion about that, everyone here would love to hear about it.
Ericson2314 45 minutes ago [-]
Ronan Farrow's expertise is investigations into elite amorality, not evaluating technical products. Why are you asking this question?
cs702 10 minutes ago [-]
I didn't asking him to evaluate them. I asked him how customer and partners perceive them.
He's had so many conversations that he likely has a sense of how perceptions of the company and its offerings have changed.
I'm curious.
irishcoffee 57 minutes ago [-]
My guess is that the answer to your question, fantastic question, is that nobody knows. I remember having the same thoughts when Covid was first “arriving” if you will: we wanted people in the know to throw us a nugget of information, and they just didn’t know.
As it turns out, and what I’m kind of going with for this LLM shit, is that it’ll play out exactly how you think it will. The companies are all too big to fail, with billionaire backers who would rather commit fraud than lose money.
unsupp0rted 4 hours ago [-]
Many of us prefer OpenAI's Codex, because we think it's a better product.
No comment on the CEO: I just find the product superior in everything but UI/UX and conversation. It's better at quality code.
mliker 4 hours ago [-]
Who is “us”? It does seem that some scientists prefer Codex for its math capabilities but when it comes to general frontend and backend construction, Claude Code is just as good and possibly made better with its extensive Skills library.
Both codex and Claude code fail when it comes to extremely sophisticated programming for distributed systems
keldaris 25 minutes ago [-]
As a scientist (computational physicist, so plenty of math, but also plenty of code, from Python PoCs to explicit SIMD and GPU code, mostly various subsets of C/C++), I can confirm - Codex is qualitatively better for my usecases than Claude. I keep retesting them (not on benchmarks, I simply use both in parallel for my work and see what happens) after every version update and ever since 5.2 Codex seems further and further ahead. The token limits are also far more generous (and it matters, I found it fairly easy to hit the 5h limit on max tier Claude), but mostly it's about quality - the probability that the model will give me something useful I can iterate on as opposed to discard immediately is much higher with Codex.
For the few times I've used both models side by side on more typical tasks (not so much web stuff, which I don't do much of, but more conventional Python scripts, CLI utilities in C, some OpenGL), they seem much more evenly matched. I haven't found a case where Claude would be markedly superior since Codex 5.2 came out, but I'm sure there are plenty. In my view, benchmarks are completely irrelevant at this point, just use models side by side on representative bits of your real work and stick with what works best for you. My software engineer friends often react with disbelief when I say I much prefer Codex, but in my experience it is not a close comparison.
unsupp0rted 4 hours ago [-]
Us = me and say /r/codex or wherever Codex users are. I've tried both, liked both, but in my projects one clearly produces better results, more maintainable code and does a better job of debugging and refactoring.
sampullman 3 hours ago [-]
That's interesting, I actively use both and usually find it to be a toss up which one performs better at a given task. I generally find Claude to be better with complex tool calls and Codex to be better at reviewing code, but otherwise don't see a significant difference.
SOLAR_FIELDS 2 hours ago [-]
If you want to find an advocate for Codex that can give a pretty good answer as to why they think it's better, go ask Eric Provencher. He develops https://repoprompt.com/. He spends a lot of time thinking in this space and prefers Codex over Claude, though I haven't checked recently to see if he still has that opinion. He's pretty reachable on Discord if you poke around a bit.
aswanson 3 hours ago [-]
Any difference in performance on mobile development?
sampullman 1 hours ago [-]
For that I'm not so sure. I tried both early 2025 and was disappointed in their ability to deal with a TCA based app (iOS) and Jetpack compose stuff on Android, but I assume Opus 4.6 and GPT 5.4 are much better.
rocketpastsix 2 hours ago [-]
yea Im not in this "us" you speak of.
zeroxfe 3 hours ago [-]
I'm in that camp -- I have the max-tier subscription to pretty much all the services, and for now Codex seems to win. Primarily because 1) long horizon development tasks are much more reliable with codex, and 2) OpenAI is far more generous with the token limits.
Gemini seems to be the worst of the three, and some open-weight models are not too bad (like Kimi k2.5). Cursor is still pretty good, and copilot just really really sucks.
7thpower 2 hours ago [-]
Not a scientist and use codex for anything complex.
I enjoy using CC more and use it for non coding tasks primarily, but for anything complex (honestly most of what I do is not that complex), I feel like I am trading future toil for a dopamine hit.
zem 2 hours ago [-]
I've found claude startlingly good at debugging race conditions and other multithreading issues though.
josephg 2 hours ago [-]
My rule of thumb is that its good for anything "broad", and weaker for anything "deep". Broad tasks are tasks which require working knowledge of lots of random stuff. Its bad at deep work - like implementing a complex, novel algorithm.
LLMs aren't able to achieve 100% correctness of every line of code. But luckily, 100% correctness is not required for debugging. So its better at that sort of thing. Its also (comparatively) good at reading lots and lots of code. Better than I am - I get bogged down in details and I exhaust quickly.
An example of broad work is something like: "Compile this C# code to webassembly, then run it from this go program. Write a set of benchmarks of the result, and compare it to the C# code running natively, and this python implementation. Make a chart of the data add it to this latex code." Each of the steps is simple if you have expertise in the languages and tools. But a lot of work otherwise. But for me to do that, I'd need to figure out C# webassembly compilation and go wasm libraries. I'd need to find a good charting library. And so on.
I think its decent at debugging because debugging requires reading a lot of code. And there's lots of weird tools and approaches you can use to debug something. And its not mission critical that every approach works. Debugging plays to the strengths of LLMs.
bko 56 minutes ago [-]
I also find Codex much more generous in terms of what you get with a Pro ($20/mo) subscription. I use it pretty much non-stop and I have yet to hit a limit. Weekly reset is much better as well.
enraged_camel 2 hours ago [-]
Yeah, there are dozens of you. Dozens!
brightbeige 12 hours ago [-]
He’s replying on this twitter thread - perhaps someone with an account can ask there and link his comment here?
The statements around the sexual abuse allegations seemed to be the most puzzling to me - his sister’s allegations and claims of underage partners because he has a tendency to hook up with younger partners. It does seem like this piece gives him a pretty clean bill of health in that matter - I guess would you be able to talk about how you investigated?
Did you do any extra investigations into Annie’s allegations? It feels to me like the unstated conclusion is recovered memory can’t be trusted, which is a popular understanding but a very wrong one put out by the now defunct and discredited False Memory Syndrome Foundation. It was founded by the parents of the psychologist who coined DARVO, directly in reaction to her accusing them of abuse.
Dissociation is real (I have a dissociative disorder, and abuse I “recovered” but did not remember for much of my adolescence and early adulthood has been corroborated by third parties) and many CSA survivors have severe memory problems that often don’t come to a head until adulthood. I know you didn’t dismiss her claim, but the way the public tends to think about recovered memories is shaped primarily by that awful organization.
ronanfarrow 8 hours ago [-]
All fair points on trauma and memory.
As noted in the piece, we spent months talking to Altman's partners and what we found and didn't is as described.
taurath 6 hours ago [-]
Thanks for the response! Cheers just fully reread the piece and appreciate your reporting.
girvo 4 hours ago [-]
It's super neat to see you here on HN taking questions, kudos :)
gowld 28 minutes ago [-]
That's not a fair assessment.
"False memory syndrome" and "repressed/recovered memory" are both outside scientific mainstream consensus.
hello_humans 3 hours ago [-]
[flagged]
stavros 17 minutes ago [-]
There's a very minor typo in the article:
> “Investors are, like, I need to know you’re gonna stick with this when times get hard,”
Should be:
> “Investors are like, I need to know you’re gonna stick with this when times get hard,”
cm2012 31 minutes ago [-]
I just spent a while reading the article. I really appreciate you writing it. In my case, it made me like Sam Altman a lot more. But I was only able to conclude this because of all the evidence you took the time to put together. It paints the picture of someone trying to do something very difficult in a rapidly changing environment and a lot of pressure, but still making the important choices and not shirking them.
ronanfarrow 12 minutes ago [-]
Interesting to hear! While this hasn’t been a commonplace reaction, I think if I do my job right it should allow people to read the facts as they will, exactly like this. It’s strenuously designed to be fair and, where appropriate, even generous.
sebmellen 2 hours ago [-]
Ronan Farrow on Hacker News. Now I’ve seen everything.
ronanfarrow 10 minutes ago [-]
I’ve really appreciated how substantive and polite the discourse here is, overall!
egonschiele 45 minutes ago [-]
Just wanted to say what an incredible person you are! Catch and Kill and the related reporting was awesome too!
ronanfarrow 10 minutes ago [-]
This is so appreciated, thank you! These stories can honestly take a lot out of me so thoughtful reactions mean a lot.
fblp 3 hours ago [-]
Hi Ronan appreciate you being here. what would help you and others continue to do journalism like this? (including commenting on HN?)
ronanfarrow 6 minutes ago [-]
This is a vast and tricky question. The business model has basically fallen out from under journalism, and especially this kind of labor-intensive investigative reporting. The media landscape is increasingly dominated by moneyed individuals and companies essentially buying up the discourse.
I would really suggest subscribing to and finding ways to amplify independent outlets and journalists, and encouraging others to do so.
cmiles8 12 hours ago [-]
Great reporting.
Altman describes his shifting views as genuine good faith evolution of thinking. Do you believe he has a clear North Star behind all this that’s not centered on himself?
ronanfarrow 8 hours ago [-]
The piece is an interrogation of this very question, at great length and with some nuance. I think what it does most usefully is scrutinize an array of different answers to the question.
My own impression after many hours of conversation is that he is identifying something of a true north star when he frames this around "winning." There are people in the story who talk about him emphasizing a desire for power (as opposed to, say, wealth). I think he probably also believes, to some extent, the story he tells that equates winning, and his gaining power, with a superabundant utopian future for all.
However, I think critics correctly highlight a tension between his statements about centering humanity writ large and his tilt into relentless accelerationism.
i7l 12 hours ago [-]
(Other people's) money.
giwook 2 hours ago [-]
Any plans to tackle any of the other folks who might be mentioned in the same sentence as Altman, like Darius Amodei?
I think the comment was out of legitimate interest rather than weighing one against the other
giwook 17 minutes ago [-]
Huh? It's a genuine question. The article is great and the writer did a fantastic job.
Please try to give people the benefit of the doubt though I know it's hard in today's society.
wyldfire 57 minutes ago [-]
Dang, can you substantiate that this is actually Mr. Farrow like he claims?
Or Mr Farrow can you post some evidence somewhere we can see?
rhlannx 3 hours ago [-]
I have the feeling that if you write an article in that style, the subject of the story becomes the hero even if you insert a couple of negatives. In the same manner that Michael Corleone becomes the hero of The Godfather.
I'm not pleased with the headline and the general framing that AI works. The plagiarism and IP theft aspects are entirely omitted. The widespread disillusion with AI is omitted.
On the positive side, the Kushner ad Abu Dhabi involvements (and threats from Kushner) deserve a wider audience.
My personal opinion is that "who should control AI" is the wrong question. In the current state, it is an IP laundering device and I wonder why publications fall silent on this. For example, the NYT has abandoned their crown witness Suchir Balaji who literally perished for his convictions (murder or not).
ronanfarrow 5 minutes ago [-]
For what it’s worth, I don’t think the piece at all avoids key areas of disillusionment with the technology. Quite the contrary.
23 minutes ago [-]
xnx 10 hours ago [-]
In depth reporting is great. This is a really tricky topic to cover over the course of 18 months. A year and a half ago OpenAI was ascendant, now it's -at best- stalling and, more likely, trending toward irrelevant.
_alternator_ 2 hours ago [-]
Do you think the recent conflict between Anthropic and the Department of War, and the apparent bootlicking by OpenAI has fundamentally altered the public perception of OAI? Are they the baddies now in the general public opinion?
Uptrenda 1 hours ago [-]
Damn, just wanted to say reporters are scary... The amount of detail here is huge. You think of hackers as the ones good at doxing... Nah, its reporters.
Stevvo 3 hours ago [-]
Love the visual. Fantastic.
felixgallo 2 hours ago [-]
This is brilliant work, guys. Did you get any pressure to soften or spike the story?
ronanfarrow 3 minutes ago [-]
I won’t get into behind-the-scenes specifics here but I think you can imagine how pressurized this topic was and the amount of heat that tends to generate. I’m used to getting a lot of blowback and it’s never fun. I just hope the work is meticulous and fair enough, and that enough people see the benefits of that, that I get to continue to do it.
12 hours ago [-]
12 hours ago [-]
sieabahlpark 2 hours ago [-]
[dead]
FloorEgg 4 hours ago [-]
Hi Ronan,
I would love to read your piece and pay you and new Yorker for it, but I am not interested in paying a subscription. If I could press a button and pay a reasonable one time license such as $3 or $5 for just this article, or better yet a few cents per paragraph as they load in, I wouldn't hesitate.
However I'm not going to pay for yet another subscription to access one article I'm interested in.
I'm sure you can't do anything about this, but I just wanted you to know.
You deserve to be compensated for great journalism. In this case, unfortunately, I won't read it and you won't earn income from me.
CookieTonsure 1 hours ago [-]
The public library [digital edition] is absolutely the correct answer. I maintain a library card at 3 different local municipal library systems. My local city's library offers access to several Digital Library apps, including Overdrive, Hoopla, and Libby.
It took me a couple searches in Libby to locate the New Yorker and it offered up the current issue right away. The article is on page 32. It is ridiculous that anyone considers to access this from "The Public Internet" or the newyorker dot com website, rather than simply turning to your public library, which has been the go-to resource for basically everyone, for hundreds of years.
You're already paying for your library with your tax dollars. If you don't use it, you may lose it, but you will certainly lose out by subsidizing bums, vagrants, and other families who use the library to their heart's content.
The public library also features lots of streaming and CD music, videos, and video games, that you can freely check out without any cost. In fact, my local library staff told me that they've abolished overdue fees. Libby and the digital apps will automatically renew or return materials. My physical books even got auto-renewed three times before I needed to manually do it, or bring them back into the building.
cloud_line 3 hours ago [-]
You could buy a physical copy (and this isn't meant to sound sarcastic).
jzymbaluk 2 hours ago [-]
You can walk down to a bookstore or anywhere that sells magazines and buy a physical copy
mattbee 4 hours ago [-]
Or just switch your browser to Reader Mode and it's free.
IrishTechie 4 hours ago [-]
I’ve often thought about a model like this and would love to see a few news outlets run it as a pilot and see how it stacks up.
mikeyouse 2 hours ago [-]
Many have tried it (as well as the oft-recommended micropayments idea) and it never justifies the added expense and overhead of the customization. Closest is probably the NYTimes’ gift article feature.
Dylan16807 2 hours ago [-]
I really doubt the implementation difficulty is the actual reason. It's not hard to have an extra table of specific article permissions.
caycep 4 hours ago [-]
You could hit up a public library...
1 hours ago [-]
eichin 3 hours ago [-]
Looking online it looks like the newsstand price of an issue is around $10 (which I'd assume is heavily ad subsidized, if anyone is still buying print ads?) which is an interesting data point for a pricing model. (Of course, I looked online because I have no idea where I'd find a newsstand around here - the nearest newsstand that show up on google maps has reviews that say "It's just snacks and scratch tickets." and "three newspapers and no magazines" - I may have to stop by just to see what three newspapers they have :-)
loloquwowndueo 12 hours ago [-]
[flagged]
LoganDark 12 hours ago [-]
Many browsers let you disable autoplay globally.
loloquwowndueo 12 hours ago [-]
Sure, there are a couple of buttons I can press to stop the video. Why do I have to? Find me one person who likes auto playing videos. The page was created with a deliberate annoying choice that I have to go out of my way to override.
binarymax 11 hours ago [-]
Why do you think the author of this piece, to who you originally replied, has any control over this?
11 hours ago [-]
LoganDark 12 hours ago [-]
I'm not talking about pausing the video after it starts playing. I'm talking about a global setting to prevent videos from playing before you manually unpause them. Safari has such a setting, for instance.
loloquwowndueo 10 hours ago [-]
Exactly what “I have to go out of my way to override” covers, from my comment.
mannyv 15 minutes ago [-]
A day late and a dollar short. The world has moved on from ChatGPT and OpenAI.
steve_adams_86 2 hours ago [-]
> Amodei, in one of his early notes, recalled pressing Brockman on his priorities and Brockman replying that he wanted “money and power.” Brockman disputes this. His diary entries from this time suggest conflicting instincts. One reads, “Happy to not become rich on this, so long as no one else is.” In another, he asks, “So what do I really want?” Among his answers is “Financially what will take me to $1B.”
I can't imagine having such uninspired thoughts and actually writing them down while in a role of such diverse and worthwhile opportunities. I'd like to ask "how the hell do these people find themselves in these positions", but I think the answer is literally what he wrote in his diary. What a boring answer. We need to filter these people out at every turn, but instead they're elevated to the highest peaks of power.
ks2048 15 minutes ago [-]
It's not surprising. I made this comment on HN before, but if you follow him on Twitter, it's pretty remarkable - the CTO of one of the most important technology companies in the world and he has never (that I've seen) posted something with some technical insight, or just anything interesting about technology. It's just boring truisms, cliches, empty statements, etc.
chromacity 1 hours ago [-]
Eh. It doesn't start or stop with people like Altman, Zuckerberg, or Nadella. I think it's a symptom of a broader problem in tech. Half the people on this site made a decision to work at companies that do shady things, and they did that to maximize personal wealth.
The difference isn't that the average techie doesn't dream of making a billion by any means necessary; it's that most of us don't think we have a shot, so we stick to enabling lesser evils to retire with mere millions in the bank.
bluefirebrand 14 minutes ago [-]
> The difference isn't that the average techie doesn't dream of making a billion by any means necessary
I hope that's not true. If it is, we live in a bleak world indeed.
I can confidently say I've never once dreamed of having billions. I've never wanted billions. Not even in a fanciful manner. What would I do with that money? Buy mansions and megayachts? That's loser stuff
Most of what I want out of life cannot be bought. The pieces that come with a price tag, like a comfortable home, do not require billions
I think only sociopaths want billions because they don't understand spending your life seeking things that actually matter, like family and human connection
kevinqi 1 hours ago [-]
it is disappointing, but is it shocking that people most driven by gaining money/power are the ones the most successful at achieving it?
steve_adams_86 50 minutes ago [-]
What sticks out to me most is that humanity consistently fails to weed these creatures out and regulate society. It's a bug in our social software; we seem to like these broken people rather than recognize that they're a liability.
hackable_sand 5 minutes ago [-]
Trust is not a bug
You need to accept that every generation some people are going to try and fuck things up.
Then you get to decide to stop or help them
basket_horse 25 minutes ago [-]
This isn’t a bug. It’s the driving force of our capitalist society. We are not trying to weed them out. We are trying to encourage them. It’s pretty simple, when they get rich, so do all their investors.
dolebirchwood 2 hours ago [-]
Sociopaths don't have much going for them in life other than winning status games.
kakacik 1 hours ago [-]
While true and we can see them literally everywhere where there is some money and/or power (even miniscule places like classic banks have easily 1/3 of the staff with clear sociopathic traits, I have to deal with them daily... or whole politics) - thats just human nature, or part of it.
Its up to rest of society to keep them in check since classic morals are highly optional and considered nuissance blocking those games. And here we the rest fail pretty miserably, while having on paper perfect tool - majority vote.
buzzerbetrayed 2 hours ago [-]
Sociopath is the next word that people seem to want to entirely destroy the meaning of
dolebirchwood 2 hours ago [-]
Struck a nerve?
JumpCrisscross 54 minutes ago [-]
> Struck a nerve?
No need to be petty. They have a point. We did this with the words racist and fascist. Overinclusion diluted the term and gave cover for the actual baddies to come in. I'm not sure debating who is and isn't a sociopath is as useful as, say, the degree to which Sam is a liar (versus visible).
greenchair 28 minutes ago [-]
Speaking of overinclusion, 'wild' is my nominee for 2026 as I'm seeing it all over the place.
lokar 2 hours ago [-]
Or, some fraction of otherwise good/normal people who “win” are turned into sociopaths by the power and sycophancy.
xorgun 59 minutes ago [-]
[dead]
morleytj 2 hours ago [-]
Wow, this is an incredibly detailed piece. Really in depth reporting and the kind of detailed investigation we need more of on important topics like this.
> "Employees now call this moment “the Blip,” after an incident in the Marvel films in which characters disappear from existence and then return, unchanged, to a world profoundly altered by their absence."
This is a very small detail, but an instinctive grimace crosses my face at the thought of these sort of Marvel references and I'm not entirely sure why.
ytoawwhra92 2 hours ago [-]
They're mass media cynically produced to extract maximum profit from lowest common denominator audiences, so the idea that people working in such influential positions find them appealing enough to reference suggests they are members of that lowest common denominator audience.
The people shaping the future have no taste.
eutropia 58 minutes ago [-]
There's a time and a place for everything, and rejecting popular media as "lowest common denominator" is the most uninspired form of cultural elitism.
Is it cynical to want your <art project> to make a profit? Or for it to make enough profit to subsidize other projects?
Is it cynical to make something accessible so more people who watch it are able to enjoy it?
I agree that it's embarrassing and feels crass when movies both try to be broadly appealing and simultaneously fail to be entertaining or well executed ... but many of the marvel movies clearly surpass that bar.
No one wants to make a bad movie that does poorly with critics and paying customers - but it does happen because making a movie is expensive and complicated and requires a lot of skilled people working together towards the same goal.
Regarding taste: do you think a michelin star chef swears off cheap food like hotdogs or fish and chips? Doubtful - because those foods have their place and the chef is able to enjoy them for what they are rather than use them as an excuse to display a superiority complex.
ytoawwhra92 36 minutes ago [-]
> There's a time and a place for everything
Yeah, I'm saying professional communication isn't the place for Marvel references, and that those who choose to include references to those movies in their professional communications are revealing something about their media tastes.
ianbutler 26 minutes ago [-]
That they relate to the common person and aren't overly snobby?
Noumenon72 1 hours ago [-]
When things reach a certain level of popularity they constitute "mental real estate". Your audience has heard of Groundhog Day, so there is an opening for a movie with that title to make money -- your film will start out already having name recognition and some understanding of what the movie is about.
Thus it is a writer's job not to make references they find appealing to reveal their good taste, but to know what references their audience will find appealing and use them to help communicate concepts. If this bothers you it's because they're insulting you by saying you might be part of the audience that watches Marvel, and you had hoped reading the New Yorker would signal that you aren't.
ytoawwhra92 40 minutes ago [-]
The writers of this piece didn't make the reference.
2 hours ago [-]
tw04 7 minutes ago [-]
I don't even need to read the article to know that he unequivocally can't be trusted. Every action he's taken to this point have shown he will say literally anything to get what he wants.
pharos92 2 hours ago [-]
We focus these critiques far too much on the face rather than the underlying mechanics. Just like in politics, we critique the personality/politician yet the underlying system architecture evades it.
Sam Altman clearly has a long history of nefarious activity. But the underlying threat posted by AI to society, the economy and human freedom persists with or without his presence.
xgulfie 2 hours ago [-]
It's because we only really know one economic system but we've known many people
j2kun 1 hours ago [-]
[dead]
einrealist 2 hours ago [-]
I don't trust anyone who claims that LLMs today are superhumanly intelligent. All they do is perform compute-intensive brute-force attacks on the problem/solution space and call it 'reasoning', all while subsidising the real costs to capture the market. So much SciFi BS and extrapolation about a technology that is useful if adopted with care.
This technology needs to become a commodity to destroy this aggregation of power between a few organizations with untrustworthy incentives and leadership.
shruggedatlas 1 hours ago [-]
Your brain is performing "compute-intensive brute-force attacks on the problem/solution space" as you read this very sentence. You trained patterns on English syntax, structure, and semantics since you were a child and it is supporting you now with inference (or interpretation). And, for compute efficiency, you probably have evolution to thank.
JohnMakin 24 minutes ago [-]
people like to say this like they’re apples to apples but this comparison isn’t remotely how the brain actually works - and even if it did, the brain does it automatically without direction and at an infitesimal percentage of the power required.
And we’re just talking about cognition - it completely ignores the automatic processes such as maintaining and regulating the body and it’s hormones, coordinating and maintaining muscles, visual/spacial processing taking in massive amounts of data at a very fine scale, and informing the body what to do with it - could go on.
One of the more annoying things about this conversation is you don’t even need to make this argument to make the point you’re trying to make, but people love doing it anyway. It needlessly reduces how amazing the human brain is to a bunch of catchy sci fi sounding idioms.
It can be simultaneously true that transformer based language models can be very smart and that the human brain is also very smart. It genuinely confuses me why people need to make it an either/or.
stonyrubbish 21 minutes ago [-]
Human cognition is nothing like AI "cognition." It really bothers me that people think AI is doing the same thing the human mind does. AI is more like a parrot which is trained to give a correct-looking response to any question. The parrot doesn't think, doesn't know what its doing etc, it just does it because it gets a treat every time a "good" answer is prompted. This is why it can't do things like know how many parenthesis are balanced here ((((()))))) (you can test this), it doesn't have any kind of genuine cognition.
chpatrick 9 minutes ago [-]
This is such a boring cliche by now. "thinking" and "knowing what it's doing" are totally vague statements that we barely understand about the human mind but in every comment section about AI people definitively state that LLMs don't do them, whatever they are.
stonyrubbish 6 minutes ago [-]
They aren't so vague that you would argue the parrot is thinking.
wil421 26 minutes ago [-]
If you think this way then why not talk to LLMs exclusively. Don’t let the oxytocin cloud your ability to problem solve.
slopinthebag 29 minutes ago [-]
I get you're trying to do the whole "humans and LLMs are the same" bit, but it's just plainly false. Please stop.
stavros 1 hours ago [-]
> All they do is perform compute-intensive brute-force attacks on the problem/solution space and call it 'reasoning'
If they discover the cure to cancer, I don't care how they did it. "I don't trust anyone who claims they're intelligent" doesn't follow from "all they do is <how they work>".
stonyrubbish 17 minutes ago [-]
> "I don't trust anyone who claims they're intelligent" doesn't follow from "all they do is <how they work>".
It kind of does if how they work is nothing like genuine intelligence. You can (rightly) think AI is incredible and amazing and going to bring us amazing new medical technologies, without wrongly thinking its super amazing pattern recognition is the same thing as genuine intelligence. It should be worrying if people begin to believe the stochastic parrot is actually wise.
bigyabai 1 hours ago [-]
That's moonshot logic that reinforces the parent's point. You'd absolutely care if the AI's cure to cancer entailed full-body transplants or dismemberment.
Noumenon72 58 minutes ago [-]
"The cure for cancer" as a phrase doesn't include those solutions. If the headline was "Pope discovers the cure for cancer" and those were his solutions you would say "No he didn't." OP was referring to AI discovering the cure for cancer that cancer research is working towards.
JumpCrisscross 51 minutes ago [-]
> You'd absolutely care if the AI's cure to cancer entailed full-body transplants or dismemberment
That's not a cure. Like yes, I'd care if the AI says it cures cancer while nuking Chicago. But that isn't what OP said.
Rover222 15 minutes ago [-]
Yeah and everything is just atoms. If you reduce anything enough it’s not real.
I am in 40s and going to be made redundant this June. In future only people who can afford to keep things like Claude, OpenAI and most importantly create value using them more than what others can do be able to survive. Otherwise, game is more or less over, and I question what's next for my own future while I learn to use Claude in FOMO. I cannot trust Sam or others if they will have any interest to keep this tech affordable for common people like me.
kmfrk 10 hours ago [-]
Gobsmacking details about Altmans' time as Y Combinator president, in case anyone's wondering.
Fantastic reporting.
ronanfarrow 8 hours ago [-]
As is always the case with incredibly precise and rigorously fact-checked reporting like this, where every word is chosen carefully (the initial closing meeting for this one was nearly eight hours long, with full deliberation about each sentence), there is more out there on that subject than is explicitly on the page.
kmfrk 7 hours ago [-]
One of the decidedly eerier parts of this story as you keep reading are all the gaps between what people are saying about Altman, and what they clearly want to say about Altman but can't.
devmor 3 hours ago [-]
Throughout my life, what colleagues/friends are unwilling to remark plainly on has been the most telling factor of someone’s character to me.
dugidugout 2 hours ago [-]
This can be true I suppose, but equally I have a few friends who practically play characters as if they've resigned themselves to a role in a sitcom. For instance: one of my friends is late to just about everything and treats everyone as if we are on-call. We plainly note this repeatedly, the friend is, I hope, equally frustrated and embarrassed by it, and in spite of this nothing changes. This is obviously a critical element to their broader character.
Perhaps you mean to distinguish social groups without much intimacy? To which I'm sure we could provide some convincing cases, but this seems like a silly heuristic generally.
rincebrain 2 hours ago [-]
I have been in or next to a number of social circles with such missing stairs, where for various reasons people in the groups have decided to not directly acknowledge certain Facts that are known about some members, because it would involve them confronting their hypocrisy.
Someone cheating regularly on their partner, flagrant substance use problems, controlling people who ostracize anyone who doesn't agree with their sometimes insane perspectives...
People will go along with quite a lot to avoid friction, especially as they get older and picking up new social circles becomes higher cost.
It's possibly the most telling thing, when you see what people say is a hard line versus how they actually respond to it.
Teever 1 hours ago [-]
You mention many proxies of Musk who post negative content about Altman.
In your investigation were you able to determine if Altman has similar proxies?
How common would you say that this is? Do these kinds of people generally have teams of people who sling mud for them?
Can you speculate on how that manifests on a site like Hackernews?
ainch 2 hours ago [-]
Great piece. And a good excuse to read up on the use of diaeresis in English (eg. coördination, reëlection) to distinguish repeated vowels - I hadn't seen the New Yorker's usage before.
goodoldneon 1 hours ago [-]
It isn’t for all repeated vowels; only for when the 2 vowels don’t make a single sound. So “chicken coop” wouldn’t have a dieresis
stavros 1 hours ago [-]
It would if the chickens formed a business structure that was owned and democratically controlled by its member-owners.
OJFord 1 hours ago [-]
Unless it was a chicken coöp... One of few cases it actually resolves an ambiguity!
krackers 2 hours ago [-]
[1] is also good to read as a follow-up, and compare the personalities
It's really interesting reading about how these folks view LLMs. Yeah, they're transformative, but I don't know that we're going to be eating ramen in a Neo-Tokyo street bar anytime soon. So much "A.G.I" mentioned in the article.
0x3f 3 hours ago [-]
It's because they're really good at the kind of busywork the average white collar job requires. Most people are out there writing documents and making presentations. Only when you use them for actual complexity does the shortfall become clear.
ergocoder 3 hours ago [-]
I wonder if Sam might abandon the ship soon. Other co-founders already did.
The main reason is that he gets all the downsides without the upsides. I know $5B is a lot but, for a 700B company, it isn't. If OpenAI was a regular for-profit, he would have been worth >$100B already.
This is probably one of the significant factors why other co-founders left too. It's just a lot of headaches with relatively low reward.
0x3f 2 hours ago [-]
But nobody is going to just gift him the same valuation on the next company. It's not like his execution is OpenAI's moat right now. So where would he be going that's a better deal for him?
ergocoder 2 hours ago [-]
Founding his own company would be one alternative. Full control. No stigma on the non-profit part. Probably get the same paper money as he got now at OpenAI.
raincole 2 hours ago [-]
And OpenAI's influence is hugely exaggerated compared to, say, Google.
ergocoder 2 hours ago [-]
Yes, and it seems people hate him more than Google co-founders, for example.
All the downsides without much upside...
georgemcbay 1 hours ago [-]
> Yes, and it seems people hate him more than Google co-founders, for example.
Sergey Brin is trying to change that lately, but Altman still has a sizable head start.
palata 2 hours ago [-]
IMHO, nobody is remotely worth $1B, period.
The fact that some (usually toxic) individuals get there shows that the system is flawed.
The fact that those individuals feel like they can do anything other than shut up, stay low and silently enjoy the fact that they got waaaay too much money shows that the system is very flawed.
We shouldn't follow billionaires, we should redistribute their money.
simonh 38 minutes ago [-]
If someone founds a company, grows it and owns $1bn of its stock, they don’t have $1bn in cash to distribute. They have a degree of control over the economic activity of that company. Should that control be taken away from them? Who should it be given to?
I can see an argument when it comes to cashing out, but I’m not clear how that should work without creating really weird incentives. Some sort of special tax?
r14c 10 minutes ago [-]
Big agree, at a certain point a company is big enough that their impact has to be managed democratically. I don't have an issue with effective leaders, the problem is that we reward a certain kind of success with transferable credits that don't necessarily align with people's actual talents or skills.
I want skilled institutional investors who have a track records of making smart bets. I don't want a random person who happened to get lucky in business dictating investment policy for substantial parts of the economy. I want accountability for abuses and mismanagement.
I know China gets a bad rep, but their bird cage market economy seems a lot more stable and predictable than this wild west pyramid scheme stuff we do in the US. Maybe there are advantages for some people in our model, but I really dislike the part where we consistently reward amoral grifters.
rafterydj 1 hours ago [-]
Well, redistributing their money is (in some cases disingenuously) exactly how they are able to pitch investors. "Sure, value my company at $10B and my shares make me $2B, but we're alllllll gonna make money when hit AGI!!!" That kind of thing.
stavros 1 hours ago [-]
I found it very interesting that Altman et al were worried that AI will become supremely intelligent and China will make a supervirus or some AI drones or whatnot, but not a single person was worried about destroying all jobs because we wouldn't need humans any more.
Or maybe they were not so much "worried" but "hopeful" that they'd amass literally all the wealth in the world.
I didn't have the mental energy to read the whole thing but man the final paragraph is some really good writing. Way to tie it all in together.
krackers 58 minutes ago [-]
The entire thing is a joy to read, you should really set aside some time to cleanse your palette in this age of LLM prose. I mean just look at this juxtaposition
>Altman continued touting OpenAI’s commitment to safety, especially when potential recruits were within earshot. In late 2022, four computer scientists published a paper motivated in part by concerns about “deceptive alignment,” in which sufficiently advanced models might pretend to behave well during testing and then, once deployed, pursue their own goals.
(plus it finally resolves the mystery of "what Ilya saw" that day)
Also since it wasn't stated clearly
>“the breach” in India. Altman, during many hours of briefing with the board, had neglected to mention that Microsoft had released an early version of ChatGPT in India
That was Sydney if I understand correctly.
the_arun 24 minutes ago [-]
The main animated picture reminded me of evil king Ravan from Ramayan with 10 heads. Not sure it is intentionally done that way.
smcg 12 minutes ago [-]
Rule of Headlines says "no"
wk_end 3 hours ago [-]
This anecdote is so absurd it sounds like satire. This is the guy with the $23M mansion?
> Amodei’s notes describe escalating tense encounters, including one, months later, in which Altman summoned him and his sister, Daniela, who worked in safety and policy at the company, to tell them that he had it on “good authority” from a senior executive that they had been plotting a coup. Daniela, the notes continue, “lost it,” and brought in that executive, who denied having said anything. As one person briefed on the exchange recalled, Altman then denied having made the claim. “I didn’t even say that,” he said. “You just said that,” Daniela replied.
simoncion 3 hours ago [-]
He's a liar and untrustworthy. Based on their public statements, that's a big part of why the board fired him.
Of course, (despite the fact that Altman previously publicly stated that it was very important that the board can fire him) he got himself unfired very quickly.
ycui1986 1 hours ago [-]
he won't. if anything, openai is falling behind recently. the trend won't change easily. it is like the old time Netscape.
just_once 11 hours ago [-]
Amazing that this article and an actual comment from Ronan Farrow is this far down the list while...Scientists Figured Out How Eels Reproduce (2022) has 6 times the points.
dang 4 hours ago [-]
This thread set off a software penalty called the flamewar detector.* I turned that off as soon as I saw it.
(* This was predictable from the title, because the question in it was inevitably going to trigger an avalanche of crap replies. Normally we'd change the title to something less baity, and indeed the article is so substantive that it deserves a considerably better one. But I'm not going to change it in this case, since the story has connections to YC - about that see https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu....)
throw4847285 7 hours ago [-]
A new Ronan Farrow piece is a rare gift (and Marantz is no slouch). Can't wait to read this in the physical magazine when it arrives!
HardwareLust 10 hours ago [-]
Of course he cannot be trusted. Anyone whose motivation is based on greed is by nature untrustworthy.
throwway120385 3 hours ago [-]
Even if your motivation is some utopian vision of the future, you should not be trusted. Utopia is a thought experiment in a philosophy of living taken too far, not something to be reached for earnestly.
hellojimbo 2 hours ago [-]
lol thats like 99% of planet earth, including the animals
383toast 2 hours ago [-]
if you have to ask if someone can be trusted, they usually can't
hungryhobbit 2 hours ago [-]
It's the golden rule of news article headlines: if the headline is a question, the answer to it is always a negative.
And if you think you don't have to ask if someone can be trusted, you're usually wrong.
brandonpollack2 31 minutes ago [-]
I haven't read it yet. The answer is no.
slg 3 hours ago [-]
One thing that stands out when reading profiles like this is the number of positive and negative descriptions of the subject that agree. For example, there seems to be little dispute that Altman will happily say something that he knows/believes isn't true, there's just a lot of people who are willing to forgive any lies if the lies are in service of something they themselves agree with.
palata 3 hours ago [-]
> there's just a lot of people who are willing to forgive any lies if the lies are in service of something they themselves agree with.
Or if the person lying is in a position of power?
innocenttop 7 hours ago [-]
Why is the story so downranked? Folks at HackerNews have something to do with it ?
dang 4 hours ago [-]
It off the flamewar detector, a,k.a. the overheated discussion detector. I've turned that off now - this is obviously a serious article.
4 hours ago [-]
randycupertino 7 hours ago [-]
HN generally downvotes and/or flags anything that paints ycombinator in a bad light. As Altman was president of yc from 2014 to 2019 that could be why this is getting downvoted.
Articles critical of Airbnb, one of yc's biggest wins, also get flagged and taken down.
dang 4 hours ago [-]
I'm not sure whether you meant this about moderator interventions or not, but our actual practice is the opposite:
I don’t think the poster you responded to was claiming that moderators directly did this. The flagging system is open to bias from the community at large and certain types of articles(ex. Anything critical of the current admin) get a bunch of real users organically flagging them.
saeranv 25 minutes ago [-]
Greg Brockman honestly sounds like a psychopath:
> In 2017, Amodei hired Page Hedley, a former public-interest lawyer, to be OpenAI’s policy and ethics adviser. In an early PowerPoint presentation to executives, Hedley outlined how OpenAI might avert a “catastrophic” arms race—perhaps by building a coalition of A.I. labs that would eventually coördinate with an international body akin to NATO, to insure that the technology was deployed safely. As Hedley recalled it, Brockman didn’t understand how this would help the company beat its competitors. “No matter what I said,” Hedley told us, “Greg kept going back to ‘So how do we raise more money? How do we win?’ ” According to several interviews and contemporaneous records, Brockman offered a counterproposal: OpenAI could enrich itself by playing world powers—including China and Russia—against one another, perhaps by starting a bidding war among them. According to Hedley, the thinking seemed to be, It worked for nuclear weapons, why not for A.I.?
40aKshl 5 minutes ago [-]
Brockman knew though that AI didn't (and still doesn't) work. So he suggested grifting.
The fact that the chatbots are now used for plausible deniability and blame shifting in Gaza or Iran wasn't known then.
pdonis 2 hours ago [-]
Does the article ever actually answer the title question?
kubik369 43 minutes ago [-]
I think you are misunderstanding the point of journalism. It can be debated whether the title should be such a question. Nevertheless, the article should just present information, ideally in a balanced way, without author's bias, so that you can decide for yourself. You can see the attempts at the balanced part in the article where an allegation/statement is made about Altman followed by parentheses saying that Altman recalls the exchange differently/does not remember.
mohamedkoubaa 2 hours ago [-]
The answer is no, he can't be trusted
pdonis 1 hours ago [-]
Oh, I agree that's the correct answer. I just don't see the article actually ending up with that answer. I see it waffling. Basically, the article ends up saying that, well, we told you about all this dodgy stuff, but what he's doing is working.
dmitrygr 2 hours ago [-]
The number of "Altman doesn’t remember this" or "Altman denies this" is hilarious
jcgrillo 33 minutes ago [-]
Life would be so much easier if I was that forgetful
panzi 39 minutes ago [-]
No. Next question.
quantified 2 hours ago [-]
A bit of a feeling of "so what" here. Maybe he's less trustworthy than some. We have people of X trustworthiness running the government, crypto exchanges, a certain space exploration and satellite company, social media companies, and so on. We know their trustworthiness. Isn't the real issue how to cope?
boc 48 minutes ago [-]
What's the point of living in an advanced society if you just sit around watching it decay around you? Our ancestors fought for our indifference today, and with attitudes like yours we'll watch our children fight for it again tomorrow.
Boxxed 2 hours ago [-]
Your point is that it's ok he's untrustworthy because lots of people in power are?
JumpCrisscross 45 minutes ago [-]
> Your point is that it's ok he's untrustworthy because lots of people in power are?
It's...weirdly a valid question. If Sam fibs as much as the next guy, we don't have a Sam problem. Focussing on him alon is, best case, a waste of resources. Worst case, it's distracting from real evil. If, on the other hand, as this reporting suggests, Sam is an outlier, then focussing on him does make sense.
TheOtherHobbes 1 hours ago [-]
No, it's that the entire ecosystem is rotten to the core, and it actively selects, rewards, and protects flawed personality types.
And when you're dealing with a potential existential threat, this is an existential problem.
12 hours ago [-]
ProAm 1 hours ago [-]
Nope, never trust this man. His history proves why you cannot. Pure greed.
If you are asking if a single human can be trusted with such a responsibility, the answer is, by default, no.
Arubis 1 hours ago [-]
This is unfair to the original article, which is well-researched and worth a read. But the answer this question is _always_ no. Nobody should have as much power as the oligarch class currently does, even if of inscrutable power.
y1n0 28 minutes ago [-]
Betteridge's law of headlines: no
game_the0ry 5 hours ago [-]
For those curious about how sama got to where he got and stayed on top for so long, I recommend you read the book: The Sociopath Next Door by Martha Stout.
I am fairly confident when I say this -- sama is a sociopath. I don't know how anyone with solid intuition could even come to any other conclusion than the guy is deeply weird and off-putting.
Some concepts from the book:
> Core trait: The defining characteristic is the absence of conscience, meaning they feel no guilt, shame, or remorse.
> Identification: Sociopaths can be charming and appear normal, but they often lie, cheat, and manipulate to get what they want.
> The Rule of Threes: One lie is a mistake, two is a concern, but three lies or broken promises is a pattern of a liar.
> Trust your instincts over a person's social role (e.g., doctor, leader, parent)
Check and check.
OpenAI is too important to trust sama with. He needs to go. In fact, AI should be considered a public good, not a commodity pay-as-you-go intelligence service.
jcgrillo 25 minutes ago [-]
I was with you right up until the final paragraph, but this made me do a double take:
> OpenAI is too important to trust sama with.
...wat? They made a chat bot. How can that possibly be so existentially important? The concept of "importance" (and its cousin "danger") has no place in the realistic assessment of what OpenAI has accomplished. They haven't built anything dangerous, there is no "AI safety" problem, and nothing they've done so far is truly "important". They have built a chat bot which can do some neat tricks. Remains to be seen whether they'll improve it enough to stay solvent.
The whole "super serious what-ifs" game is just marketing.
unsupp0rted 4 hours ago [-]
I suspect there's some other category, which isn't really a sociopath and isn't really a not-sociopath, which we don't have a good definition for.
We only say a lot of CEOs are sociopaths because they're in that third category we haven't named, where they're very good at manipulating people, but also can feel conscience, guilt, remorse, etc, perhaps just muted or easier to justify against.
E.g. if you think you're doing something for the betterment of mankind, it doesn't really matter if you lie to some board members some year during the multi-decade pursuit.
game_the0ry 13 minutes ago [-]
> I suspect there's some other category, which isn't really a sociopath and isn't really a not-sociopath, which we don't have a good definition for.
There is -- I call it "corpo sociopath." The corpo sociopath really comes out in the workplace, less so in personal life.
xg15 2 hours ago [-]
That's not a third category, that's just a sociopath as seen by themself.
unsupp0rted 1 hours ago [-]
I doubt most sociopaths, when they’re honest, would agree they feel much guilt or remorse at all.
Whereas the people in the category I’m describing might feel those things, but prioritize those feelings far below the benefits of achieving what they set out to achieve.
game_the0ry 13 minutes ago [-]
> I doubt most sociopaths, when they’re honest, would agree they feel much guilt or remorse at all.
Yes that is the core trait I highlighted in the 1st bullet.
KellyCriterion 4 hours ago [-]
Na, it will be Dario instead of Sam, Id say? :-))
2 hours ago [-]
almostdeadguy 8 hours ago [-]
Seems this got buried from the front page very quickly
dang 4 hours ago [-]
It set off the flamewar detector. I've turned that off now.
I only saw this thread by chance and almost didn't look, because the title made the piece sound like a flamebait blog post. Fortunately I saw newyorker.com beside the title and looked more closely.
ronanfarrow 8 hours ago [-]
There is dwindling space for sincere independent accountability reporting on big tech like this to a) be created, since it's incredibly resource-intensive and so many resources flow from Silicon Valley, and b) actually reach people, since more platforms are now owned or otherwise influenced by interested parties.
Thank you for looking. Please do spread this kind of reporting in your communities, and subscribe to investigative outlets when you can.
walterbell 2 hours ago [-]
> OpenAI has closed many of its safety-focussed teams
A paper with "ideas to keep people first" was (coincidentally?) published today:
• Worker perspectives
• AI-first entrepreneurs
• Right to AI
• Accelerate grid expansion
• Accelerate scientific discovery and scale the benefits.
• Modernize the tax base
• Public Wealth Fund
• Efficiency dividends
• Adaptive safety nets that work for everyone
• Portable benefits
• Pathways into human-centered work
This was an excellent piece with many new pieces of information in it. Thanks to you and your coauthor for getting it released.
big_toast 7 hours ago [-]
You can see the vote history here[1]. It's always hard to know exactly why something gets buried. I was a little sad to see the story down-ranked when I saw that you were here in the comments.
But the discussion is generally pretty low quality with these sort of posts. People react without having read the story, or with whatever was on their mind already, or are insubstantive, or simply low effort. I don't think you'll lose k-factor not having a bigger post here.
Sometimes if you talk to the mods, they'll let you know their perspective. I generally find they're correct that people are much better at contributing/disseminating new knowledge to the world on more technical topics here.
Yes, I was surprised that it was downranked when I saw that too. Then I realized it had set off the flamewar detector and it was a simple matter to turn it off. I'm glad we got to this in time, because sometimes we don't, and this was an important case not to miss.
throw4847285 5 hours ago [-]
But isn't that circular? If the ranking algorithm used by the mods tends to devalue articles like this because they don't trust the user base to comment intelligently, doesn't that alter the culture of this site to make that more true?
dang 3 hours ago [-]
I'm not sure what big_toast meant, but we do trust the user base to comment intelligently (which sometimes works and sometimes not), and we don't devalue articles like this.
We do tend to devalue titles like this, or more likely change them to something more substantive (preferably using a representative phrase from the article body), but I'm worried that if I did that here we would get howls of protest, since YC is part of the story.
throw4847285 1 hours ago [-]
I'm sure you're sick of comments about moderation, but I will say, this makes me more sympathetic to the position you're in.
It's an interesting dilemma. Many very respected publications use provocative titles because of the attention economy. And I'm sure you have good data that provocative titles lead to drive-by comments and flame wars.
But I don't think big_toast was entirely wrong that there is a side effect of sometimes burying articles that are by their nature provocative. And how do you distinguish a flame war over a title from a flame war over content? That's not a leading question. I don't know.
4 hours ago [-]
9 hours ago [-]
jader201 3 hours ago [-]
Am I the only one that feels like Claude is clearly winning code generation, and Gemini in general LLM?
I just don’t feel like OpenAI has a legitimate shot at winning any of the AI battles.
Therefore, I feel like “Sam Altman may control our future” is a far stretch.
guelo 3 hours ago [-]
Well I just canceled my Claude Pro subscription because of the mysterious limits that I don't experience with codex, even after paying for "extra usage". If Anthropic can't figure out their capacity problems they are in trouble.
chrisjj 2 hours ago [-]
I doubt Anthropic see this as their capacity problem. They like "extra usage", and users who don't, well its their capacity problem.
dominotw 3 hours ago [-]
how is gemini winning in general llm. what is general llm .
SwellJoe 3 hours ago [-]
General LLM is what Apple is paying Google for.
gambiting 2 hours ago [-]
>>and Gemini in general LLM?
You might be. Or at least I feel like Gemini is actually dumber than a house of bricks - I have multiple examples, just from last week, where following its advice would have lead to damage to equipment and could have hurt someone. That's just trying to work on an electronics project and askin Gemini for advice based on pictures and schematics - it just confidently states stuff that is 100000% bullshit, and I'm so glad that I have at least a basic understanding of how this stuff works or I would have easily hurt myself.
It's somewhat decent at putting together meal plans for me every week, but it just doesn't follow instructions and keeps repeating itself. It hardly feels worth any money right now, like it's some kind of giant joke that all these companies are playing on us, spending billions of these talking boxes that don't seem that intelligent.
I also use claude at work, and for C++ programming it behaves like someone who read a C++ book once and knows all the keywords, but has never actually written anything in C++ - the code it produces is barely usable, and only in very very small portions.
Edit: I just remembered another one that made me incredibly angry. I've been reading the Neuromancer on and off, and I got back into it, but to remind myself of the plot I asked Gemini to summarise the plot only up to chapter 14, and I specifically included the instruction that it should double check it's not spoiling anything from the rest of the book. Lo and behold, it just printed out the summary of the ending and how the characters actions up to chapter 14 relate to it. And that was in the "Pro" setting too. Absolute travesty. If a real life person did that I'd stop being friends with them, but somehow I'm paying money for this. Maybe I'm the clown here.
therobots927 12 hours ago [-]
Excellent work. I’ll have to wait until we get the print version delivered to finish as I’m not signed into the new Yorker on my phone.
I’ve always been a huge fan of Ronan Farrow’s journalism and willingness to speak truth to power. I think he’s pulling at exactly the right thread here, and it’s very important to counteract Altman’s reputation laundering given that we run a very real risk of him weaseling his way into the taxpayer’s wallet under the current administration.
I suspect that they are perfectly capable of clicking an archive link or better yet logging in as they are already a subscriber. Maybe, like me, they enjoy reading the physical magazine.
zoklet-enjoyer 1 hours ago [-]
I believe Annie Altman.
s5300 55 minutes ago [-]
[dead]
primer42 2 hours ago [-]
"Any headline that ends in a question mark can be answered by the word no."
Can Sam "The board can fire me, I think that's important." Altman be trusted?
If for no other reason, given what happened when the board fired him... no. I'd say not.
jesterson 10 hours ago [-]
Watch Altman's reaction in Tucker Carlson interview to the question about (alleged) murder of OpenAI researcher Suchir Balaji.
The overall response and particularly the body language speaks a lot.
lnenad 12 hours ago [-]
This whole situation goes to show that yesterday's conspiracy theorists are today's realists. What's happening to USA's leadership and as a country and what's happening with with their top companies is really scary for the rest of us. If this trend continues we're all definitely gonna end up in a kleptocracy.
asK1ajsh 3 hours ago [-]
The New Yorker is owned by Conde Nast just as Reddit. Conde Nast has a deal with OpenAI:
This is a damage control piece, and you see that the most stinging comments here get downvoted.
cake_robot 1 hours ago [-]
What might feel like "damage control" is more likely to be the outcome of the even-handedness you get with serious, rigorous reporting. Something the New Yorker is known for.
nickphx 2 hours ago [-]
speak for yourself, he doesnt control my future.
vntok 1 hours ago [-]
Please don't leave us hanging; what makes you immune?
GlibMonkeyDeath 9 hours ago [-]
Disclaimer: I have no association with any AI company and have never met Altman or any of the other top AI scientists.
The real question is: can anyone be trusted if the fever dreams of super-intelligence come true? Go ahead and replace Sam Altman with someone else - will it make a difference? Any other CEO is going to be under the same overwhelming pressure to make a profit somehow. I think the OpenAI story is messier because it was founded for supposedly altruistic reasons, and then changed.
Methinks many of Altman's detractors protesteth too much. He's doing his job as it is defined (make OpenAI profitable.) Nothing of substance in this article seemed to make him exceptionally "sociopathic" compared to any other tech CEO. It goes with the territory.
What depressed me most is that trillions of dollars are being raised for building what will undoubtedly be used as a weapon. My guess is the ROI on that money is going to be extremely bad for the most part (AI will make some people insanely rich, but it is hard to see how the big investors will get a return.) Could you imagine if the world shared the same vision for energy infrastructure (so we could also stop fighting wars over control of fossil fuels and spewing CO2?) A man can dream...
tim333 6 hours ago [-]
People do vary even if none are perfect. Demis Hassabis has a pretty good reputation amongst the AI leaders. Altman seems unusually shifty.
thm 14 hours ago [-]
Hybris.
aduty 2 hours ago [-]
LOL, no.
guzfip 7 hours ago [-]
> Lehane—whose reported motto, after Mike Tyson, is “Everyone has a game plan until you punch them in the mouth”
lol do you think these guys have ever been hit? Let alone in the face. They’d probably be less eager to mouth off as much as they do if so.
Aboutplants 12 hours ago [-]
Seeing Sam Altman slowly degrade into the realization that he is in fact not as smart as others in this space has been fascinating to watch. He used to speak with enthusiasm and confidence and now he’s like a scared little boy who got in way too deep.
The last person that this happened to was Sam Bankman Fried as investors and regular folk finally realized he was full of complete shit and could only talk the game for so long until the truth emerged.
the_doctah 10 hours ago [-]
And they both peddle the same altruism smokescreen. Sociopath leader playbook.
therobots927 12 hours ago [-]
[flagged]
jjtheblunt 4 hours ago [-]
which of the two are you referring to as possibly angling for a pardon?
Findecanor 3 hours ago [-]
Bankman-Fried has already done it.
throwawayq3423 4 hours ago [-]
I have a feeling he might be angling for a pardon if he ends up bringing the whole global economy down.
jojobas 2 hours ago [-]
The guy called out for being a sociopath by a multitude of Silicon Valley CEOs of all people, sure we can trust him our future.
seba_dos1 12 hours ago [-]
Looks like Betteridge's law of headlines applies here too.
ambicapter 3 hours ago [-]
> The day that Altman was fired, he flew back to his twenty-seven-million-dollar mansion in San Francisco, which has panoramic views of the bay and once featured a cantilevered infinity pool, and set up what he called a “sort of government-in-exile.” Conway, the Airbnb co-founder Brian Chesky, and the famously aggressive crisis-communications manager Chris Lehane joined, sometimes for hours a day, by video and phone. Some members of Altman’s executive team camped out in the hallways of the house. Lawyers set up in a home office next to his bedroom. During bouts of insomnia, Altman would wander by them in his pajamas. When we spoke with Altman recently, he described the aftermath of his firing as “just this weird fugue.”
These sociopaths are so good at giving away nothing. He managed to engender sympathy instead of saying "I'm not gonna talk about anything that happened then".
Also very weird how many of these people are so deeply-linked that they'll drop everything they're doing just to get this guy back in power? Terrifying cabal.
It is, at best, incredibly hard to accumulate that much wealth without doing shady things. Microsoft's monopolistic practices in the 90s for example. The only person I can think of that ever cracked a billion without their money coming through dirty means was, funny enough, JK Rowling who has her own set of issues separate from the value she got out of Harry Potter.
balls187 3 hours ago [-]
John Lithgow had a take I agreed with: Her opinions were heavily misconstrued though she chose to double down at her own peril.
i7l 12 hours ago [-]
I feel the "always have been" meme might be a suitable insert here.
aleph_minus_one 11 hours ago [-]
>
Why are all billionaires (especially tech) such villains?
Not all billionaires are villians. But it is long-known in organizational psychology that dark triad [1] traits are very "helpful" if one wants to climb career ladders fast.
I'm not 100% sure if it's strictly necessary to be a villain in order to become and remain a billionaire, but it seems like it could be and even if it's not it surely helps.
burnt-resistor 11 hours ago [-]
Money often changes people's attitude in a fashion similar to chronic substance abuse. Plus, there's a insular and detached bubble effect that grows around them.
Also, there's the psychopathic and narcissistic tendencies of greedier people and the false "virtue" "greed is good" that is contrary to the values espoused by Adam Smith.
We need standard income tax brackets of 90% after $20M/y and 99% after $100M/y.
romeroej 4 hours ago [-]
Can anybody tho?
morleytj 3 hours ago [-]
Yeah, some people can more than others.
neya 12 hours ago [-]
[flagged]
12 hours ago [-]
1 hours ago [-]
FpUser 3 hours ago [-]
>"Sam Altman may control our future"
TLDR but just the heading is already ugly. No single person no matter how nice they're should be able to control our future. Power corrupts, what fucking trust. We are supposed to be democratic society (well looking at what is going on around this is becoming laughable)
josefritzishere 11 hours ago [-]
Betteridge's law of headlines is an adage that states: "Any headline that ends in a question mark can be answered by the word "NO."
gchokov 12 hours ago [-]
He is cooked. Only a matter of time before the whole thing blows up.
Once a scammer, always a scammer.
sumeno 12 hours ago [-]
Betteridge strikes again
Cheyana 13 hours ago [-]
Harvey Dent…
the_doctah 10 hours ago [-]
The brighter the picture, the darker the negative
LetsGetTechnicl 12 hours ago [-]
No
gonzo41 12 hours ago [-]
just like Zuck.
nielsbot 3 hours ago [-]
No one person control our future. Stop there.
_moof 3 hours ago [-]
Some people have far, far more power over our lives than others. More than they deserve, frankly.
mikkupikku 3 hours ago [-]
Yeah, but one person can fuck a lot of shit up.
killbot5000 3 hours ago [-]
No. Why is this a question?
catigula 12 hours ago [-]
1. No.
2. You cannot "control" superintelligent AI.
ekjhgkejhgk 12 hours ago [-]
No.
aksss 4 hours ago [-]
"could", "may", "might" - these words do so much heavy lifting in "journalism". Almost always it's an invitation to worry and be miserable.
arionhardison 3 hours ago [-]
Hi @ronanfarrow — I have only had one interaction with Sam Altman in person, and I was advised to keep it to myself. I know this crowd may not care, but Altman is absolutely terrified of Black people — not in any contextual sense, but in a visceral, instinctive way. For someone who, as you put it, "controls our future," this should matter.
FYI: I am by far not the only one to have experienced this and it 100% impacts hiring and other decisions at OpenAI.
edbaskerville 3 hours ago [-]
Can you give more details?
It wouldn't particularly surprise me if Sam Altman were racist, but I'm curious what the specific incident you observed was.
arionhardison 2 hours ago [-]
Yes, but first I want to be very clear on some things.
1. I could have hidden my identify behind a throwaway. I did not feel that would be appropriate when making this calim.
2. I am not looking for anything, literally at all. Any follow ups for blogs; anything that would benefit I will not answer.
3. This is NOT a new account, I am very easy to find; I am 6'1 140lbs
I was working for a company called NationBuilder and I had the opportunity to go on a work trip. Outside of a talk he had just given I was waiting for my ride and I looked over like...damn thats the speaker. I wanted to say Hi; he damn near flagged down the police. I apologized and just decided to move on.
Note: It was in Reno, and no I don't want to go into details; the others are not hard to find because I happened upon them via blog posts so i'm sure if someone with the accumen of RF wants to know, he will find.
I have heard similar stores from several people in the years since. I AM NOT CALLING THIS PERSON RACIST. I am saying; he is observably scared of black people and that is not someone I want making descions about how the world moves foward.
arionhardison 2 hours ago [-]
Note: To all the downvotes; I did this publicly and not anon for a reason, if you will do the same I am more than willing to provide evidence for all of these claims as long as its done publicly and in the open.
arionhardison 53 minutes ago [-]
PG said something along the lines of: "There should be no truth that is increasingly unpopular to speak."
If you don't believe what I shared is true, address that directly. But seeing my post sitting at 1 point and [flagged] after 2 hours is not OK. Just as DJT can't flag away his issues, you shouldn't be able to do so on HN.
One of the things I've loved most about HN is that it was real — grounded in observability, empirical evidence, not bias or feelings. I really hope that what happened to my post is not the beginning or a continuance of the end for that ethos.
bijowo1676 3 hours ago [-]
This article is just another typical New Yorker fluff piece that tries to look deep but misses the actual point.
The biggest flaw is that it spends way too much time on high-school level drama and "he-said-she-said" gossip about Sam Altman’s personal life instead of focusing on the actual technical and corporate capture of OpenAI.
The author treats the "nonprofit mission" like some holy quest that was "betrayed," when anyone with a brain in tech saw the Microsoft deal as the moment the original vision died. Instead of a hard-hitting look at how compute-monopolies are actually forming (MSFT AMZN NVDA and circular debt dealing inflating the AI bubble that could crash the economy), we get 5,000 words of hand-wringing over whether Sam is a "nice guy" or a "liar."
Who cares???????
The board failed because they had no real leverage against billions of dollars, not because they didn't write enough Slack messages. It's a long-winded way of saying "Silicon Valley has internal politics," which isn't news to anyone here.
ninjahawk1 3 hours ago [-]
OpenAI is like #3 or #4 of the AI companies right now in terms of power, and last place in the court of public opinion.
I’d be more concerned about Anthropic both being in the good graces of the public and having access to all of our computers indirectly with Claude Code.
0x3f 3 hours ago [-]
OpenAI has ~30x the userbase of Anthropic.
aduffy 2 hours ago [-]
I'm not sure how much of that converts to revenue. If it's free plan users, that's just cost. You can say what you want about "creating a training data moat" but that doesn't seem like it's prevented the other labs from putting out excellent models.
0x3f 2 hours ago [-]
Well we were talking about power and reputation and being well-known and all that. Being more ubiquitous is surely a big part of that. GP seems to think Anthropic is doing better because of the DoD thing. In my estimation, 90% of people do not care about that at all.
ninjahawk1 2 hours ago [-]
They’re all in the negative excluding subsidies, hard core coders are more valuable than high schoolers cheating on homework.
hellojimbo 2 hours ago [-]
Around the same revenue due to Anthropics strong enterprise strategy
0x3f 2 hours ago [-]
Perhaps, but I'd venture the ear of the regime is even more valuable.
estearum 3 hours ago [-]
makes sense if you think the point of journalism is just to take everyone down a notch instead of... um... informing the public of bad actors
"the local drug-dealing pimp is so passe, we need to investigate the most upstanding members of the community just to be sure" is a frankly insane strategy
cm2012 2 hours ago [-]
I don't see anything bad about Altman in this article that cant be explained by the chaos of growing a billion dollar company in a few years.
Rendered at 01:21:47 GMT+0000 (Coordinated Universal Time) with Vercel.
All evidence today suggests Anthropic is passing OpenAI in relative and absolute growth. So where's the critical reporting? The DOD coverage was framed around the Pentagon's decisions, not Anthropic's. And nobody seems interested in examining whether the company that branded itself as the ethical AI lab actually is one. That seems like a story worth writing.
FWIW I have two(!!) close friends working for Anthropic, one for nearly two years and one for about 4 months.
Both of them tell me that this is not just marketing, that the company actually is ethical and safety conscious everywhere, and that this was the most surprising part about joining Anthropic for them. They insist the culture is actually genuine which is practically unicorn rarity in corporate America.
We have worked for FAANG so I know where they're coming from; this got me to drop my cynicism for once and I plan on interviewing with them soon. Hopefully I can answer this question for myself.
You have a point in that Anthropic deserves some coverage too and that there are interesting perspectives that we've not heard of on that front either.
But just because that's true doesn't mean this article isn't very much relevant and needed.
Because it is.
Now, they may have heard the word "Anthropic" due to recent media coverage. But they don't know what it is and don't remember what it makes. The fact that all businesses use "Anthropic" is about as relevant to them as knowing the overseas shipping company for all the shit they buy off Amazon.
So articles about OAI will always produce more revenue for the media, because it's related to what normies actually use day to day.
As a reader, am I supposed to infer anything about evidentiary weight from these stylistic choices? When a single anonymous source's testimony is presented in a "declarative" narrative style like here (with the attribution in a less prominent position), should we read that as reflecting high confidence on your end (perhaps from additional corroboration not fully spelled out)? And does the fact that Altman’s non-recollection appears in parentheses carry any epistemic signal (e.g. that you assign it less evidentiary weight)? Or is that mostly a matter of (say) prose rhythm?
she feels very frustrated that she was dragged through the mud, with the WSJ and Satya Nadella saying she wasn't qualified for being on the board, when in fact she was totally right, along with the other PhDs, that Sam Altman - and for that matter Greg Brockman - had to go.
everybody else has the wisdom to not talk to Ronan Farrow.
My question is, how do you know when an enormous project like this, conducted over an 18-month time span is "done"? I assume you get a lot of leeway from editors and publishers on this matter. How do you make the decision to finally pull the trigger on publishing?
This is a fantastic piece, very timely, evidently well-researched, and also well-written. Judging by the little that I know, it's accurate. Thank you for doing the work and sharing it with the world.
OpenAI may be in a more tenuous competitive position than many people realize. Recent anecdotal evidence suggests the company has lost its lead in the AI race to Anthropic.[b]
Many people here, on HN, who develop software prefer Claude, because they think it's a better product.[c]
Is your understanding of OpenAI's current competitive position similar?
---
[a] You may want to provide proof online that you are who you say you are: https://en.wikipedia.org/wiki/On_the_Internet%2C_nobody_know...
[b] https://www.latimes.com/business/story/2026-04-01/openais-sh...
[c] For example, there are 2x more stories mentioning Claude than ChatGPT on HN over the past year. Compare https://hn.algolia.com/?dateRange=pastYear&page=0&prefix=tru... to https://hn.algolia.com/?dateRange=pastYear&page=0&prefix=tru...
The piece captures some of the anxieties within OpenAI right now about their competitive position. This obviously ebbs and flows but of late there has been much focus on Anthropic's relative position. We of course mention the allegations of "circular deals" and concerns about partners taking on debt.
I was asking more about your informed view on how OpenAI's technology, products, and roadmap are perceived, particularly by customers and partners, in comparison to those of competitors.
If you have an opinion about that, everyone here would love to hear about it.
He's had so many conversations that he likely has a sense of how perceptions of the company and its offerings have changed.
I'm curious.
As it turns out, and what I’m kind of going with for this LLM shit, is that it’ll play out exactly how you think it will. The companies are all too big to fail, with billionaire backers who would rather commit fraud than lose money.
No comment on the CEO: I just find the product superior in everything but UI/UX and conversation. It's better at quality code.
Both codex and Claude code fail when it comes to extremely sophisticated programming for distributed systems
For the few times I've used both models side by side on more typical tasks (not so much web stuff, which I don't do much of, but more conventional Python scripts, CLI utilities in C, some OpenGL), they seem much more evenly matched. I haven't found a case where Claude would be markedly superior since Codex 5.2 came out, but I'm sure there are plenty. In my view, benchmarks are completely irrelevant at this point, just use models side by side on representative bits of your real work and stick with what works best for you. My software engineer friends often react with disbelief when I say I much prefer Codex, but in my experience it is not a close comparison.
Gemini seems to be the worst of the three, and some open-weight models are not too bad (like Kimi k2.5). Cursor is still pretty good, and copilot just really really sucks.
I enjoy using CC more and use it for non coding tasks primarily, but for anything complex (honestly most of what I do is not that complex), I feel like I am trading future toil for a dopamine hit.
LLMs aren't able to achieve 100% correctness of every line of code. But luckily, 100% correctness is not required for debugging. So its better at that sort of thing. Its also (comparatively) good at reading lots and lots of code. Better than I am - I get bogged down in details and I exhaust quickly.
An example of broad work is something like: "Compile this C# code to webassembly, then run it from this go program. Write a set of benchmarks of the result, and compare it to the C# code running natively, and this python implementation. Make a chart of the data add it to this latex code." Each of the steps is simple if you have expertise in the languages and tools. But a lot of work otherwise. But for me to do that, I'd need to figure out C# webassembly compilation and go wasm libraries. I'd need to find a good charting library. And so on.
I think its decent at debugging because debugging requires reading a lot of code. And there's lots of weird tools and approaches you can use to debug something. And its not mission critical that every approach works. Debugging plays to the strengths of LLMs.
https://xcancel.com/RonanFarrow/status/2041127882429206532#m
https://x.com/RonanFarrow/status/2041127882429206532
Unfortunately it probably doesn't even matter here on HN considering how brigaded down this story is predictably getting.
But yeah, it was a fantastic piece.
Did you do any extra investigations into Annie’s allegations? It feels to me like the unstated conclusion is recovered memory can’t be trusted, which is a popular understanding but a very wrong one put out by the now defunct and discredited False Memory Syndrome Foundation. It was founded by the parents of the psychologist who coined DARVO, directly in reaction to her accusing them of abuse.
Dissociation is real (I have a dissociative disorder, and abuse I “recovered” but did not remember for much of my adolescence and early adulthood has been corroborated by third parties) and many CSA survivors have severe memory problems that often don’t come to a head until adulthood. I know you didn’t dismiss her claim, but the way the public tends to think about recovered memories is shaped primarily by that awful organization.
As noted in the piece, we spent months talking to Altman's partners and what we found and didn't is as described.
> “Investors are, like, I need to know you’re gonna stick with this when times get hard,”
Should be:
> “Investors are like, I need to know you’re gonna stick with this when times get hard,”
I would really suggest subscribing to and finding ways to amplify independent outlets and journalists, and encouraging others to do so.
Altman describes his shifting views as genuine good faith evolution of thinking. Do you believe he has a clear North Star behind all this that’s not centered on himself?
My own impression after many hours of conversation is that he is identifying something of a true north star when he frames this around "winning." There are people in the story who talk about him emphasizing a desire for power (as opposed to, say, wealth). I think he probably also believes, to some extent, the story he tells that equates winning, and his gaining power, with a superabundant utopian future for all.
However, I think critics correctly highlight a tension between his statements about centering humanity writ large and his tilt into relentless accelerationism.
Please try to give people the benefit of the doubt though I know it's hard in today's society.
Or Mr Farrow can you post some evidence somewhere we can see?
I'm not pleased with the headline and the general framing that AI works. The plagiarism and IP theft aspects are entirely omitted. The widespread disillusion with AI is omitted.
On the positive side, the Kushner ad Abu Dhabi involvements (and threats from Kushner) deserve a wider audience.
My personal opinion is that "who should control AI" is the wrong question. In the current state, it is an IP laundering device and I wonder why publications fall silent on this. For example, the NYT has abandoned their crown witness Suchir Balaji who literally perished for his convictions (murder or not).
I would love to read your piece and pay you and new Yorker for it, but I am not interested in paying a subscription. If I could press a button and pay a reasonable one time license such as $3 or $5 for just this article, or better yet a few cents per paragraph as they load in, I wouldn't hesitate.
However I'm not going to pay for yet another subscription to access one article I'm interested in.
I'm sure you can't do anything about this, but I just wanted you to know.
You deserve to be compensated for great journalism. In this case, unfortunately, I won't read it and you won't earn income from me.
You're already paying for your library with your tax dollars. If you don't use it, you may lose it, but you will certainly lose out by subsidizing bums, vagrants, and other families who use the library to their heart's content.
The public library also features lots of streaming and CD music, videos, and video games, that you can freely check out without any cost. In fact, my local library staff told me that they've abolished overdue fees. Libby and the digital apps will automatically renew or return materials. My physical books even got auto-renewed three times before I needed to manually do it, or bring them back into the building.
I can't imagine having such uninspired thoughts and actually writing them down while in a role of such diverse and worthwhile opportunities. I'd like to ask "how the hell do these people find themselves in these positions", but I think the answer is literally what he wrote in his diary. What a boring answer. We need to filter these people out at every turn, but instead they're elevated to the highest peaks of power.
The difference isn't that the average techie doesn't dream of making a billion by any means necessary; it's that most of us don't think we have a shot, so we stick to enabling lesser evils to retire with mere millions in the bank.
I hope that's not true. If it is, we live in a bleak world indeed.
I can confidently say I've never once dreamed of having billions. I've never wanted billions. Not even in a fanciful manner. What would I do with that money? Buy mansions and megayachts? That's loser stuff
Most of what I want out of life cannot be bought. The pieces that come with a price tag, like a comfortable home, do not require billions
I think only sociopaths want billions because they don't understand spending your life seeking things that actually matter, like family and human connection
You need to accept that every generation some people are going to try and fuck things up.
Then you get to decide to stop or help them
Its up to rest of society to keep them in check since classic morals are highly optional and considered nuissance blocking those games. And here we the rest fail pretty miserably, while having on paper perfect tool - majority vote.
No need to be petty. They have a point. We did this with the words racist and fascist. Overinclusion diluted the term and gave cover for the actual baddies to come in. I'm not sure debating who is and isn't a sociopath is as useful as, say, the degree to which Sam is a liar (versus visible).
> "Employees now call this moment “the Blip,” after an incident in the Marvel films in which characters disappear from existence and then return, unchanged, to a world profoundly altered by their absence."
This is a very small detail, but an instinctive grimace crosses my face at the thought of these sort of Marvel references and I'm not entirely sure why.
The people shaping the future have no taste.
Is it cynical to want your <art project> to make a profit? Or for it to make enough profit to subsidize other projects?
Is it cynical to make something accessible so more people who watch it are able to enjoy it?
I agree that it's embarrassing and feels crass when movies both try to be broadly appealing and simultaneously fail to be entertaining or well executed ... but many of the marvel movies clearly surpass that bar.
No one wants to make a bad movie that does poorly with critics and paying customers - but it does happen because making a movie is expensive and complicated and requires a lot of skilled people working together towards the same goal.
Regarding taste: do you think a michelin star chef swears off cheap food like hotdogs or fish and chips? Doubtful - because those foods have their place and the chef is able to enjoy them for what they are rather than use them as an excuse to display a superiority complex.
Yeah, I'm saying professional communication isn't the place for Marvel references, and that those who choose to include references to those movies in their professional communications are revealing something about their media tastes.
Thus it is a writer's job not to make references they find appealing to reveal their good taste, but to know what references their audience will find appealing and use them to help communicate concepts. If this bothers you it's because they're insulting you by saying you might be part of the audience that watches Marvel, and you had hoped reading the New Yorker would signal that you aren't.
Sam Altman clearly has a long history of nefarious activity. But the underlying threat posted by AI to society, the economy and human freedom persists with or without his presence.
This technology needs to become a commodity to destroy this aggregation of power between a few organizations with untrustworthy incentives and leadership.
And we’re just talking about cognition - it completely ignores the automatic processes such as maintaining and regulating the body and it’s hormones, coordinating and maintaining muscles, visual/spacial processing taking in massive amounts of data at a very fine scale, and informing the body what to do with it - could go on.
One of the more annoying things about this conversation is you don’t even need to make this argument to make the point you’re trying to make, but people love doing it anyway. It needlessly reduces how amazing the human brain is to a bunch of catchy sci fi sounding idioms.
It can be simultaneously true that transformer based language models can be very smart and that the human brain is also very smart. It genuinely confuses me why people need to make it an either/or.
If they discover the cure to cancer, I don't care how they did it. "I don't trust anyone who claims they're intelligent" doesn't follow from "all they do is <how they work>".
It kind of does if how they work is nothing like genuine intelligence. You can (rightly) think AI is incredible and amazing and going to bring us amazing new medical technologies, without wrongly thinking its super amazing pattern recognition is the same thing as genuine intelligence. It should be worrying if people begin to believe the stochastic parrot is actually wise.
That's not a cure. Like yes, I'd care if the AI says it cures cancer while nuking Chicago. But that isn't what OP said.
Fantastic reporting.
Perhaps you mean to distinguish social groups without much intimacy? To which I'm sure we could provide some convincing cases, but this seems like a silly heuristic generally.
Someone cheating regularly on their partner, flagrant substance use problems, controlling people who ostracize anyone who doesn't agree with their sometimes insane perspectives...
People will go along with quite a lot to avoid friction, especially as they get older and picking up new social circles becomes higher cost.
It's possibly the most telling thing, when you see what people say is a hard line versus how they actually respond to it.
In your investigation were you able to determine if Altman has similar proxies?
How common would you say that this is? Do these kinds of people generally have teams of people who sling mud for them?
Can you speculate on how that manifests on a site like Hackernews?
https://harpers.org/archive/2026/03/childs-play-sam-kriss-ai...
The main reason is that he gets all the downsides without the upsides. I know $5B is a lot but, for a 700B company, it isn't. If OpenAI was a regular for-profit, he would have been worth >$100B already.
This is probably one of the significant factors why other co-founders left too. It's just a lot of headaches with relatively low reward.
All the downsides without much upside...
Sergey Brin is trying to change that lately, but Altman still has a sizable head start.
The fact that some (usually toxic) individuals get there shows that the system is flawed.
The fact that those individuals feel like they can do anything other than shut up, stay low and silently enjoy the fact that they got waaaay too much money shows that the system is very flawed.
We shouldn't follow billionaires, we should redistribute their money.
I can see an argument when it comes to cashing out, but I’m not clear how that should work without creating really weird incentives. Some sort of special tax?
I want skilled institutional investors who have a track records of making smart bets. I don't want a random person who happened to get lucky in business dictating investment policy for substantial parts of the economy. I want accountability for abuses and mismanagement.
I know China gets a bad rep, but their bird cage market economy seems a lot more stable and predictable than this wild west pyramid scheme stuff we do in the US. Maybe there are advantages for some people in our model, but I really dislike the part where we consistently reward amoral grifters.
Or maybe they were not so much "worried" but "hopeful" that they'd amass literally all the wealth in the world.
>Altman continued touting OpenAI’s commitment to safety, especially when potential recruits were within earshot. In late 2022, four computer scientists published a paper motivated in part by concerns about “deceptive alignment,” in which sufficiently advanced models might pretend to behave well during testing and then, once deployed, pursue their own goals.
(plus it finally resolves the mystery of "what Ilya saw" that day)
Also since it wasn't stated clearly
>“the breach” in India. Altman, during many hours of briefing with the board, had neglected to mention that Microsoft had released an early version of ChatGPT in India
That was Sydney if I understand correctly.
> Amodei’s notes describe escalating tense encounters, including one, months later, in which Altman summoned him and his sister, Daniela, who worked in safety and policy at the company, to tell them that he had it on “good authority” from a senior executive that they had been plotting a coup. Daniela, the notes continue, “lost it,” and brought in that executive, who denied having said anything. As one person briefed on the exchange recalled, Altman then denied having made the claim. “I didn’t even say that,” he said. “You just said that,” Daniela replied.
Of course, (despite the fact that Altman previously publicly stated that it was very important that the board can fire him) he got himself unfired very quickly.
(* This was predictable from the title, because the question in it was inevitably going to trigger an avalanche of crap replies. Normally we'd change the title to something less baity, and indeed the article is so substantive that it deserves a considerably better one. But I'm not going to change it in this case, since the story has connections to YC - about that see https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu....)
Or if the person lying is in a position of power?
Articles critical of Airbnb, one of yc's biggest wins, also get flagged and taken down.
https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu...
As those comments explain, this has been the #1 rule of HN moderation from the beginning. See also https://hn.algolia.com/?dateRange=all&page=0&prefix=true&que....
> In 2017, Amodei hired Page Hedley, a former public-interest lawyer, to be OpenAI’s policy and ethics adviser. In an early PowerPoint presentation to executives, Hedley outlined how OpenAI might avert a “catastrophic” arms race—perhaps by building a coalition of A.I. labs that would eventually coördinate with an international body akin to NATO, to insure that the technology was deployed safely. As Hedley recalled it, Brockman didn’t understand how this would help the company beat its competitors. “No matter what I said,” Hedley told us, “Greg kept going back to ‘So how do we raise more money? How do we win?’ ” According to several interviews and contemporaneous records, Brockman offered a counterproposal: OpenAI could enrich itself by playing world powers—including China and Russia—against one another, perhaps by starting a bidding war among them. According to Hedley, the thinking seemed to be, It worked for nuclear weapons, why not for A.I.?
The fact that the chatbots are now used for plausible deniability and blame shifting in Gaza or Iran wasn't known then.
It's...weirdly a valid question. If Sam fibs as much as the next guy, we don't have a Sam problem. Focussing on him alon is, best case, a waste of resources. Worst case, it's distracting from real evil. If, on the other hand, as this reporting suggests, Sam is an outlier, then focussing on him does make sense.
And when you're dealing with a potential existential threat, this is an existential problem.
https://www.reddit.com/r/AskReddit/s/VWJVBNzc2u
I am fairly confident when I say this -- sama is a sociopath. I don't know how anyone with solid intuition could even come to any other conclusion than the guy is deeply weird and off-putting.
Some concepts from the book:
> Core trait: The defining characteristic is the absence of conscience, meaning they feel no guilt, shame, or remorse.
> Identification: Sociopaths can be charming and appear normal, but they often lie, cheat, and manipulate to get what they want.
> The Rule of Threes: One lie is a mistake, two is a concern, but three lies or broken promises is a pattern of a liar.
> Trust your instincts over a person's social role (e.g., doctor, leader, parent)
Check and check.
OpenAI is too important to trust sama with. He needs to go. In fact, AI should be considered a public good, not a commodity pay-as-you-go intelligence service.
> OpenAI is too important to trust sama with.
...wat? They made a chat bot. How can that possibly be so existentially important? The concept of "importance" (and its cousin "danger") has no place in the realistic assessment of what OpenAI has accomplished. They haven't built anything dangerous, there is no "AI safety" problem, and nothing they've done so far is truly "important". They have built a chat bot which can do some neat tricks. Remains to be seen whether they'll improve it enough to stay solvent.
The whole "super serious what-ifs" game is just marketing.
We only say a lot of CEOs are sociopaths because they're in that third category we haven't named, where they're very good at manipulating people, but also can feel conscience, guilt, remorse, etc, perhaps just muted or easier to justify against.
E.g. if you think you're doing something for the betterment of mankind, it doesn't really matter if you lie to some board members some year during the multi-decade pursuit.
There is -- I call it "corpo sociopath." The corpo sociopath really comes out in the workplace, less so in personal life.
Whereas the people in the category I’m describing might feel those things, but prioritize those feelings far below the benefits of achieving what they set out to achieve.
Yes that is the core trait I highlighted in the 1st bullet.
I only saw this thread by chance and almost didn't look, because the title made the piece sound like a flamebait blog post. Fortunately I saw newyorker.com beside the title and looked more closely.
Thank you for looking. Please do spread this kind of reporting in your communities, and subscribe to investigative outlets when you can.
A paper with "ideas to keep people first" was (coincidentally?) published today:
https://openai.com/index/industrial-policy-for-the-intellige...But the discussion is generally pretty low quality with these sort of posts. People react without having read the story, or with whatever was on their mind already, or are insubstantive, or simply low effort. I don't think you'll lose k-factor not having a bigger post here.
Sometimes if you talk to the mods, they'll let you know their perspective. I generally find they're correct that people are much better at contributing/disseminating new knowledge to the world on more technical topics here.
[1]: https://news.social-protocols.org/stats?id=47659135
We do tend to devalue titles like this, or more likely change them to something more substantive (preferably using a representative phrase from the article body), but I'm worried that if I did that here we would get howls of protest, since YC is part of the story.
It's an interesting dilemma. Many very respected publications use provocative titles because of the attention economy. And I'm sure you have good data that provocative titles lead to drive-by comments and flame wars.
But I don't think big_toast was entirely wrong that there is a side effect of sometimes burying articles that are by their nature provocative. And how do you distinguish a flame war over a title from a flame war over content? That's not a leading question. I don't know.
I just don’t feel like OpenAI has a legitimate shot at winning any of the AI battles.
Therefore, I feel like “Sam Altman may control our future” is a far stretch.
You might be. Or at least I feel like Gemini is actually dumber than a house of bricks - I have multiple examples, just from last week, where following its advice would have lead to damage to equipment and could have hurt someone. That's just trying to work on an electronics project and askin Gemini for advice based on pictures and schematics - it just confidently states stuff that is 100000% bullshit, and I'm so glad that I have at least a basic understanding of how this stuff works or I would have easily hurt myself.
It's somewhat decent at putting together meal plans for me every week, but it just doesn't follow instructions and keeps repeating itself. It hardly feels worth any money right now, like it's some kind of giant joke that all these companies are playing on us, spending billions of these talking boxes that don't seem that intelligent.
I also use claude at work, and for C++ programming it behaves like someone who read a C++ book once and knows all the keywords, but has never actually written anything in C++ - the code it produces is barely usable, and only in very very small portions.
Edit: I just remembered another one that made me incredibly angry. I've been reading the Neuromancer on and off, and I got back into it, but to remind myself of the plot I asked Gemini to summarise the plot only up to chapter 14, and I specifically included the instruction that it should double check it's not spoiling anything from the rest of the book. Lo and behold, it just printed out the summary of the ending and how the characters actions up to chapter 14 relate to it. And that was in the "Pro" setting too. Absolute travesty. If a real life person did that I'd stop being friends with them, but somehow I'm paying money for this. Maybe I'm the clown here.
I’ve always been a huge fan of Ronan Farrow’s journalism and willingness to speak truth to power. I think he’s pulling at exactly the right thread here, and it’s very important to counteract Altman’s reputation laundering given that we run a very real risk of him weaseling his way into the taxpayer’s wallet under the current administration.
https://en.wikipedia.org/wiki/Betteridge%27s_law_of_headline...
If for no other reason, given what happened when the board fired him... no. I'd say not.
The overall response and particularly the body language speaks a lot.
https://www.reuters.com/technology/openai-signs-deal-with-co...
This is a damage control piece, and you see that the most stinging comments here get downvoted.
The real question is: can anyone be trusted if the fever dreams of super-intelligence come true? Go ahead and replace Sam Altman with someone else - will it make a difference? Any other CEO is going to be under the same overwhelming pressure to make a profit somehow. I think the OpenAI story is messier because it was founded for supposedly altruistic reasons, and then changed.
Methinks many of Altman's detractors protesteth too much. He's doing his job as it is defined (make OpenAI profitable.) Nothing of substance in this article seemed to make him exceptionally "sociopathic" compared to any other tech CEO. It goes with the territory.
What depressed me most is that trillions of dollars are being raised for building what will undoubtedly be used as a weapon. My guess is the ROI on that money is going to be extremely bad for the most part (AI will make some people insanely rich, but it is hard to see how the big investors will get a return.) Could you imagine if the world shared the same vision for energy infrastructure (so we could also stop fighting wars over control of fossil fuels and spewing CO2?) A man can dream...
lol do you think these guys have ever been hit? Let alone in the face. They’d probably be less eager to mouth off as much as they do if so.
The last person that this happened to was Sam Bankman Fried as investors and regular folk finally realized he was full of complete shit and could only talk the game for so long until the truth emerged.
These sociopaths are so good at giving away nothing. He managed to engender sympathy instead of saying "I'm not gonna talk about anything that happened then".
Also very weird how many of these people are so deeply-linked that they'll drop everything they're doing just to get this guy back in power? Terrifying cabal.
No, he cannot.
Not all billionaires are villians. But it is long-known in organizational psychology that dark triad [1] traits are very "helpful" if one wants to climb career ladders fast.
[1] https://en.wikipedia.org/wiki/Dark_triad
Also, there's the psychopathic and narcissistic tendencies of greedier people and the false "virtue" "greed is good" that is contrary to the values espoused by Adam Smith.
We need standard income tax brackets of 90% after $20M/y and 99% after $100M/y.
TLDR but just the heading is already ugly. No single person no matter how nice they're should be able to control our future. Power corrupts, what fucking trust. We are supposed to be democratic society (well looking at what is going on around this is becoming laughable)
2. You cannot "control" superintelligent AI.
FYI: I am by far not the only one to have experienced this and it 100% impacts hiring and other decisions at OpenAI.
It wouldn't particularly surprise me if Sam Altman were racist, but I'm curious what the specific incident you observed was.
1. I could have hidden my identify behind a throwaway. I did not feel that would be appropriate when making this calim.
2. I am not looking for anything, literally at all. Any follow ups for blogs; anything that would benefit I will not answer.
3. This is NOT a new account, I am very easy to find; I am 6'1 140lbs
I was working for a company called NationBuilder and I had the opportunity to go on a work trip. Outside of a talk he had just given I was waiting for my ride and I looked over like...damn thats the speaker. I wanted to say Hi; he damn near flagged down the police. I apologized and just decided to move on.
Note: It was in Reno, and no I don't want to go into details; the others are not hard to find because I happened upon them via blog posts so i'm sure if someone with the accumen of RF wants to know, he will find.
I have heard similar stores from several people in the years since. I AM NOT CALLING THIS PERSON RACIST. I am saying; he is observably scared of black people and that is not someone I want making descions about how the world moves foward.
If you don't believe what I shared is true, address that directly. But seeing my post sitting at 1 point and [flagged] after 2 hours is not OK. Just as DJT can't flag away his issues, you shouldn't be able to do so on HN.
One of the things I've loved most about HN is that it was real — grounded in observability, empirical evidence, not bias or feelings. I really hope that what happened to my post is not the beginning or a continuance of the end for that ethos.
The biggest flaw is that it spends way too much time on high-school level drama and "he-said-she-said" gossip about Sam Altman’s personal life instead of focusing on the actual technical and corporate capture of OpenAI.
The author treats the "nonprofit mission" like some holy quest that was "betrayed," when anyone with a brain in tech saw the Microsoft deal as the moment the original vision died. Instead of a hard-hitting look at how compute-monopolies are actually forming (MSFT AMZN NVDA and circular debt dealing inflating the AI bubble that could crash the economy), we get 5,000 words of hand-wringing over whether Sam is a "nice guy" or a "liar."
Who cares???????
The board failed because they had no real leverage against billions of dollars, not because they didn't write enough Slack messages. It's a long-winded way of saying "Silicon Valley has internal politics," which isn't news to anyone here.
I’d be more concerned about Anthropic both being in the good graces of the public and having access to all of our computers indirectly with Claude Code.
"the local drug-dealing pimp is so passe, we need to investigate the most upstanding members of the community just to be sure" is a frankly insane strategy