To answer a few people at once: I did mention compensation as a factor in the post, but I didn't elaborate details, so easy to miss. Comp is important of course, but so are the other factors. It feels like I can't go for a day without reading about the cost of AI datacenters in the news, and I can do something about it.
Again, many comments here saying I only care about the money, and while comp is an important factor I think it characterizes me as someone I'm not, and forgets what I've been doing for the past two decades. I've spent thousands of hours of my life writing textbooks for roughly minimum wage, as I want to help others like me (I came from nothing, with no access tech meetups or conferences, and books were the gateway to a better job). I've published technologies as open source that have allowed others to make millions and are the basis for many startups. I'm also helping pioneer remote work and hoping to set a good example for others to follow (as I've published about before). So I think I'm well known for caring about a lot of things during the past couple of decades.
It's okay to want to make money. You don't really have to justify it this hard unless you want people to really think you don't think comp is important, which is a bit sus to be fair.
Even if people don't want it to be about the money, it's still about the money because of the world we live in. Good vibes don't pay the mortgage or put food on the table. More money equates to better health and future outcomes for a person and their family, so how couldn't it always be about the money?
Of course once someone has money they can say it's not about the money, but that privilege is literally bought with...money.
Reducing runtime energy use over years won't really make up for the resource use that goes into building the data center. It's just moved around, similar to how Elon moves around money as needed to bolster the financials of a particular project.
Like with the airline industry it's not just the smog they blow on our food. Drink carts, seat belts, barf bags all have a resource intensive energy and materials pipeline.
From my reading of what you said, you think comp is important and so are other things. You outlined those a bit too, but I already forgot them.
Quite frankly, I think some people here are too quickly spooked and think what you say is sus. I simply see that as a sign that they aren't fully having a good faith discussion. Or they simply read things way differently than I do.
I'm simply writing this because I think there are enough people that have a similar reading to what you wrote. They simply don't mention it as people who feel "outraged" (a bit too dramatic of a term but English is my 2nd language). "Outraged" people seem simply more vocal to me.
For clarity: I feel neutral about this whole thing. I do appreciate the work you've done in the past.
The issue is that you're doing lot, but not saving the planet.
What do you think is happening with the efficiency gains? You're making rich people richer and helping AI to become an integral (i.e. positive ROI from business perspective) part of our lives. And that's perfectly fine if it aligns with your philosophy. It's not for quite a few others, and you not owning up to it leads to all kinds of negativity in the comments.
>What do you think is happening with the efficiency gains?
may it happen that the efficiency gains decrease demand and thus postpone investment into and development of new and better energy sources? If one couldn't get by just by bringing 20 trucks with gas turbines, may be he would have invested in fusion development :)
> may it happen that the efficiency gains decrease demand
What mechanism would make this happen?
Demand could decrease if AI became worse, but efficiency doesn't make AI worse - it actually makes possible at all to run bigger, better models (see the other comment with a link to Jevon's paradox), which increase, not decrease demand (more powerful models may have new capabilities that people want to use)
Alternatively, AI demand could decrease through political pressure (either anti-AI sentiment takes a foothold on the public, and/or government regulation strangle demand on the sector like it did for eg. on tobacco industry). But another way to reap the benefits of more efficient AI datacenters is to make it a talking point on how AI environmental impacts can be mitigated, which could curb anti-AI sentiment.
Either way, those possibilities don't decrease demand for AI - they are either neutral, or increase demand instead.
Great to see you're in Sydney Brendan, and let the haters hate.
You have done a brilliant job elevating your chosen specialty to the world, and encouraging and inspiring others in the industry for a long time - so you should be fairly compensated for that lofty position. I don't envy the late nights or very early mornings you have ahead of you on conference calls with SF, but good luck at OpenAI mate !
I think it’s not about compensation or passion, but something a bit more abstract.
I’ll give it a shot; I think you’re successful in what you do and very altruistic and open, not only in your discoveries but also your opinions. You also have a higher sense of duty. as oscar wilde once said, we’re all in the gutter, but some of us are looking at the stars. compensation is boring gutter talk. It’s hard for people to reconcile your benevolence with your success, and just as the trope of joining a company to change the world is a veil for making money, so is the trope of criticizing the agent of change entering an industry because the industry is bad.
Personally I can’t wait to read about the inefficiencies you find and have a little glimpse into openai tech from your opinionated point of view.
I am a long time fan, I have the physical copy of each and every book that you have authored, I have watched each and every video that you are in, and I walk team members and clients through your USE method at every engagement I am on.
I would say to you that the "make the world a better place" has been excessively misquoted. Even the Silicon Valley episode on Tech Crunch parodies show how anything and everything is intended to "make the world a better place".
Please reconsider your use of the phrase given the well-earned negativity around it.
I apologise for the incorrect "make the world a better place" attribution to you.
It does appear that for many of us "saving the planet" has become akin to "making the world a better place".
I was excited last night thinking about what you might uncover at OpenAPI. While we can all speculate (device drivers? File systems? Python itself? DB queries? Memory architecture? Inference algos? ) I recall “I have no data yet. It is a capital mistake to theorize before one has data” by Sherlock Holmes in "A Scandal in Bohemia".
All the best with your work at OpenAPI. We will all hopefully learn about writing more energy efficient code and about more efficient LLMs and more thanks to your work!
I mean, I don't know you well, but, I see your posts on here from time to time and from what I gather you are very, very, exceptional at what you do.
Reality is, these AI giants are here and they are using a massive amount of resources. Love them or hate them, that is where we are. Whether or not you accept the job with them, OpenAI is gonna OpenAI.
Given how much the detractors scream about resource uses, you'd think they'd welcome the fact that someone of your calibre is going in and attempting to make a difference.
Which, leads me to believe you're encountering a lot of projecting from people who perhaps can't land the highest of comp roles, and shield their ego by ascribing to the concept of it being selling out, which they would of course never do.
It's probably impossible to prove I'm not projecting..
However. I am putting my curious foot forward here:
What were the toughest ethical quandaries you faced when deciding to join OpenAI?
To give a purely hypothetical example which is probably not relevant to your case: if I had to choose DeepSeek or OpenAI, I think I would struggle with openness of the weights..
Ignore the haters (who sadly have become extremely common on HN now).
I loved your work back when I was an IC, and I'm sure this is a common sentiment across the industry amongst those of us who started systems adjacent! I still refer to your BPF tools and Systems Performance books despite having not written professional code for years now.
Can't wait to read content similar to what you wrote about when at Netflix and Intel albeit about the newer generation of GPUs and ASICs and the newer generation of performance problems!
It’s not a crime if you do something for money. Those who comment are likely doing the same and they couldn’t get into a company like OpenAI and hence the hatred! Keep doing the great work you always did! Excited to see what you ll do with all the resources in the world.
Hey Brendan - first time listener, first time caller.
Inferring the overall tone from the comments, I think the folks here are struggling with what sounds like a logical fallacy from someone who is certainly a logical thinker.
> how I could lead performance efforts and help save the planet.
The problem on the face of it being: Performance gains will not translate to less energy usage (and by extension less heat released into the atmosphere). Rather, performance gains will mean that more effective compute can be squeezed from the existing hardware.
If performance gains translate to better utilization of the hardware, it also follows that it will translate to more money for the company, allowing for the purchase of more GPUs. Ad infinitum.
My stance is that this is just businesses doing what they do. It's always required regulation to slow down the direct/indirect negative byproducts (petro companies being the most obvious example). I don't see how AI would inherently be different.
Is there another angle that I (we) am (are) missing where the performance efficiencies translate to net benefits for the planet?
Given the widely reported issue of AI environmental cost, and being among the top people in the world who can fix it, I don't feel comfortable with _inaction_. I can reduce datacenter sizes, so I'll do that.
As a performance engineer I'm familiar with Jevons paradox, but it does not discourage improving efficiency.
Read this Gregg.
I'm the first one to always put your work and books in any comments related to you or your work here. But...
This is a company which at the first opportunity seized and stopped doing open research, cut open source contributions, converted itself to for profit after years of fiscal benefits, that scrapped its ethics committee and removed all engineers who opposed any of this.
Don't come with the excuse there is any work being done for the better of something.
One should never input one's own expectations into another, but, I feel disappointed. It is having the guy I saw growing from the first posts working for an evil machine on his own volition.
Do what you want. But that's what I feel about this disheartening news
It would be good if the performance improvements made can be applicable across the industry so everyone benefits. But it doesn't sound unbelievable that OpenAi may want to keep some of it secret to keep an advantage over others?
Unbelievable? It’s unfathomable - out of all existing AI companies, OpenAI is the least open of them all. They have stopped contributing any useful research into the public domain. Even infamous “villains” like Meta and China are doing leaps and bounds more compared to /Open/AI and the like.
> I stood on the street after my haircut and let sink in how big this was, how this technology has become an essential aide for so many, how I could lead performance efforts and help save the planet.
Brendan.
First of all congratulations on your new job. However,
It is easier to just say to everyone it is about the money, compensation and the stock options.
You're not joining a charity, or to save the planet, this company is about to unload on the public markets at an unfathomable $1TN valuation.
Thanks for taking the risk in this environment and posting about your experience from a personal standpoint. [environment: people will come at you from all angles with very passionate opinions]
Interesting. Out of curiosity, how long do you think OpenAI can survive as a company? Put another way, what would be your guesses for probability of failure on 1yr, 3yr, and 5yr horizons?
EDIT: possibly a corollary--does Mia pay money for chatgpt or use a free plan?
'Before you joined' seems to that she doesn't anymore.
I find it a little amusing that AI companies provide free AI subscriptions to their employees and their families. Perhaps because I'd never thought of it that way.
That's kinda nice actually; I assume employees get early access to models to test (5.3 codex, for example). Do families get it too?
I use LLMs daily and they've become an essential part of my life. Really hoping your work can help make LLMs and AI more broadly cheaper and more abundant.
Would be fantastic if you can find a way to make optimizations you find available more openly. The whole ecosystem benefits when efficiency improvements are shared. Looking forward to seeing where this goes and don't let the negativity from some get to you
Humans are complex and have multiple sources of motivation. You don't know whether he took the offer with the highest pay. He's likely wealthy enough that he can pay less attention to his income and focus on his other sources of motivation if he wants to. That's not to say pay is not a factor in his choice, but it need not be the only or primary one. This is a luxury of the privileged for sure, which can make it difficult to relate to.