Hacker Newsnew | past | comments | ask | show | jobs | submit | LZ_Khan's commentslogin

OpenAI's just trading equity for GPU credits at this point?

But also a billion users is ChatGPT's biggest weakness. So many free users burning compute up. So many incentives to nerf the intelligence to affordable levels. Sounds like a nightmare.

A very obvious and likely to happen strategy is to turn the emotional manipulation dial up. Make users more dependent on validation and attention the model provides.

They're already doing it, but wonder how far they'll take it.


At least they're paying. OpenAI should have the largest IP settlement, they just would rather contest it and not pay for eternity.

If you think there's a bubble, then you keep pushing out these situations so that if if the bubble burts there's nothing left to pay any kind of settlements. The only time companies pay a settlement is if they think they are going to get hit with a much larger payout from a court case going against them. Even then, there's chances to appeal the amounts in the ruling. Dear Leader did this very thing.

The crazy thing is, the blatant AI generated replies with unnecessary gravitas and cliche writing are just the obvious ones.

Those are probably replies crafted by non-English speaking scammers from India / Russia / China.

There's probably a whole sea of undetectable replies from people who know how to prompt the models properly.


If you care about improvement of models, you would support the US labs here.

It costs hundreds of millions of dollars to train a frontier model. It's not just "scraping the web."

Distillation allows labs to replicate these results at 1/100th of the cost. This creates a prisoner's dilenmma which incentivizes labs to withhold their models from the public.


How much did it cost to produce all the data on the internet and every book ever published? Surely even the most conservative calculations put it at multiple years of planetary GDP. The same argument can be made to say that letting the big labs get away with pirating it will disincentivize people to publish anything.

I personally have stopped publishing publicly, since my research is still on the fuzzy boundary of AI's current knowledge, my website gets scraped daily, and I don't want to contribute to paid models for zero acknowledgement or compensation.

> I personally have stopped publishing publicly, since my research is still on the fuzzy boundary of AI's current knowledge, my website gets scraped daily, and I don't want to contribute to paid models for zero acknowledgement or compensation.

I don't know about your works so pardon me but thinking about it, would a better solution be for gated communities at the very least, say matrix or xmpp or irc be better?

I suppose that scraping bots of matrix would be quite hard for AI companies to setup? but anyone interested in reading your contents can still find the data if they are interested plus you get the additional benefit of a community/like-minded people.


Not only publishing, it has already disincentivised a huge part of what made Web 2.0: public APIs for data access to platforms.

It was amazing to be able to create some toy projects using data from big platforms, now they're all afraid LLM trainers will scrape their contents and create a competitor to their moat, the data.

It just sucks at many different levels.


This reads a bit like over-moralizing to me. US labs will continue improving their models because they have to make money in a competitive market. Chinese distillations have arguably improved the status-quo, with Qwen and R1 forcing GPT-OSS to be released to the public. American businesses are competing, and American customers are getting better products because of the competitive pressure on them.

Your purported "prisoner's dilemma" hasn't happened yet to my knowledge, instead we seem to see the opposite. The high-speed development velocity has forced US labs to release more often with less nebulous results. Supporting either side will contribute to healthier competition in the long run.


If 'we' really cared about the improvement of models all of them would be public.

Anything else just proves someone prefers making money to improving the models.


> incentivizes labs to withhold their models from the public.

Does it really? How would they get revenue if they withhold their models? And doesn't economics generally say that if it's easier for your competitor to catch up, you have a higher incentive to maintain your lead?


I think that the bigger conversation to be had here is about the environmental damage - if by using distillation we can really train new models at 1% of the cost in energy, it is ethically imperative that we do this.

> If you care about improvement of models, you would support the US labs here.

I guess I don't care then.


that's fair!

Tell me how they obtained that data?

Nobody feels sorry for big Multinationals trying to skirt Copyright for their own good, but then cry about it when their competition ignores it too.

You can't have your cake and eat it


> incentivizes labs to withhold their models from the public.

This is the only way they make money.


biggest problem is that it's slow. also safety seems overtuned at the moment. getting some really silly refusals. everything else is pretty good.

Improving the UI would only take you a few prompts

lmao

No, TSLA purposely does not list the details of the incident.

I do this but instead in a google doc. Even better because I can use LLM's to query it aftewards.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: