Hacker Newsnew | past | comments | ask | show | jobs | submit | derrida's commentslogin

Not a haiku, more a koan.

Has ChatGPT gotten worse over past few months or is it I just have seen other things higher quality, or they stopped caring about user or something?

All of a sudden feels like it gives me boilerplate and boiler plate of PR and cheesy reasoning, and like no actual answers - worse even - highly confident wrong answers that it then seeks to justify or explain (like it doesn't seem humble enough to be like "Actually, got that wrong" or if challenged it just caves over, accepts too readilythe assumptions in what the user is asking, or just blindly accepts a premise of the question) it's almost useless, like before it used to seem like could get it to emulate the way a certain writer or discourse speaks, now it seems like this derpy highschool just wants to be in kid that went into public relations and the language no matter what the topic seems always the same, it's really spammy feeling,

I could be asking it questions about like how medieval monks talked about light and the breath in latin and it will be replying like I'm interested in monetising or improving my lifestyle or some b.s. I don't think it used to be this way?

reminds of a circa 2003-6 wordpress sites - blackhat seo - feeling to generate back links to push affiliate links or something, with markov generated content designed to push back links for the actual human written landing page

It's not like this on the other llms, something's up.

Or maybe they have just found the niche and it is a bunch of people who do think like that - like I dunno - middle management the world over

that is scary ... bonus ghastly incantations of the epistemology of middle management


That's how ChatGPT always seemed to me. One of the reasons I exclusively use other models is it would rather make something up than say "I can't find anything about that."

But I'm starting to wonder about something.

I've noticed a lot of people claiming the models—all the models from all the big providers—are deteriorating, and then go on to describe the problems that skeptics picked up on during their first few days of usage.

The models really could be getting worse. I haven't noticed anything but I don't know.

But do you think its possible that this is more akin to a honeymoon period? Depending on how you use the system and a fair bit of luck, the problems may show up for you pretty early, or may take a while to become obvious.


It's gotten bad enough that I finally cancelled this month in protest. It's not just you.

people have been talking about "models of models" for arbitration opportunity in inference for about 1.5 yrs.

Arbitration idea: if a user doesn't need high QOS of newest LLM, slip them a cheaper LLM, run their query at reduced quality. measure if they cost you fewer $s in the lower QOS. => profit.

For chatgpt the arbitration opportunity looks more like "we could allocate this amount of gpu to training or inference, we are losing money if we offer the highest quality infra"

In addition there's other interesting economics scaling that can be done outside of "models of models" that are far more profitable. I won't go over all of them (and some of them I feel are quite powerful) but the laziest one is that subscription models count on some zombie users as a counterweight to highly expensive single users, and as a source of stable cashflow.

Zombie users are ones that are paying for sub but not actively or barely using the service


They made a big point of explicitly advertising this as a feature with the GPT-5 rollout, no? Routing to cheaper models/less reasoning depending on the input prompt.

There's potentially some discussion of this publicly to investors. I feel there's more going on there and is re: quality issues described above.

I think you have it set to the wrong mode. If you set it to Thinking with “Extended” thinking effort, is it slower but almost never wrong (because it searches the web to get verify all its assumptions and answers).

Yeah I set thinking as my default and never looked back. It’s my daily driver and extended thinking is usually not too slow. The way that the “instant” model trades quality for speed is not worth it and I don’t need the instant gratification. (But I also don’t do entertainment chatting so ymmv.)

All models are variable in quality simply because they need to do some financial engineering to make the financial standing somewhat healthier. There’s a lot of fear in the market and need for signalling around the viability of the business of llm models and generating returns on invested capital.

If I recall correctly, in their pivot to Codex they took a sizable amount of compute away from ChatGPT

Set your default mode to thinking and set some custom instructions. Night and day difference in UX.

Are you using the free or the paid version? Did you try personalization settings other than the default?

Are you using the free service or paid? Because when the free service drops back to older or smaller models, there are noticeable quality differences.

> ghastly incantations of the epistemology of middle management

I mean, LLM writing has been like that from the early on. Its most perfect niche for writing is the LinkedIn blog post.


Is it perhaps because this is for claude code but there's other tools that use anthropics api like custom agents? (some i prefer to use than claude code - e.g sketch.dev what is now called shelley at exe.dev) perhaps?


No, because this doesn’t actually “fix” any existing code. It’s only useful for helping an LLM to modify your code to adjust the caching parameters in the right place, but it doesn’t have the correct API for that.


I have no idea about any of that but like I wasn't thinking of github until you mentioned it and this comment I upvoted because was informative and relevant to the discussion and I don't know about R.E but curious to try and this kind of activity just seems like the sort of things people who are interested in software, learning and aware of security do... like to find bugs or malware or something... FOSS or not - actually "especially if not FOSS" you'd kinda like people to scan their binaries at <big tech corp> and have that knowledge indigenous wouldn't you? while thinking of code security etc, anyway

Is this a bad look for Derrida.org?

Anyway, "not my business"


The other LLMs probably don't have the training data in the first place.


Er...

"Study uncovers presence of CSAM in popular AI training dataset"

https://www.theregister.com/2023/12/20/csam_laion_dataset/.


I wouldn't equate the two.

There's someone who was being held responsible for what was in encrypted chats.

Then there's someone who published depictions of sexual abuse and minors.

Worlds apart.


Telegram isn't encrypted. For all the marketing about security, it has none, apart from TLS, and an optional "secret chat" feature that you have to explicitly select, only works with 2 participants and doesn't work very well.

They can read all messages, so they don't have an excuse for not helping in a criminal case. Their platform had a reputation of being safe for crime, which is because they just... ignored the police. Until they got arrested for that. They still turn a blind eye but not to the police.


ok thank you! I did not know that, I'm ashamed to admit! sort of like studying physics at university a decade later forgetting V=IR when I actually needed it for some solar install. I took "technical hiatus" about 5 years and recently coming back.

Anyway cut to the chase, I just checked out Mathew Greens post on the subject, he is on my list of default "trust what he says about cryptography" along with some others like djb, nadia henninger etc

Embarrased to say I did not realise, I should of known! 10+ years ago I used to lurk the IRC dev chans of every relevant cypherpunk project, including of text secure and otr-chat when I saw signal being made and before that was witnessing chats with devs and ian goldberg and stuff, I just assumed Telegram was multiparty OTR,

OOPS!

Long winded post because that is embarrassing (as someone who studied cryptography undergrad in 2009 mathematics, 2010 did postgrad wargames and computer security course and worse - whose word once about 2012-2013 was taken on these matters by activists, journalists, researchers with pretty knarly threat model - like for instance - some guardian stories and former researcher into torture - i'm also the person that wrote the bits of 'how to hold a crypto party' that made it a protocol without an organisation and made clear the threat model was anyone could be there, oops oops oops

Yes thanks for letting me know I hang my head in shame for missing that one or some how believing that one without much investigation, thankfully it was just my own personal use to contact like friend in the states where they aren't already on signal etc.

EVERYONE: DON'T TRUST TELEGRAM AS END TO END ENCRYPTED CHAT https://blog.cryptographyengineering.com/2024/08/25/telegram...

Anyway as they say "use it or lose it" yeah my assumptions here no longer valid or considered to have educated opinion if I got something that basic wrong.


[flagged]


In November 2012, Epstein sent Musk an email asking “how many people will you be for the heli to island”.

“Probably just Talulah and me. What day/night will be the wildest party on your island?” Musk replied, in an apparent reference to his former wife Talulah Riley.

https://www.theguardian.com/technology/2026/jan/30/elon-musk...

I think there's just as much evidence Clinton did as Musk. Gates on the other hand.


To my knowledge Musk asked to go but never actually went. Clinton, however, went a dozen or so times with Epstein on his private jet?

Has the latest release changed that narrative?


Yes. He went at least once in 2012, then asked to go again in 2013 and Epstein refused.


Oof.


Musk did ask to go after Epstein was sentenced.


I hate to be the “source” guy but can I get one?


The Epstein files


Additionally Clinton is listed several times on the Lolita express flight logs, Elon never

Elon didn't ask to go, he was invited multiple times


If Elon never asked to go, why do the Epstein files have an email from Elon to Jeff where Elon asks to go? Was it fabricated?


... Eh? This isn't about Musk's association with Epstein, it's about his CSAM generating magic robot (and also some other alleged dodgy practices around the GDPR etc).


Business as usual.


They could use sex to sell.

"imagine everything is a file"


That's the Newcastle connection and "..." supplementing "." And ".."


OS is designed by Bell Labs, of Unix and C fame, but ok.

EDIT: I think some of the in-jokey, arty farty culture about it is because the rest of us didn't run with it, it's basically for hobby and personal use - so its culture now. But about any one I know who has used it enough to "grok" it kinda is like "wow if this won" because its kinda such a paradigmatically good thing whole kind of cottage industries of software may not even exist because the user would be able to write a pipe from a file that is a device to write to another one type thing.


Plan 9 is supposed to be absurd. Pull the cork out of your ass lmao


Ah, the hell site, no click.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: