Hacker Newsnew | past | comments | ask | show | jobs | submit | ebonnafoux's commentslogin

They repeat only six sentences during 100+ comments:

Worked like a charm, much appreciated.

This was the answer I was looking for.

Thanks, that helped!

Thanks for the tip!

Great explanation, thanks for sharing.

This was the answer I was looking for.


Over the last ~15 years I have been shocked by the amount of spam on social networks that could have been caught with a Bayesian filter. Or in this case, a fairly simple regex.

It's the bear trash lock problem all over again.

It could be solved by the filter but filter would also have a bunch of false positives


It seems like if the content is this hollow and useless, it shouldn't matter if it was a human or spambot posting it.

Well, large companies/corporations don't care about Spam because they actually benefit from spam in a way as it boosts their engagement ratio

It just doesn't have to be spammed enough that advertisers leave the platform and I think that they sort of succeed in doing so.

Think about it, if Facebook shows you AI slop ragebait or any rage-inducing comment from multiple bots designed to farm attention/for malicious purposes in general, and you fall for it and show engagement to it on which it can show you ads, do you think it has incentive to take a stance against such form of spam


> Well, large companies/corporations don't care about Spam because they actually benefit from spam in a way as it boosts their engagement ratio

I'm not sure that's actually true. It's just that at scale this is still a hard problem that you don't "just" fix by running a simple filter as there will be real people / paying customers getting caught up in the filter and then complain.

Having "high engagement" doesn't really help you if you are optimizing for advertising revenue, bots don't buy things so if your system is clogged up by fake traffic and engagement and ads don't reach the right target group that's just a waste.


Yeah, I almost included that part in my comment, but it still sucks.

The question is can you be trained? Beside the obvious case, some IA generated photo could not be distinged from real one.


You can easily make a RLAIF loop.

- Take a list of n animals * m vehicule

- Ask a LLM to generate SVG for this n*m options

- Generate png from the svg

- Ask a Model with vision to grade the result

- Change your weight accordingly

No need to human to draw the dataset, no need of human to evaluate.


There can be a lot of definition but I propose " a wage where you do not need to annoy your customer for tip".


China is a 1.4 billion people country, more than US + Europe, so it is expected that there is a wild gap between high end product and low cost there.


It's wild that people check the box

> I have searched existing issues and this hasn't been reported yet

when the first 50 issues are about 500 error.


https://ebonnafoux.bearblog.dev/ Started two weeks ago, but my resolution is to publish at least once a month


Yes but in practice, if you compute K=X.wk, Q=X.wq and then K.tQ you make three matrice multiplication. Wouldn't be faster to compute W=wk.twq beforhand and then just X.W.tX which will be just two matrices multiplication ? Is there something I am missing ?


Most models have a per-head dimension much smaller than the input dimension, so it's faster to multiply by the small wk and wk individually than to multiply by the large matrix W. Also, if you use rotary positional embeddings, the RoPE matrices need to be sandwiched in the middle and they're different for every token, so you could no longer premultiply just once.


Now divide "coal consumption", "executions" and " CO2 emission" by capita, it is also interesting.


I have seen some codebase doubling the number of LoC after "refactoring" made by humans, so I would say no.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: