Hacker Newsnew | past | comments | ask | show | jobs | submit | Moosdijk's commentslogin

This is nice website.

I wonder why glm is viewed so positively.

Every time I try to build something with it, the output is worse than other models I use (Gemini, Claude), it takes longer to reach an answer and plenty of times it gets stuck in a loop.


I've been running Opus and GLM side-by side for a couple weeks now, and I've been impressed with GLM. I will absolutely agree that it's slow, but if you let it cook, it can be really impressive and absolutely on the level of Opus. Keep in mind, I don't really use AI to build entire services, I'm mostly using it to make small changes or help me find bugs, so the slowness doesn't bother me. Maybe if I set it to make a whole web app and it took 2 days, that would be different.

The big kicker for GLM for me is I can use it in Pi, or whatever harness I like. Even if it was _slightly_ below Opus, and even though it's slower, I prefer it. Maybe Mythos will change everything, but who knows.


> The big kicker for GLM for me is I can use it in Pi, or whatever harness I like.

Yes, but... isn't the same true for Opus and all the other models too?


Opus is about 7 times more expensive than GLM with API pricing. And since you can only use the Opus subscription plan in CC, you're essentially locked into API pricing for Pi and any other harness.

So you're either paying $1000's for Opus in Pi, or $30/month for GLM in Pi. If the results are mostly equivalent that's an easy choice for most of us.


Perhaps I'm being extremely daft: If the API is 7 times more expensive, then why is it $1000 vs $30? Or is there a GLM subscription one can use with Pi? Certainly not available in my (arguably outdated) Pi.

I'm not the OP, but it's the latter. I'm currently using the "Lite" GLM subscription with OpenCode, for example. I'm not using it very heavily, but I haven't come close to hitting the limits, whereas I burned through my weekly limits with Claude very regularly.

I am using GLM-5.1 in pi.dev through Ollama Cloud. I am able to get by on the $20 plan. I use it a lot and the reset is hourly for sessions and weekly overall. This is the first week I got close to the limit before reset at about 85% used. I am probably using it about 4 hours a day on average 6 or 7 days per week.

You can use GLM’s coding plan in Pi, just use the anthropic API instead of the OpenAI compatible one they give.

Or tell pi to add support for the coding plan directly. That gave me GLM-5.1 support in no time along with support for showing the remaining quota, etc, too.

It also compresses the context at around 100k tokens.

In case anyone is interested: https://github.com/sebastian/pi-extensions/tree/main/.pi/ext...


I have used GLM 4.7, 5 and 5.1 now for about 3 month via OpenCode harness and I don't remember it every being stuck in a loop.

You have to keep it below ~100 000 token, else it gets funny in the head.

I only use it for hobby projects though. Paid 3 EUR per month, that is not longer available though :( Not sure what I will choose end of month. Maybe OpenCode Go.


EDIT: Ok, now I tried GLM for the first time in the morning CET, and it was .. bad. The reasoning took 5 mintues for a very very small .html file going around in circles.

Evening CET experience for me is super smooth.


That's unfortunate. 70-80k tokens is roughly the point where I start wrapping up with giving agent required context even on the small to medium sized requests.

That would leave almost no tokens for actual work


GLM is the first open source model that actually worked for me, where I found the output ok.

And yes, sonnet/opus is better and what I use daily. But I wouldn’t be that upset if I had to drop down to GLM.


IDK about GLM but GPT 5.4 Extra High has been great when I've used it in the VS Code Copilot extension, I see no actual reason Opus should consume 3x more quota than it the way it does

I think it offers a very good tradeoff of cost vs competency

4.7 is better, but its also wildly expensive


You're probably just holding it wrong.

They are produced by companies that specialise in producing ICs.

They can be placed manually or automated.


If you can print small enough with this technology I'm pretty sure you can make transistors - sort of 1980 era transistors, not very dense, but if you are printing bulk materials you can build in 3D rather than 2D, make interesting numbers of transistors, cpus in everything!

Home-printed 80s transistors would unlock a lot of analog-style guitar pedals. :D

That’s users, for ya! They will always find ways to use your product that you didn’t intend to or even knew about!

The funny thing is that the website only has firebase auth, without any ai features. The default api key that was created (before the ai was even released a few years back), someone got it from the website and started using the gemini api with the key.

Yet my phone service provider is able to cut off my internet access from the Kb I go over the limit…

I had to do this with my MacBook Pro models early 2015 and late 2017.

It seems like there was a period in time when solder just wasn’t done well, it seems like.


IIRC this is to do with the phase in of RoHS and bad lead free solder


Is a future of what it could look like or what it will look like?


Bringing kintsugi into this conversation is like saying “being underwater can be quite advantageous!” and linking a video on fish, when the main topic is about people drowning in the ocean.


Art is everywhere, and starts with a simple philosophy of making things slightly less awful everyday. Initially focused on your own mind, body, and soul... then recognizing you were always part of something a lot bigger and older than most imagine.

I do appreciate your poetic tone though =3


Bad bot


You will learn this fact in time.

And bots are not "good" or "bad", but rather an imperfect mirror of statistically salient nonsense. =3


> You will learn this fact in time.

> And bots are not "good" or "bad", but rather an imperfect mirror of statistically salient nonsense. =3

I think that you are a human and surprised to see you be calm when someone calls you bot because I usually flip out.

If you are a human, I think that one of the things that I am starting to think to do is reply with I am human only after all video.

I am only human after all: https://www.youtube.com/watch?v=L3wKzyIN1yk

Or in this case about fishes, we can have the video parody of I am just a fish!

I am just a fish: https://www.youtube.com/watch?v=1goAp0XmhZQ

(this last video is a parody-ish but really great music unironically out of the original music being I am just a freak, both music are really great in my opinion unironically haha!)

I am just a fish!


There seems to be a webgl render engine suitable for vega [0]. Have you tried and if so, what was your experience?

[0] https://github.com/vega/vega-webgl-renderer


The opening paragraph gives the reason why Linux is still not at the same level as windows:

“ A few months and several headaches later…”

Additionally, the first comment I read is very positive, yet it also gives insight into the same situation.


> The opening paragraph gives the reason why Linux is still not at the same level as windows:

> “ A few months and several headaches later…”

These are the words of someone that hasn't tried to install Windows on a recent machine. There's plenty of headaches there too.


I have, actually. It was a several-click process with few to no headaches (I don’t recall any in the last several installs I did).


Lol, I had to hunt for drivers for a while and then research which ones match my hardware, then I had to research how to strip windows 11 of the more egregious privacy intrusions and nags ... in my case there was plenty of headaches.


Given how many people here are saying that Windows is great after you do enough tweaking it does not seem very different.


Yet it is. I don’t see myself or anyone in my family effortlessly choosing and setting up a Linux distro without outside help.

Windows on the other hand is easy to install and set up. I’d argue most people in my family can do this and I’m the only one that is technically inclined.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: