Hacker Newsnew | past | comments | ask | show | jobs | submit | try-working's commentslogin

I wanted to like Discworld and read a couple of the books but they're frankly not funny.

He definitely grew significantly as a writer over time, and I would agree that some of his early work isn't particularly strong (The Light Fantastic, for example, is relatively bog-standard comic fantasy without any of the depth his later work showed).

If you start reading at the very beginning of the Discworld, you're slogging through the weaker stuff, and it's easy to get discouraged. A smoother path is to pick one of the defined sub-series (the guards are very popular, but my vote goes to the witches) and start along just that track; you'll get to the strong stuff much faster.


My advice has always been to start with Small Gods. It is a standalone book that references but does not rely on any others, and is far enough into his career that it’s fair to say that if you don’t like it, you won’t like his work in general.

You must have had a very good surgeon then! Congrats!

Looks alright for a "first" but there's no reason for anyone to really use until they open source it.

it's an agent product that you interact with over email. it has skills and stuff. it has multiplayer. so it's something like openclaw, but different.

you can counter the context rot and requirement drift that is experienced here by many users by using a recursive, self-documenting workflow: https://github.com/doubleuuser/rlm-workflow

I use a self-documenting recursive workflow: https://github.com/doubleuuser/rlm-workflow

you can use lurkkit.com to build your own chronological youtube feed with only your subscriptions

> Traditional Chinese relies on context: “Rain heavy, not go”, “雨大,不去了”. > Modern Chinese demands explicit logic: “Because the rain is heavy, therefore I will not go.””因为雨下得很大,所以我决定不去了。”

I would say "下雨了,我不去“ or something like that. The second example is perhaps what a language learner would say in order to "speak correctly", but nobody actually speaks or writes like that.


Totally. I also feel such a disconnect with HSK material, no one speaks like that or even uses that vocabulary. But I guess thats the case with almost every language/language course.

What's gone unnoticed with the Gemma 4 release is that it crowned Qwen as the small model SOTA. So for the first time a Chinese lab holds the frontier in a model category. It is a minor DeepSeek model, because western labs have to catch up with Alibaba now.

on my 16 GB GPU Gemma 4 is better and faster than Qwen 3.5, both at 4-bit

so it's not so clear cut


depends on usage, Gemma 4 is better on visuals/html/css and language understanding (Which probably plays a role in prompting). But it's worse at code in general compared to Qwen 3.5 27B.

Which in the series specifically?

It's unnoticed because it didn't. In Google's own benchmarks they are on par, and I've seen 3rd party benchmarks where Qwen beats G4 with high margin

The day a western anything will need to catch up with alibaba will be a notable day indeed. Also, this will never happen.

most codebases dont have traces to train on. if you use rlm-workflow you will build up rich traceability in the form of requirements, plans, implementation artifacts, along with worktree diffs. with these, you can then use self-distillation on models or use autoagent to improve your harness. https://github.com/doubleuuser/rlm-workflow

For anyone that believes Chinese labs will stop open sourcing their models, let me tell you why that won't happen.

First, try signing up for Z.ai's coding plan. I know how to but I bet you won't be able to.

The absolute disaster that is Z.ai's internet presence shows that these small labs have no ability to market themselves and drive direct sales.

For marketing, they lack capabilities, and releasing open models is the only way for them to remain in the conversation.

For sales, they rely on distribution via OpenRouter, OpenCode etc. Interest with their users is driven by open model performance.

Open sourcing for Chinese labs is not some large national scheme. It is their only way to commercialization.


Well, can’t they just direct their model to do some marketing for them?

Only partially tongue in cheek - if it’s not good at marketing itself, that seems like a red flag for capabilities?


> First, try signing up for Z.ai's coding plan. I know how to but I bet you won't be able to.

What's the issue with signup up for Z.ais coding plan?


you can't find it

1. Go to https://z.ai

2. Top right corner, "API"

3. https://z.ai/subscribe


China can't get good chips. But I don't understand why they can't license their closed source models to US inference providers so we can get more than 80% reliability on their models on OpenRouter.

I think they already are licensing their biggest models to third party inference providers.

Agree, it’s surprising that these companies don’t thoroughly test their own workflows to ensure a smooth and seamless user experience.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: