Hacker Newsnew | past | comments | ask | show | jobs | submit | stuaxo's commentslogin

Even if it wasn't sycophantic this sort of thing should be checked outside of an LLM but how was he to know that ?

I don't know: we brand LLMs as "artificial intelligence", like it's the computer out of star trek.

On this website most of us know what an LLM is and how it works, but I could well see a young person asking it for advice and not knowing it could say the wrong thing.


Is this Palentir related ?

Palantir is best viewed as a React dev shop.

They have a bunch of dumb react components that can become the gui for live targeting systems, but that all happens post contract and in the liability of their customers, the organizations willing to expend human capital coming and going, e.g. militaries


Ah fuel and food - those classic unimportant things.

Fuel and food are excluded from core inflation not because they're unimportant (they are in fact incredibly important) but because they are much more volatile in price--going up and down in bigger increments--so that you get a more stable view of inflation by excluding them.

But it's a bit of a nasty trick because food, in particular, has inflated in price a lot the past 3 years. Some items, like sugar, are legitimately double the price they were.

Well - you converge to a system, but do that by pruning what you don't want.

If you care about maintainability and quality (and I include maintaining using LLM based tools) then you need to understand what it does (in doing so you will find lots of things for it to fix - you'll probably find that the architecture it's chosen is not right for what you want too).


Devs still have to maintain this code, the Python devs can definitely get the LLM to write (some kind of) Rust, but when it goes wrong and you hit the wall with the LLM then they can will have to learn Rust which might take a while, this sounds like a bit of a project risk.

A cynic would put this together with the plan to close all the NHSs source code: since the public wouldn't like further Palentir integration + be able to see it happening.

The whole model of slicing the app into so many pieces (proprietary wrappers it seems like) to make more opportunities for billing is terrible.

What harness are you using ?

I'm going to switch to local LLMs for most stuff soon.


Overall using screentime as the metric, derived from some imperfect logging and vibes it's about 50% OpenCode 15% Continue 15% my homebrew bullshit 13% Claude Code and 7% Cline. I've been deep on agentic stuff lately (1.3wks aka 3 months of AI time), there are only so many hours in the day to duplicate work and AB test, but in the past I've sworn by Qwen Coder + llama.vim and I still enjoy that workflow for deep work far more than I like prompting agents, but there's a lot of dross I'm learning to delegate.

Interesting.

I stopped doing local stuff for a bit when I realised I didn't know how well it is supposed to work so have been on Claude for a few months now.

I think I'll try OpenCode this time.

Usually I do stuff in devcontainers, qwen code (non local) was the only time I managed to lose some work as it got confused when I ran out of tokens.

There's still quite a way to go - it does seem like Claude code itself is pretty badly coded, so I think there is a space for open source to come in with a high quality harness at some point.


Harnessed seem to be a big part of what makes stuff good or not.

I tried Cline and couldn't get it working well and part of this was that at the time it expected OpenAIs output format.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: