Hacker Newsnew | past | comments | ask | show | jobs | submit | adam_arthur's commentslogin

LLMs have clearly accelerated development for the most skilled developers.

Particularly when the human acts as the router/architect.

However, I've found Claude Code and Co only really work well for bootstrapping projects.

If you largely accept their edits unchanged, your codebase will accrue massive technical debt over time and ultimately slow you down vs semi-automatic LLM use.

It will probably change once the approach to large scale design gets more formalized and structured.

We ultimately need optimized DSLs and aggressive use of stateless sub-modules/abstractions that can be implemented in isolation to minimize the amount of context required for any one LLM invocation.

Yes, AI will one shot crappy static sites. And you can vibe code up to some level of complexity before it falls apart or slows dramatically.


>If you largely accept their edits unchanged, your codebase will accrue massive technical debt over time and ultimately slow you down vs semi-automatic LLM use.

Worse, as its planning the next change, it's reading all this bad code that it wrote before, but now that bad code is blessed input. It writes more of it, and instructions to use a better approach are outweighed by the "evidence".

Also, it's not tech debt: https://news.ycombinator.com/item?id=27990979#28010192


People can take on debt for all sorts of things. To go on vacation, to gamble.

Debt doesn't imply it's productively borrowed or intelligently used. Or even knowingly accrued.

So given that the term technical debt has historically been used, it seems the most appropriate descriptor.

If you write a large amount of terrible code and end up with a money producing product, you owe that debt back. It will hinder your business or even lead to its collapse. If it were quantified in accounting terms, it would be a liability (though the sum of the parts could still be net positive)

Most "technical debt" is not buying the code author anything and is materialized through negligence rather than intelligently accepting a tradeoff


All those examples were borrowing money. What you're describing as "technical debt" doesn't involve borrowing anything. The equivalent for a vacation would be to take your kids to a motel with a pool and dress up as Mickey Mouse and tell them its "Disney World debt". You didn't go in debt. You didn't go to Disney World. You just spent what money you do have on a shit solution. Your kids quite possibly had fun, even.

> term technical debt has historically been used

There are plenty of terms that we no longer use because they cause harm.


Agreed.

What I've found is that AI can be alright at creating a Proof of Concept for an app idea, and it's great as a Super Auto-complete, but anything with a modicum of complexity, it simply can't handle.

When your code is hundreds of thousands of lines, asking an agent to fix a bug or implement a feature based on a description of the behavior just doesn't work. The AI doesn't work on call graphs, it basically just greps for strings it thinks might be relevant to find things. If you know exactly where the bug lies, it can usually find it with context given to it, but at that point, you're just as good fixing the bug yourself rather than having the AI do it.

The problem is that you have non-coders creating a PoC, then screaming from the rooftops how amazing AI is and showing off what it's done, but then they go quiet as the realization sets in that they can't get the AI to flesh it out into a viable product. Alternatively, they DO create a product that people start paying to use, and then they get hacked because the code is horribly insecure and hard-codes API keys.


> We ultimately need optimized DSLs and aggressive use of stateless sub-modules/abstractions that can be implemented in isolation to minimize the amount of context required for any one LLM invocation.

Containment of state also happens to benefit human developers too, and keep complexity from exploding.


Yes!

I've found the same principles that apply to humans apply to LLMs as well.

Just that the agentic loops in these tools aren't (currently) structured and specific enough in their approach to optimally bound abstractions.

At the highest level, most applications can be written in simple, plain english (expressed via function names). Both humans and LLMs will understand programs much better when represented this way


The most interesting thing for me is that I am sure it does.

I have been coding for 20+ years and I have used AI agents for coding a lot, especially for the last month and a half. I can't say for sure they make me faster.They definitely do for some tasks, but over all? I can solve some tasks really quickly, but at the same time my understanding of the code is not as good as it was before. I am much less confident that is is correct.

LLMs clearly make junior and mid level engineers faster, but it is much harder to say for Senior.


Valknut is pretty good at forcing agents to build more maintainable codebases. It helps them dry out code, separate concerns cohesively and organize complexity. https://github.com/sibyllinesoft/valknut

> LLMs have clearly accelerated development for the most skilled developers.

Have they so clearly? What's the evidence?


Most people's "truth" nowadays is what they've heard enough people say is true. Not objective data/measures. What people believe is true, and say is true, IS truth, to them.

> accrue massive technical debt

The primary difference between a programmer and an engineer.


> We ultimately need optimized DSLs and aggressive use of stateless sub-modules/abstractions that can be implemented in isolation to minimize the amount of context required for any one LLM invocation

Wait till you find out about programming languages and libraries!

> It will probably change once the approach to large scale design gets more formalized and structured

This idea has played out many times over the course of programming history. Unfortunately, reality doesn’t mesh with our attempts to generalize.


Relying on the model for security is not security at all.

No amount of hardening or fine-tuning will make them immune to takeover via untrusted context


Comment definitely reads like AI


Implies they will pay cash value to equity holders as opposed to issuing NVDA shares.

(Electronically)


Is this to somehow screw the employees with RSUs or what?


No, it doesn't really matter if they pay in cash or stock. If you think NVDA has room to run you're welcome to use your buyout money to buy NVDA on the open market.


Well, this isn't framed as a buyout/takeover, so I was curious how existing RSUs would be cashed out?

This deal is framed as IP transfer and talent transfer without owning the full company. Probably to skirt anti trust, among other things.


I'm not sure in this specific case. They could choose to pay the employees some portion of the funds.

If not, the owners are likely liable to be sued for "selling in effect" without paying equity holders.

Presuming the company becomes a defacto subsidiary of Nvidia (even if not legally so)

My guess, without researching it, is they will compensate existing equity holders to avoid that possibility. I mean the valuation multiple is enormous, it's worth it simply to derisk the legal aspect.


For vested RSUs it's likely that the Groq husk will pay out the $20B as a dividend or buyback or something. I don't know if unvested RSUs are accelerated or just canceled. Of course the employees will receive new RSUs when they join Nvidia.


Local tools/skills/function definitions can already invoke any API.

There's no real benefit to the MCP protocol over a regular API with a published "client" a local LLM can invoke. The only downside is you'd have to pull this client prior.

I am using local "skill" as reference to an executable function, not specifically Claude Skills.

If the LLM/Agent executes tools via code in a sandbox (which is what things are moving towards), all LLM tools can be simply defined as regular functions that have the flexibility to do anything.

I seriously doubt MCP will exist in any form a few years from now


Storage doesn't require the same capex/upfront investment to get that margin.

How much does it cost to train a cutting edge LLM? Those costs need to be factored into the margin from inferencing.

Buying hard drives and slotting them in also has capex associated with it, but far less in total, I'd guess.


  How much does it cost to train a cutting edge LLM? Those costs need to be factored into the margin from inferencing.
They don't, though! I can buy hardware off of the shelf, host open source models on it, and then charge for inference:

https://parasail.io, https://www.baseten.co


Yes, which is why the companies that develop the models aren't cost viable. (Google and others who can subsidize it at a loss obviously are excepted)

Where is the return on the model development costs if anybody can host a roughly equivalent model for the same price and completely bypass the model development cost?

Your point is inline with the entire bear thesis on these companies.

For any use cases which are analytical/backend oriented, and don't scale 1:1 with number of users (of which there are a lot), you can already run a close to cutting edge model on a few thousand dollars of hardware. I do this at home already


Open source models are still a year or so behind the SotA models released the last few months. The price to performance is definitely in favor of Open Source models however.

DeepMind is actively using Google’s LLMs on groundbreaking research. Anthropic is focused on security for businesses.

For consumers it’s still a better deal for a subscription than to invest a few grand in a personal LLM machine. There will be a time in the future where diminishing returns shortens this gap significantly, but I’m sure top LLM researchers are planning for this and will do whatever they can to keep their firm alive beyond the cost of scaling.


Definitely

I am not suggesting these companies can't pivot or monetize elsewhere, but the return on developing a marginally better model in-house does not really justify the cost at this stage.

But to your point, developing research, drugs, security audits or any kind of services are all monetization of the application of the model, not the monetization of the development of new models.

Put more simply, say you develop the best LLM in the world, that's 15% better than peers on release at the cost of $5B. What is that same model/asset worth 1 year later when it performs at 85% of the latest LLM?

Already any 2023 and perhaps even 2024 vintage model is dead in the water and close to 0 value.

What is a best in class model built in 2025 going to be worth in 2026?

The asset is effectively 100% depreciated within a single year.

(Though I'm open to the idea that the results from past training runs can be reused for future models. This would certainly change the math)


For sure, all these companies are racing to have the strongest model, and as time goes on we quickly start reaching diminishing returns. DeepSeek came out at the beginning of this year, blew everyone's minds, and now look at how far the industry has progressed beyond it.

It doesn't even seem like these companies are in a battle of attrition to not be the first to go bankrupt. Watching this would be a lot more exciting if that was the case! I think if there was less competition between LLMs developers could slow down, maybe.

Looking at the prices of inference of open-source models, I would bet proprietary models are making a nice margin on API fees, but there is no way OpenAI will make their investors whole because they make a few dollars of revenue for a million tokens. I am terrified of the world we will live in if OpenAI will be able to reverse their balance sheet. I think there's no where else that investors want to put their money.


The other nightmare for these companies, is that any competitor can use their state of the art model for training another model. As some Chinese models are suspected to do. I personally think it's only fair, since those companies in the first place trained on a ton of data and nobody agreed to it. But it shows that training the frontier models have really low returns on investment


Yes you’re right. Capex spend is definitely higher.

In the end it comes all down to the value provided as you see in the storage example.


Probably it's "operationally profitable" when ignoring capex, depreciation, dilution and other required expenses to stay current.

Of course that means it's unprofitable in practice/GAAP terms.

You'd have to have a pretty big margin on inference to make up for the model development costs alone.

A 30% margin on inference for a GPU that will last ~7 years will not cut it


I'm confused by all the takes implying decode is more important than prefill.

There are an enormous number of use cases where the prompt is large and the expected output is small.

E.g. providing data for the LLM to analyze, after which it gives a simple yes/no Boolean response. Or selecting a single enum value from a set.

This pattern seems far more valuable in practice, than the common and lazy open ended chat style implementations (lazy from a product perspective).

Obviously decode will be important for code generation or search, but that's such a small set of possible applications, and you'll probably always do better being on the latest models in the cloud.


The US nuked Japan in the 1940s and only a short few years later were close allies.

Similarly, Germany was the de facto enemy of the Ally aligned world in the 40s and only a few short decades later best friends with most.

You can identify countless similar atrocities in US history after which relations stabilized within a few years.

Most in life are short term oriented and it's rare that these things produce lasting effects on perception.

Does anyone care or even remotely think about what Bush did as president 20 years ago? Some politically oriented and historically minded folks yes, 99% no.

They really have to be long lasting and persistent transgressions to produce generational distrust e.g. Japan invading China/Korea many times over the last few hundred years.

However the new gen seems far less concerned about this too.

Given Trump's larger than life character/ego/presence, it's more likely that anything he does will be attributed to him instead of the country as a whole. Which makes his actions perhaps even less impactful than a more neutral presenting president doing the same.


> The US nuked Japan in the 1940s and only a short few years later were close allies.

I don't think that's a great example. A few years after the bombs, Douglas MacArthur was ruling Japan. They didn't have much of a choice. Japan was occupied for almost 7 years, had their constitution rewritten to make them essentially reliant on the US for security.


I'm talking about the sentiment of the populace toward the US, not which legislators or leaders were in charge.

It's undeniable that by and large Japanese sentiment towards the US bounced back quickly, despite a traumatic and expansive war


That goodwill is almost entirely due to the US' need to rebuild the Japanese economy rapidly to counter threats from China, North Korea and the Soviet Union. That happened through favourable trade deals and mass outsourcing of automotive and electronics jobs and diverting capital away from domestic military as per the US-written constitution of Japan.

Western and Japanese companies alike started moving production outside Japan once wages started eating into profit margins. Today, underemployment is common in Japan and the low birth rate is one of several symptoms of the economic stagnation that began in the early '90s. Populist governments won't be far behind.


I don't think Japan after WW2 is the right comparison here.

My understanding is that generally they expected a MUCH more severe set of penalties and occupation after losing - especially given the unconditional surrender, and instead got a stable and functional provisional government for the next 7 years.

---

Basically - Japanese sentiment is not a parallel to this. That time:

1. Japan started hostilities with a surprise attack

2. Lost, complete with unconditional surrender

3. Was then occupied by a government that was more stable, left much of the existing civilian infrastructure in place, and forgave many key figures (not the least of which was the emperor)

4. Then that government helped them roll out "new deal" style social reforms.

That is absolutely, utterly at odds with the current situation. There - the Allied powers were relatively graceful, culturally aware, and interested in a stable, functional government.

Here - We're insulting our friends, from a position where there's no moral high ground to stand on. Personally, I don't think they'll forget so quickly, and I think things will get much worse if the US continues down this path.

Can we rebuild those relationships? Sure, seems likely on a long enough timespan, but it takes a hell of a lot more effort to get it back than it does to throw away.


Those two examples relied on heavy investment over many years by the worlds sole superpower to bounce back. See the Marshall Plan and MacArthur’s reconstruction of Japan. These things do not happen automatically, and unlike those examples, there doesn’t seem to be a rich, well-run superpower waiting in the wings to lift anyone back up. I’m sure some relations will normalize post-47 but economic might and attractiveness to the world may not.


The US military physically occupied and restructured the governments of Japan and Germany. A better example might be Vietnam, which has become a US partner but it took a couple generations.


It took about 50 years before India started to trust that relations with America could be positive. Then it took a lot of trust and concessions on both sides to strengthen that over the next 20 years under Bush, Obama, Trump and Biden - all of whom went out of their way to court India as an ally and a counterweight to China.

Then Trump decided his shitty Nobel Prize mattered more and threw 20 years of hard work in the dump.

The consensus in India is that America is perfidious. You claim that countries have goldfish memories, but that doesn’t seem to be the case with Indo-American relations. It was immensely hard to build this relationship and easy to burn it.

But who knows? Maybe you have an insight into how Indian people think.


I think it's important to note that they were friends after a regime change.


Like the US presidential elections every four years?


Going from an empire/reich to a democracy ain't quite the same thing as a presidential election. The USA will have a hard time recovering from institutional damage being done to it right now.


It's irrelevant whether change in policy is due to losing a war, revolution, or democratic election.

If the numbers swing 5%, the US may suddenly have a radical change in foreign policy and governorship.

Seen as recently as the 2024 election.

Regardless, I'm discussing the opinion of the populace, not the political leadership.

You can't endear a country's people to you by installing a puppet government


There's a regime change like "the second TV debate was a clown show, most swing states saw +5% move away from the incumbent, with the usual stronghold states still holding their established positions, and the latest poll listed inflation as voters' top concern."

...and then there's a regime change like "your capital city and every industrial center is firebombed to oblivion, kindergarteners are begging on streets, soldiers are coming back from POW camps to find their home burnt to ground and their whole family dead, and the occupying forces are executing high ranking officers of the previous regime for war crimes, just to drive the point home that their ideology will never be tolerated ever again."


> Does anyone care or even remotely think about what Bush did as president 20 years ago?

If they don't they are idiots. To name a few things: 9/11, the foreign policy disaster and money pit of the Iraq War/GWOT, the easily-avoidable GFC.

Those were both significant events in their own right and fundamental causes of the Trump presidencies.


No, uncertainty and stability matter in business, a lot. Not that Trump is the reason we lack political stability. It’s the two party system Washington warned us about, boaz and jachin. But that lack of stability is arguably pretty stable itself. American dynamism has pros and cons. Much of today’s context goes back to political and trade decisions made back in the 90s.


Google Search has AI responses at the top of the fold.

Eventually those answers will be sufficient for most and give people no reason to move to alternatives.

Allowing them to pay to be default seems to mostly guarantee this outcome


Those answers with the awful model they're using are frequently wrong and aggravating to have displace search results.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: