Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Would you mind sharing some of these projects?

I've found Claude's usefulness is highly variable, though somewhat predictable. It can write `jq` filters flawlessly every time, whereas I would normally spend 30 minutes scanning docs because nobody memorizes `jq` syntax. And it can comb through server logs in every pod of my k8s clusters extremely fast. But it often struggles making quality code changes in a large codebase, or writing good documentation that isn't just an English translation of the code it's documenting.





It is always "I'm producing 300 projects in a nanosecond" but it's almost never about sharing or actually deploying these ;)

The problem I had that the larger your project gets, the more mistakes Claude makes. I (not a parent commenter) started with a basic CRUD web app and was blown away by how detailed it was, new CSS, good error handling, good selection and use of libraries, it could even write the terminal commands for package management and building. As the project grew to something larger Claude started forgetting that some code already existed in the project and started repeating itself, and worse still when I asked for new features it would pick a copy at random leaving them out of sync with eachother. Moving forward I've been alternating between writing stuff with AI, then rewriting it myself.

> The problem I had that the larger your project gets, the more mistakes Claude makes

I think the reason for this is because these systems get all their coding and design expertise from training, and while there is lots of training data available for small scale software (individual functions, small projects), there is much less for large projects (mostly commercial and private, aside from a few large open source projects).

Designing large software systems, both to meet initial requirements, and to be maintainable and extensible over time, is a different skill than writing small software projects, which is why design of these systems is done by senior developers and systems architects. It's perhaps a bit like the difference between designing a city and designing a single building - there are different considerations and decisions being made. A city is not just a big building, or a collection of buildings, and large software system is not just a large function or collection of functions.


Yeah, great analogy. Thanks!

Well said, and good analogy.

Have it produce CLAUDE.md files in every directory giving a summary of what code is where, and a system directive to keep these updated.

I’ve got Claude on 10k loc projects, which is probably mid range.

But I also have it document and summarise its own work.


> I also have it document and summarise its own work.

Could you share some of your prompts or CLAUDE.md? I'm still learning what works.


So, just like a human on a growing codebase?

Here's mine fully deployed, https://hackernewsanalyzer.com/. I use it daily and have some users. ~99.7% LLM code. About 1 hour to first working prototype then another 40 hours to get it polished and complete to current state.

It shows, quite an interesting wrapper over GPT with unauthorized access to prompting it you assembled there ;) Very much liked the part where it makes 1000 requests pulling 1000 comments from the firebase to the client and then shoots them back to GPT via supabase

take care


To be clear, roughly 39.8 hours of just prompting and output to make this website?

41 hours total of prompting, looking at code diffs, reverting, reprompting, and occasional direct code commits. I do review the full code changes nearly every step of the way and often iterate numerous times until I'm satisfied with the resulting code approach.

Have you tried to go back to the old way, maybe just as an experiment, to see how much time you are actually saving? You might be a little surprised! Significant "reprompting" time to me indicates maybe a little too much relying on it rather than leading by example. Things are much faster in general if you find the right loop of maybe using Claude for like 15%-20% of stuff instead of 99.7%. You wouldn't give your junior 99.7% ownership of the app unless they were your only person, right? I find spending time thinking through certain things by hand will make you so much more productive, and the code will generally be much better quality.

I get that like 3 years ago we were all just essentially proving points building apps completely with prompts, and they make good blog subjects maybe, but in practice they end up being either fragile novelties or bloated rat's nests that end up taking more time not less.


I’ve done things in days that in the before times would have took me months. I don’t see how you can make that time difference up.

I have at least one project where I can make that direct comparison - I spent three months writing something in the language I’ve done most of my professional career in, then as a weekend project I got ChatGPT to write it from scratch in a different language I had never used before. That was pre-agentic tools - it could probably be done in an afternoon now.


I'm not a fulltime developer, but manage a large dev team now. So, this project is basically beyond my abilities to code myself by hand. Pre llm, I would expect in neighborhood of 1.5-2 months for a capable dev on my team to produce this and replicate all the features.

If you haunt the pull requests of projects you use I bet you'll find there's a new species of PR:

> I'm not an expert in this language or this project but I used AI to add a feature and I think its pretty good. Do you want to use it?

I find myself writing these and bumping into others doing the same thing. It's exciting, projects that were stagnant are getting new attention.

I understand that a maintainer may not want to take responsibility for new features of this sort, but its easier than ever to fork the project and merge them yourself.

I noticed this most recently in https://github.com/andyk/ht/pulls which has two open (one draft) PRs of that sort, plus several closed ones.

Issues that have been stale for years are getting traction, and if you look at the commit messages, it's AI tooling doing the work.

People feel more capable to attempt contributions which they'd otherwise have to wait for a specialist for. We do need to be careful not to overwhelm the specialists with such things, as some of them are of low quality, but on the whole it's a really good thing.

If you're not noticing it, I suggests hanging out in places where people actually share code, rather than here where we often instead brag about unshared code.


> People feel more capable to attempt contributions

That does not mean that they are more capable, and that's the problem.

> We do need to be careful not to overwhelm the specialists with such things, as some of them are of low quality, but on the whole it's a really good thing.

That's not what the specialists who have to deal with this slop say. There have been articles about this discussed here already.


What would you have us do, Keep the fixes to ourselves?

Yes, keep AI slop “fixes” to yourself and only create PRs for your own work.

At this point my prior is that all these 300/ns projects are some kind of internal tools, with very narrow scope and many just for a one-off use.

Which is also fine and great and very useful and I am also making those, but it probably does not generalize to projects that require higher quality standards and actual maintenance.


Sure, but 80% of software is probably internal and short lived like that, if not more.

Solving the problems of the business that isn’t a software business.


Places that aren't software businesses are usually the inverse. The software is extremely sticky and will be around for ages, and will also bloat to 4x the features it was originally supposed to have.

I worked at an insurance company a decade ago and the majority of their software was ancient. There were a couple desktops in the datacenter lab running Windows NT for something that had never been ported. They'd spent the past decade trying to get off the mainframe and a majority of requests still hit the mainframe at some point. We kept versions of Java and IBM WebSphere on NFS shares because Oracle or IBM (or both) wouldn't even let us download versions that old and insecure.

Software businesses are way more willing to continually rebuild an app every year.


I also see a lot of this so I can't blame you for thinking it! See my other post about some projects build _only_ using LLMs.

https://news.ycombinator.com/item?id=46133458


There's a massive incentive not to share them. If I wrote a project using AI I'd be reluctant to publish it at all because of the backlash I've seen people get for it.

People are and always were reluctant to share their own code just the same. There is nothing to be gained, the chances of getting positive reviews from fellow engineers are slim to none. We are a critical and somewhat hypocritical bunch on average.

Building something is easy

Building something that works ? Not so easy

Pushing that thing in production ? That the hardest part


I came with receipts

Claude has taught me so much about how to use jq better. And really, way more efficient ways of using the command line in general. It's great. Ironically, the more I learn the less I want to ask it to do things.

In an ideal world we function in exactly this way - using LLMs to bootstrap our skill/knowledge improvement journeys.

Yeah, if you pay attention to its output you can pick up little tips and tricks all over the place.

Not the OP you're replying to, but I've put together quite a few projects using only LLMs, no hand crafted code anywhere (I couldn't do it!)

https://dnbfamily.com

https://eventme.app

https://blazingbanana.com

https://play.google.com/store/apps/details?id=com.blazingban...

https://play.google.com/store/apps/details?id=com.blazingban...

https://play.google.com/store/apps/details?id=com.blazingban...

Are they perfect? No probably not, but I wouldn't have been able to make any of these without LLMs. The last app was originally built with GPT-3.5.

There is a whole host of other non-public projects I've built with LLMs, these are just a few of the public ones.


Maybe the most depressing part of all this is if people start thinking they would not have been able to do things without the LLM. Of course they would have, it's not like LLMs can do anything that you cannot. Maybe it would have taken more time at least the first time and you would have learned a few things in the process.

Sure, I can write all of it. But I simply won’t. I have Claude generated Avalonia C# applications and there is no way I would have written the thousands of lines of xaml they needed for the layouts. I would just have done it as a console app with flags.

Surely nobody writes this XAML by hand?

But reducing friction, eliminating the barrier to entry, is of fundamental importance. It's human psychology; putting running socks next to your bed at night makes it like 95% more likely you'll actually go for a run in the morning.

Yes, "I couldn't have bothered..." is different from " I wouldn't have been able to make...".

You might not go for a run when the socks are not there, but I don't think you would start questioning your ability to run.


It would be more depressing if our imagination didn't exceed the finite time we have to learn and master new skills.

Or if we stopped imagining.

I understand the point, and to some degree agree. For myself, I really couldn't (not to say it wouldn't have been possible). I tried many many times over so many years and just didn't have the mental stamina for it, it would never "click" like infra/networking/hardware does etc and I would always end up frustrated.

I have learnt so much in this process, nowhere near as much as someone that wrote every line (which is why I think being a good developer will be a hot commodity) but I have had so much fun and enjoyment, alongside actually seeing tangible stuff get created, at the end of the day, that's what it's all about.

I have a finite amount of time to do things, I already want to do more than I can fit into that time, LLMs help me achieve some of them.


This is a "scratch an itch" project I initially started to write manually in the past, but never finishing. I then used claude to do it basically on the side while watching the world series http://nixpkgs-pr-explorer.s3-website-us-west-2.amazonaws.co...

It’s not just good for small code bases. In the last six months I’ve built a collaborative word processor with its own editor engine and canvas renderer using Claude, mostly Opus. It’s practically a mini Google Docs, but with better document history and an AI agent built in. I could never have built this in 6 months by myself without Claude Code.

https://revise.io

I think if you stick with a project for a while, keep code organized well, and most importantly prioritize having an excellent test suite, you can go very far with these tools. I am still developing this at a high pace every single day using these tools. It’s night and day to me, and I say that as someone who solo founded and was acquired once before, 10 years ago.


https://github.com/lawless-m

You can see by Contributors which ones Claude has done.

I have no idea if the code is any good, I’ve never looked at it and I have no idea how to code in Rust or Racket or Erlang anyway.


> I have no idea if the code is any good, I’ve never looked at it and I have no idea how to code in Rust or Racket or Erlang anyway.

In that case, are you really producing multiple projects per week? If you've never looked at the code, have you verified that they work?


yes, I am using my voice agent, my head tracker, my sql writer, my odbc client, my shopping list, my sharepoint file uploader, my Timberborn map generator, my wireguard routing, my oxygen not included launch scripts, my i3wm config, my rust ATA over Ethernet with Content Addressable storage

the list could go on


The former tasks are directly from the training material, directly embedded into the model. For the latter task, it needs a context window and intelligence.

At the end of this week we are releasing https://github.com/eqtylab/cupcake

You can see all of Claude’s commits.

I’ve shipped so much with ai.

Favorite has been metrics dashboards of various kinds - across life and business.


Something about restricting an AI agent by using another AI to write code to restrict it is quite funny

It'll be a common paradigm. Some agents support the coding agent discover relevant context for a plan, others will help the agent stay on track and ensure no rules break.

They really should have been supplying at least a week worth of readymade "projects" to every freelance AI promoter out there to demonstrate x9000 AI productivity gains for the skeptics.

Because vibing the air about those gains without any evidence looks too shilly.


As opposed to ad hominems and handwave dismissals, which are highly credible?

Pointing out the where the burden of proof lies is not an ad hominem. Calling it such is in fact a good example of poisoning the well. all the fan girls have to do is post links to code they have vibe coded. some people have even done that in this thread. it's not an unreasonable standard.

Where did I attack the parent personally? I just described a steady pattern when someone makes drive-by statements and never replies to back them up.

I handwave dismiss bigfoot, too



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: