This resonates. I've shipped 7 side projects in the past year using AI heavily. The code writing part got 10x faster. But I've noticed something counterintuitive: the total time from idea to shipped product barely decreased.
Why? Because the bottleneck was never typing code. It was always understanding the problem, making architectural decisions, debugging edge cases, and most importantly - knowing what NOT to build.
AI made me faster at producing code, but it also made me produce MORE code, which means more surface area for bugs, more maintenance burden, more complexity to reason about. The discipline of "write less code" is harder now because writing code costs almost nothing.
The engineers who thrive will be the ones who can resist the temptation to over-engineer when the marginal cost of adding complexity drops to near zero.
We need to have more metrics for this. Like I hear people making this claim on HN all the time as if they know absolutely for sure but I doubt it's this simple.
I can guarantee you this... the story is not absolute. Depending on who you are and what you need to work on dev time could be slower, same or faster for you. BUT what we don't know is the proportion. Is it faster for 60% of people? 70%, 80%?
This is something we don't know for sure yet. But i suspect your instinct is completely wrong and that 90% of people are overall faster... much faster. I do agree that it produces more bugs and more maintenance hurdles but it is that much faster.
The thing is LLMs can bug squash too. AND they are often much faster at it then humans. My agentic set up just reads the incoming slack messages on the issue, makes a ticket, fixes the code and creates a PR in one shot.
> Why? Because the bottleneck was never typing code. It was always understanding the problem, making architectural decisions, debugging edge cases, and most importantly - knowing what NOT to build.
For me, this is a bit different. Writing code has always been the bottleneck. I get most of my joy out of solving edge cases and finding optimizations. My favorite projects are when I’m given an existing codebase with the task, “When mars and venus are opposite eachother, the code gets this weird bug that we can’t reproduce.”
When a project requires me to start from scratch, it takes me a lot longer than most other people. Once I’ve thought of the architecture, I get bored with writing the implementation.
AI has made this _a lot_ easier for me.
I think the engineers who thrive wi be the ones know when to use what tool. This has been the case before AI, AI is just another tool allowing more people to thrive.
Fascinating- I've always loved the big picture, architecture, and I've also loved stable software but have one hell of a time fixing bugs. AI helped me a ton with that.
Well if you're ever in need for a complementary mind in side projects- huh, how does one connect over HackerNews?
I have my own side project that I vibe coded. I probably did what would take one team 6 montns and produced it myself in one month.
I'm not afraid of breaking stuff because it is only a small set of users. However for my own code for my professional job no way I would go that fast because I would impact millions of users.
It is insane that companies think they can replace teams wholesale while maintaining quality.
A missing link right now is automated high-quality code reviews. I would love an adversarial code review agent that has a persona oriented around all incoming code being slop, that leverages a wealth of knowledge (both manually written by the team and/or aggregated from previous/historical code reviews). And that agent should pull no punches when reviewing code.
This would augment actual engineer code reviews and help deal with volume.
Yeah I vibe coded an addition game for my 4 year old that lets him do addition problems where the answer is always 10 or less. It’s very “juicy”. There’s a lot of screen shake and spinning and flashy rainbow insanity going on. If I had done all that stuff myself it would have take a week because I would have been picky about each little animation. The thing that saved me the most time was just being ok with the good enough animations the ai spit out.
It’s amazing for him and it works on his iPad.
However when I tried it on my iPhone it was a broken mess. Completely unusable (not because of screen size differences).
I tried getting Claude to fix it but it couldn’t do it without changing too much of the look and feel, so I dug into the code and it was thousands of lines of absolute madness. I know from using this at work that there are things I could have done. Write tests to lock in things I like etc…
But so much of the speed up was about not caring about the specifics that once I started caring about making an actual product, I was not much faster maybe not any faster at all. The bottleneck in writing a game was never in banging out code.
> I probably did what would take one team 6 montns and produced it myself in one month.
I find it… Amusing? That’s not quite the word. That programmers—a group notoriously for making wrong estimates of how long something will take to build—continuously and confidently spew a version of this.
And it’s not even estimating how long we ourselves would take to build something, now we’re onto estimating what an undetermined team of completely made up strangers could do. It’s bonkers. It has no basis in reality.
The issue is that before AI, 1% of the population was capable of creating 1 side project per year. After AI, 10% of the population is capable of creating 10 side projects per year. The competition grew by 100x. The pessimist in me thinks that the window of opportunity to create something successful is shrinking.
Why do you look at it that way? Why does anyone beside you have to care about what you do?
Just build something for yourself. You will always have things you'd like to build for yourself. You will be in competition with yourself only and your target audience will be yourself.
Market forces do not apply to side-projects, because that's what people do for fun.
Just because there are chess computers, doesn't mean that no one plays chess anymore at home.
Isn't it obvious? The reward that a personal project can generate for you is limited. It's not remotely close to what a successful project would give you - money, fulfillment, social capital, feeling good about yourself, etc.
Ironically I had a very smart and otherwise reasonable math professor who, shortly after Kasparov lost to Deep Blue, said in class that chess was no longer interesting.
Maybe, but LLMs solve but one issue (maybe two). Take me, for example. I am highly proficient regarding software development in most aspects. Except for that tiny problem: I wouldn't even know what to build. And at least for me, LLMs could not help with that.
The whole side project or even private project thing doesn't just hinge on being able to produce software. There's a lot more.
It's like the business of selling electric drills. People don't really want drills they want holes. But holes are difficult to sell so the selling the drills is a proxy for that.
In software it's the same thing. People don't really want software they want data and data transformation. But traditionally the proxy for that has been selling the software (either as a desktop app or then later as sole kind of service).
You could argue that in either case the proxy is not what people want but yet because of the difficulty of selling the "actual" thing the proxy market has flourished.
We're now inventing a new tool that will completely disrupt that market and any software business that is predicated on the complexity required to create the software to transform the data is going to get severely disrupted. Software itself will be worthless.
Are most side projects in competition? I wouldn't think so.
Even if they were I disagree that 10x more ideas being produced means 10x more products in competition. You could leverage AI to execute but still have terrible ideas, leadership, product stewardship etc.
I think some clever people with a real and valuable insight will finally be able to turn that insight into a product. I also think the other 9 products will be get rich quick attempts by people with nothing to offer.
Yes it become much easier to fail fast and iterate, but also a lot of these fail fast projects are trivial for anyone to implement themselves. Differentiating your project is going to be tougher too.
A lot of the moats are gone, but quality (and security) is in a nose dive. AI built project might be the Ikea furniture. Good for the masses, but there's still a market (much smaller) for well crafted applications and services. It's hard to say what it'll look like in a couples years though. Maybe even the crafting is eventually gone. /shrug
I think we need to change our perspective of what success is. I believe there will be a ton of small companies popping up instead of a few big ones that eats everyone's lunch. Like Google, Microsoft and others giants have done until now.
I can relate. Sincerely debating whether I quit my well-paying and comfortable corporate job and just go full-time entrepreneur before the opportunities disappear.
When the goal is to ship (the result) I'll happily leverage LLM's to try an idea or 3 out. However, it wouldn't be fair to claim that my side projects have exactly one goal. That's why I choose to use AI generated code when I deal with stuff that I already know how to do, done a lot of times, and the only thing that I gain from using AI is time typing it out.
Anything else? I'll struggle and grow as a developer, thanks. And before anyone says "but there are architecture decisions etc. so you still grow"... those existed anyways. If I have to practice, I'll practice micro AND macro skills.
This tracks with the way a lot of heavily vibecoded projects have issues with beeing feature heavy, while those features often don't fully work and most importantly don't fit together cohesively. In other words, the quality is low.
> AI made me faster at producing code, but it also made me produce MORE code, which means more surface area for bugs, more maintenance burden, more complexity to reason about
I think from time to time, it's better to ask the AI whether the codebase could be cleaned and simplified. Much better if you use different AI than what you use to make the project.
Well how many side projects did you ship last year? I’ve written small programs in the last few months over a weekend that would have taken me a month to do a couple years ago, and they’re better. Not in terms of code quality, but in terms of features I wanted and knew how to implement but couldn’t be bothered, Opus can do in one minute and even if it’s not the optimal implementation it’s completely functional, fine, and costs me almost nothing.
You are putting sentences together just like an LLM would - quite fitting for an AI generated article. You might want to get it checked out, these days you never know if you are a real person or not.
> The engineers who thrive will be the ones who can resist the temptation to over-engineer when the marginal cost of adding complexity drops to near zero.
One area --and many may not like that fact-- where it can help greatly is that the cost of adding tests also drops to near zero and that doesn't work against us (because tests are typically way more localized and aren't the maintenance burden production code is). And a some us were lazy and didn't like to write too many tests. Or take generative testing / fuzzy testing: writing the proper generators or fuzzers wasn't always that trivial. Now it could become much easier.
So we may be able to use the AI slop to help us have more correct code. Same for debugging edge cases: models can totally help (I've had case as simple as a cryptic error message which I didn't recognize: passed it + the code to a LLM and it could tell me what the error was).
But yup it's a given that, as you put it, when the marginal cost of adding complexity drops to near zero, we're opening a whole new can of worms.
TFA is AI slop but fundamentally it may not be incorrect: the gigantic amount of generated sloppy code needs to be kept in check and that's where engineering is going to kick in.
I mean if you've built 7 side projects (and we assume it's the same phase since total time from idea to shipped product barely decreased), how are these things still a bottleneck to you? I'm assuming you're building in a domain/language you're comfortable with by now (unless you're crazy and try something fundamentally different on each of those shipped products).
Why will the 8th project still have those things as the bottleneck given your experience?
Also if you're not seeing any real gains in productivity, why are you using AI for your side projects and wasting tokens/money?
I use AI for side projects because Google gives me a stupid large number of tokens that refresh every 6-24 hours on my existing $10/mo Google One plan. I see it as my civic duty to help increase their costs by producing slop that I generally throw away anyways because it doesn't actually work after it gets generated.
At work, I was told to use AI but it doesn't actually work for anything that I couldn't have handed off to a brand new undergraduate intern. So I use it for things that I don't care about then go spend twice as long rewriting what it output because it made the task longer by being wrong.
There's nothing new about this pattern. When the tractor was invented, the farmer didn't get to knock off early. He just started producing 10x more. Then the tractors got bigger and more powerful, and the things you used them with got more sophisticated too and suddenly you're producing 100x more.
Its worth mentioning that this essay has some signs of being either partially AI generated or heavily edited through an LLM. Some of the signs are there (It's not X, it's Y), With the blog having gone from nearly zero activity between 2015 and 2025 to have it explode in posts and text output since then also raises an eyebrow.
I think it is very fair to say that in the same way that LLM's have given english majors access to programming, LLMs have also given engineers access to clear communication.
I'm not shy to admit that LLMs even from 2 years ago could communicate ideas much better than me, especially for a general audience.
As someone who has written a few deeply personal articles with LLM assistance, I see the signs and I'm almost certain this was generated off a few bullet points. The repetition and cadence strongly resembles the LLM output. Its the kind of fluff that I remove from a piece, because it lacks humanity and offers little substance.
Even the linkedin profile has a studio-ghibli-style avatar. People are going to assume that he is just an "analog interface" to an LLM. Which is sad, because he might be a good programmer. In fact, I tend to see a lot of english-as-second-language people embrace LLMs as a kind of "equalizer", not realizing that in 2026 it is the opposite (not saying that it's right either way, just pointing out that it is becoming a kind of anti-marketing, like showing up to a conference without any clothing, and getting banned from the conference permanently).
We should probably normalize publishing things in our native languages, and expecting the audience to run it through a translator. (I have been toying with the idea of writing everything in Esperanto (not my native language, but a favorite) and just posting links to auto-translated English versions where the translation is good enough).
EDIT: as someone with friends and family from Eastern Europe, I can tell you that the prevailing attitude is: "everything is bullshit anyway" (which, to be fair, has a lot of truth to it), and so it is no surprise that people would enthusiastically embrace a pocket-sized bullshit factory, hook it up to a fire-hose, and start spraying. We saw it with spam, and we see it now with slop. It won't stop unless the system stops rewarding it.
This was my thought after getting through a few paragraphs as well. At first, I was thinking, this is interesting, maybe worth sharing with colleagues. But then it became too obvious it was AI written or "assisted". Can't take that seriously.
It's funny how seemingly easy it is to tell articles like this have that AI generated whiff to them. The first bit that raised my suspicion was the "The Identity Crisis Nobody Talks About" headline. This "The x nobody talks about" feels like such a GenAI thing.
It is almost 90% generated using AI text. So many paragraphs to say basically nothing at all.
Like look at this paragraph:
> Junior engineers have traditionally learned by doing the simpler, more task-oriented work. Fixing small bugs. Writing straightforward features. Implementing well-defined tickets. This hands-on work built the foundational understanding that eventually allowed them to take on more complex challenges.
The first sentence was enough to convey everything you needed to know, but it kept on adding words in that AI cadence. The entire post is filled with this style of writing, which, even if it is not AI, is extremely annoying to read.
My point is that there's nothing to be written there "instead", it just is not needed text that is added to make the text longer, typical of AI writing that parrots the same points over and over to make up for word count.
Here's another example from the blog:
> Here is something that gets lost in all the excitement about AI productivity: most software engineers became engineers because they love writing code.
> Not managing code. Not reviewing code. Not supervising systems that produce code. Writing it. The act of thinking through a problem, designing a solution, and expressing it precisely in a language that makes a machine do exactly what you intended. That is what drew most of us to this profession. It is a creative act, a form of craftsmanship, and for many engineers, the most satisfying part of their day.
can just be:
> Most software engineers became engineers because they love writing code. It is a creative act, a form of craftsmanship, and for many engineers, the most satisfying part of their day.
Clarity is something that is taught in every writing class but AI generated text always seems to have this weird cadance as follows: The sound is loud. Not a whimper, not a roar, a simple sound that is very loud. And that's why... blah blah blah.
You have to care about your readers if you're writing something seriously. Throwing just a bunch of text that all mean the same thing in your writing is one of the bigger sins you can do, and that's why most people hate reading AI writing.
The part you'd like to remove ("Not managing code...") may be not required to convey the objective meaning of the sentence, but humans have emotions, too. I could have written stuff like that. To build up a bigger emotional picture.
> The act of thinking through a problem, designing a solution, and expressing it precisely in a language that makes a machine do exactly what you intended.
This sentence may not be relevant for whatever you experience to be the relevant message of the text. But it still says something the remaining paragraph does not. And also something I can relate to.
Also, as LLMs are statistical models, one has to assume that they write like this because their training data tells them to. Because humans write like this. Not when they do professional writing maybe, but when they just ramble. Not all blogs are written by professionals. I'd say most aren't. LLM training data consists mostly of humans rambling.
I also sometimes write long comments on the internet. And while I have no example to check, I feel like I do write such sentences, expanding on details to express more emotional context. Because I'm not a robot and I like writing a lot. I think it's a perfectly human thing to do. I find it sad that "writing more than absolutely needed" is now regarded as a sign of AI writing.
One of the good book about writing I read was William Zinsser's "On Writing Well". Striving for simplicity and avoiding clutter was the two first principles described in the book. AI writing feels more like ramblings than communication.
I feel like it's such a lack of self respect and respect for others when people write using AI on personal blogs.
Reading AI code is very pleasant. It's well annotated and consistent - how I like to read code (although not how I write code LOL). Reading language/opinions is not meant to be this way. It becomes repetitive, boring, and feels super derivative. Why would you turn the main way we communicate with each other into a soulless, tedious, chore?
I think with coding it's because I care* about what the robot is doing. But, with communication, I care about what the person is thinking in their mind, not through the interpretation of the robot. Even if the person's mind isn't as strong. At least then I can size the person up - which is the other reason understanding each other is important and ruined when you put a robot in between.
I saw someone point out something like: ai makes every sentence count. There’s no building or allowing a point to breathe. Every sentence is an axiom to get the meaning across, and its so grating
One problem I have seen IRL is AI deployment mistakes and IMO Vibe Coders need an IT/Dev Father Figure type to avoid these simple mistakes. Here is one example:
A surgeon (no coding experience) used Claude to write a web app to track certain things about procedures he had done. He deployed the app on a web hosting provided (PHP LAMP stack). He wanted to share it with other doctors, but wasn't sure if it was 'secure' or not. He asked me to read the code and visit the site and provide my opinion.
The code was pretty reasonable. The DB schema was good. And it worked as expected. However, he routinely zipped up the entire project and placed the zip files in the web root and he had no index file. So anyone who navigated to the website saw the backups named Jan-2026.backup, etc. and could download them.
The backups contained the entire DB, all the project secrets, DB connection strings, API credentials, AWS keys, etc.
He had no idea what an 'index' file was and why that was important. Last I heard he was going to ask Claude how to secure it.
The post is right superficially. It made being an engineer harder because it took away the easy parts that anyone can do and it forces engineers to think of the hard ones.
No jobs get easier with automation - they always move a step up in abstraction level.
An accountant who was super proficient in adding numbers no longer can rely on those skills once calculator was invented.
This is the key. I haven't found that things have become harder. The hard parts are still hard, and those have been the most important and prominent parts of my job once I reached a certain level.
I've unfortunately stopped reading articles before reading comments here as it's all mostly garbage now. I'm not sure what people are trying to accomplish with generating blogs aside from either clout farming or marketing for their companies.
What I never enjoyed was looking up the cumbersome details of a framework, a programming language or an API. It's really BORING to figure out that tool X calls paging params page and pageSize while Y offset and limit. Many other examples can be added.
For me, I feel at home in so many new programming languages and frameworks that I can really ship ideas. AI really helps with all the boring stuff.
Agree, it’s made programming so much fun. The other day I wrote a C# app just because it was the best language for the job, I’ve never touched .Net in my life. Worked great, clients loved it.
I can actually build nice UIs as a traditional ML engineer (no more streamlit crap). People are using them and genuinely impressed by them
I can fly through Rust and C++ code, which used to take ages of debugging.
The main thing that is clear to me is that most of the ecosystem will likely converge toward Rust or C++ soon. Languages like Python or Ruby or even Go are just too slow and messy, why would you use them at all if you can write in Rust just as fast? I expect those languages to die off in the next several years
The scenario I'm somewhat worried about is that instead of 1 PM, 1 designer and 5 developers, there will be 1 PM, 1 designer and 1 developer. Even if tech employment stays stable or even slightly increases due to Jevons paradox, the share of software developers in tech employment will shrink.
It might be worth mentioning studies that show the lack of productivity gains from LLM usage. These posts take it as an unequivocal given. Management might still have the expectations that certain tasks are faster. But they aren’t always connected to reality because they’re not thinking as engineers.
AI made it so individual developers can outsource their work, not just companies. Maybe there are some lessons to be learned from companies that manage outsourced work successfully.
> Yeah but a manager can do those things. You don't need an engineer for that.
Can they really? Engineering is about keeping the whole picture in mind so that you know which lever to push and which to not push for a certain goal. Trying until you're lucky can get you to that goal, but it's costly and not sustainable. So you need someone that can work out a model for experimentation in a less costly manner.
Judgment in this case is about deciding which path to direct the project, tradeoffs is being aware that there are other paths that are better in some aspects. And responsibility is acknowledging that a bad decision will bear a personal cost.
Everyone does the above in their own domain. But I don't think I've ever see a manager wanting to do it in the engineering domain. It's more about pushing the engineer to accept the responsibility, but denying them the power of judgment.
The author introduces the term "Supervision Paradox", but IMHO this is simply one instance of the "Automation Paradox" [1], which has been haunting me since I started working in IT.
Interestingly, most jobs don't incentivize working harder or smarter, because it just leads to more work, and then burn-out.
You seem to be right. The author is pumping out one such article per day. I think I've spent more time in forming my comment than they did in generating the article. Oh well :)
> ...most software engineers became engineers because they love writing code. Not managing code. Not reviewing code. Not supervising systems that produce code. Writing it...
An SWE who bases their entire identity and career around writing code is not an engineer - they are a code monkey.
The entire point of hiring a Software ENGINEER is to help translate business requirements into technical requirements, and then implement the technical requirements into a tangible feature or product.
The only reason companies buy software is because the alternative means building in-house, and for most industries software is a cost-center not a revenue generator.
I don't pay (US specific) 200K-400K TCs for code monkeys, I pay that TC for Engineers.
And this does a disservice to the large portion of SWEs and former SWEs (like me) who have been in the industry because we are customer-outcome driven (how do we use code to solve a tangible customer need), not write pretty code.
You might be missing that a lot of companies are giddy that the mgmt can just vibe code stuff and there's no opportunity for engineers to be involved, (except for when it crashes?). I use AI tools and they are nice, but the mgmt are mostly not logical and need someone to sort through their bullshit.
Yeah no. Almost all companies I've chatted with - from MSPs to C-Suite of F10s - expect and demand humans-in-the-loop. I'm also on a couple boards and we've aligned on the same expectation as well.
Look, AI/ML and especially LLMs are powerful, but there does remain a degree of instability and non-determinism which will require human intervention to remediate.
That said, there is a lot of dev work in companies that is a cost-center, and those are the portions that will start getting vibe coded and deployed in product with little-to-no oversight (eg. a support portal for SMBs at an enterprise), but the equivalent feature would have already been an afterthought even without LLMs and probably given to a couple SWEs we'd be fine re-orging in a quarter anyhow.
> Yes just not driven or owned by engineers. That's what I'm seeing from company and a few peer's companies
I mean, it depends on the feature/product and how critical it is to the health of the business.
Like I mentioned in my edited comment, there is a lot of dev work in companies that is a cost-center, and those are the portions that will start getting vibe coded and deployed in product with little-to-no oversight (eg. a support portal for SMBs at an enterprise), but the equivalent feature would have already been an afterthought even without LLMs and probably given to a couple SWEs we'd be fine re-orging in a quarter anyhow because we cannot justify spending $750K a year (the backend cost of 3 FT SWEs for a company) on a customer form which nets $0 in revenue and is not directly tied with pipeline generation.
This section very much resonated with me, even though I still haven't tried any of the AI tools:
... most software engineers became engineers because they love writing code. Not managing code. Not reviewing code. Not supervising systems that produce code. Writing it. The act of thinking through a problem, designing a solution, and expressing it precisely in a language that makes a machine do exactly what you intended. That is what drew most of us to this profession. It is a creative act, a form of craftsmanship, and for many engineers, the most satisfying part of their day.
Actually surprised none of the other comments have picked up on this, as I don't think it's especially about AI. But the periods of my career when I've been actually writing code and solving complicated technical problems have been the most rewarding times in my life, and I'd frequently work on stuff outside work time just because I enjoyed it so much. But the other times when I was just maintaining other people's code, or working on really simple problems with cookie-cutter solutions, I get so demotivated that it's hard to even get started each day. 100%, I do this job for the challenges, not to just spend my days babysitting a fancy code generation tool.
I've always been motivated by making simple solid foundations in my code the fastest way possible.
So for me being able to have AI wrote certain things extremely fast with me just doing voice to text with my specific approach, is amazing.
I am all in on everything AI and have a discord server just for openclaw and specialized per repo assistants. It really feels like when I'm busy I can throw it an issue tracker number for things.
Then I will ssh via vs code or regular ssh which forwards my ssh key from 1password. My agents have read only repo access and I can push only when I ssh in. Super secure. Sorry for the tangent to the article but I have always loved coding now I love it even more.
I'm not sure if it's made engineering harder, but it's certainly changing what it means to be a good engineer. It's no longer just about writing code. Now it's increasingly about having good taste, making the right decisions, and sometimes just being blessed with the Midas touch.
In any case, I think we should start treating the majority of code as a commodity that will be thrown away sooner or later.
I feel like there's a market out there for a weekly newsletter that summarises all the AI takes like this and collects the one meaningful snippet of insight (if any)
AI learned this figure of speech from humans. Even the frequency in which it is used is copied from humans. So you can't really use it to determine if something is written by an AI or not.
LLMs might follow the frequencies of the training data in their raw form, but nobody uses raw LLMs, they use models which have been RLHFed to hell and back to bias them towards specific patterns. Then newer models were trained on the output of those RLHFed models, and further RLHFed, and so on, and so on.
The humans tasked with RLHF tuning aren't tuning to their own preference or personal style, their job is to tune it how they've been told to tune it. You don't get a subserviant, sycophantic chat interface without putting your thumb on the scale, hard.
If you think that the article is written by human or that is is unclear, please go ahead. Others here on HN also have pointed out that the author shoots out such lengthy blog posts every day. And you can also see the typical emoji AI slop here: https://www.ivanturkovic.com/services/
But I have no issue with your argumentation whatsoever, it is just that I think there is more than sufficient evidence, and you think there is not.
I still feel like I'm writing code. I tell Claude what to write and I am very specific about it. There's still tons of problems for which Claude has no particular solution and it's on me and other humans to figure out what to do. For those cases where I tell it to just go off and write a whole script that I'm not even looking at, those are throwaway / low-value cases I dont care about where previously I'd not have even taken on that particular job.
Developers will become admins. Responsible for supervising and owning the outcomes of increasingly agentic engineering outputs. Trust is the most important thing in business and it’s worth more than ever.
There's always a grain of truth in everything, but the recent article by the Redis guy (sorry for the lack of name) resonated more with me. It's correct that the load in other areas is increasing also because these tools are not there yet when it comes to for lack of a better word "good taste". I work with someone who hasn't written a line of code in a year and it shows and I'm about tired dealing with the slop. But also there's a bunch of things at work that you either did a million times already, aren't really challenging problems just annoying problems hard to solve because of all the cruft, a lot of boring manual work etc. and for this it's just an amazing help to the point I am more relaxed at work than I was previously. And when it does something that is not quite there, I can either fix it manually or tell it to fix it and it usually "gets it". Of course it it ultimately replaces me I will not be relaxed but that's a different topic.
Another little thing that resonated was a tweet that said "some will use it to learn everything and some so that they don't have to learn anything ". Of course it's not really a hard truth. It's questionable how much you can learn without really getting your hands dirty. But I do think people looking at it as a tool that helps then and/or makes them better will profit more than people looking to cut corners.
Not really, I disagree. The article did slightly touched on the real issue on why people enjoy writing code, a “craftsmanship”, yes, coding is NOT engineering, it is writing, and the people who enjoy doing it are actually writers not engineers, and I always keep mentioning that. With AI however, those writers have to be doing the engineering work: the goals, architecture design, managing blueprints, process design and refining, among many other things, and that job is not easy hence why engineers are “supposedly” paid well, AI now took the writing role, and you have to do the engineering one.
Why? Because the bottleneck was never typing code. It was always understanding the problem, making architectural decisions, debugging edge cases, and most importantly - knowing what NOT to build.
AI made me faster at producing code, but it also made me produce MORE code, which means more surface area for bugs, more maintenance burden, more complexity to reason about. The discipline of "write less code" is harder now because writing code costs almost nothing.
The engineers who thrive will be the ones who can resist the temptation to over-engineer when the marginal cost of adding complexity drops to near zero.
I can guarantee you this... the story is not absolute. Depending on who you are and what you need to work on dev time could be slower, same or faster for you. BUT what we don't know is the proportion. Is it faster for 60% of people? 70%, 80%?
This is something we don't know for sure yet. But i suspect your instinct is completely wrong and that 90% of people are overall faster... much faster. I do agree that it produces more bugs and more maintenance hurdles but it is that much faster.
The thing is LLMs can bug squash too. AND they are often much faster at it then humans. My agentic set up just reads the incoming slack messages on the issue, makes a ticket, fixes the code and creates a PR in one shot.
For me, this is a bit different. Writing code has always been the bottleneck. I get most of my joy out of solving edge cases and finding optimizations. My favorite projects are when I’m given an existing codebase with the task, “When mars and venus are opposite eachother, the code gets this weird bug that we can’t reproduce.”
When a project requires me to start from scratch, it takes me a lot longer than most other people. Once I’ve thought of the architecture, I get bored with writing the implementation.
AI has made this _a lot_ easier for me.
I think the engineers who thrive wi be the ones know when to use what tool. This has been the case before AI, AI is just another tool allowing more people to thrive.
Well if you're ever in need for a complementary mind in side projects- huh, how does one connect over HackerNews?
I'm not afraid of breaking stuff because it is only a small set of users. However for my own code for my professional job no way I would go that fast because I would impact millions of users.
It is insane that companies think they can replace teams wholesale while maintaining quality.
Tech-savvy people might understand this feeling, but those who are responsible for hiring will easily proceed with another candidate that goes fast.
When push comes to shove, then, programmers will opt to have food to eat over handling technical debt generation.
This would augment actual engineer code reviews and help deal with volume.
It’s amazing for him and it works on his iPad.
However when I tried it on my iPhone it was a broken mess. Completely unusable (not because of screen size differences).
I tried getting Claude to fix it but it couldn’t do it without changing too much of the look and feel, so I dug into the code and it was thousands of lines of absolute madness. I know from using this at work that there are things I could have done. Write tests to lock in things I like etc…
But so much of the speed up was about not caring about the specifics that once I started caring about making an actual product, I was not much faster maybe not any faster at all. The bottleneck in writing a game was never in banging out code.
I find it… Amusing? That’s not quite the word. That programmers—a group notoriously for making wrong estimates of how long something will take to build—continuously and confidently spew a version of this.
And it’s not even estimating how long we ourselves would take to build something, now we’re onto estimating what an undetermined team of completely made up strangers could do. It’s bonkers. It has no basis in reality.
Why do you look at it that way? Why does anyone beside you have to care about what you do?
Just build something for yourself. You will always have things you'd like to build for yourself. You will be in competition with yourself only and your target audience will be yourself.
Market forces do not apply to side-projects, because that's what people do for fun.
Just because there are chess computers, doesn't mean that no one plays chess anymore at home.
This is just a correction of something that managed to remain in an invalid state for an impressively long time.
If you have no reaction at all, you probably weren't paying attention.
Eventually though, people _should_ recover and return after having processed the changes. So maybe the professor was still recovering at the time?
The whole side project or even private project thing doesn't just hinge on being able to produce software. There's a lot more.
In software it's the same thing. People don't really want software they want data and data transformation. But traditionally the proxy for that has been selling the software (either as a desktop app or then later as sole kind of service).
You could argue that in either case the proxy is not what people want but yet because of the difficulty of selling the "actual" thing the proxy market has flourished.
We're now inventing a new tool that will completely disrupt that market and any software business that is predicated on the complexity required to create the software to transform the data is going to get severely disrupted. Software itself will be worthless.
Even if they were I disagree that 10x more ideas being produced means 10x more products in competition. You could leverage AI to execute but still have terrible ideas, leadership, product stewardship etc.
I think some clever people with a real and valuable insight will finally be able to turn that insight into a product. I also think the other 9 products will be get rich quick attempts by people with nothing to offer.
A lot of the moats are gone, but quality (and security) is in a nose dive. AI built project might be the Ikea furniture. Good for the masses, but there's still a market (much smaller) for well crafted applications and services. It's hard to say what it'll look like in a couples years though. Maybe even the crafting is eventually gone. /shrug
Anything else? I'll struggle and grow as a developer, thanks. And before anyone says "but there are architecture decisions etc. so you still grow"... those existed anyways. If I have to practice, I'll practice micro AND macro skills.
I think from time to time, it's better to ask the AI whether the codebase could be cleaned and simplified. Much better if you use different AI than what you use to make the project.
You can use it to discuss about what you should build, identify edge cases, ask you questions to force you to take decisions, etc.
Were you able to fairly split test?
One area --and many may not like that fact-- where it can help greatly is that the cost of adding tests also drops to near zero and that doesn't work against us (because tests are typically way more localized and aren't the maintenance burden production code is). And a some us were lazy and didn't like to write too many tests. Or take generative testing / fuzzy testing: writing the proper generators or fuzzers wasn't always that trivial. Now it could become much easier.
So we may be able to use the AI slop to help us have more correct code. Same for debugging edge cases: models can totally help (I've had case as simple as a cryptic error message which I didn't recognize: passed it + the code to a LLM and it could tell me what the error was).
But yup it's a given that, as you put it, when the marginal cost of adding complexity drops to near zero, we're opening a whole new can of worms.
TFA is AI slop but fundamentally it may not be incorrect: the gigantic amount of generated sloppy code needs to be kept in check and that's where engineering is going to kick in.
Why will the 8th project still have those things as the bottleneck given your experience?
Also if you're not seeing any real gains in productivity, why are you using AI for your side projects and wasting tokens/money?
At work, I was told to use AI but it doesn't actually work for anything that I couldn't have handed off to a brand new undergraduate intern. So I use it for things that I don't care about then go spend twice as long rewriting what it output because it made the task longer by being wrong.
That this kind of writing puts a great number of us off is not important to many who seek their fortune in this industry.
I hear the cry: "it's my own words the LLM just assisted me". Yes we have to write prompts.
I'm not shy to admit that LLMs even from 2 years ago could communicate ideas much better than me, especially for a general audience.
We should probably normalize publishing things in our native languages, and expecting the audience to run it through a translator. (I have been toying with the idea of writing everything in Esperanto (not my native language, but a favorite) and just posting links to auto-translated English versions where the translation is good enough).
EDIT: as someone with friends and family from Eastern Europe, I can tell you that the prevailing attitude is: "everything is bullshit anyway" (which, to be fair, has a lot of truth to it), and so it is no surprise that people would enthusiastically embrace a pocket-sized bullshit factory, hook it up to a fire-hose, and start spraying. We saw it with spam, and we see it now with slop. It won't stop unless the system stops rewarding it.
I hate it. I couldn't read much more after that.
Like look at this paragraph:
> Junior engineers have traditionally learned by doing the simpler, more task-oriented work. Fixing small bugs. Writing straightforward features. Implementing well-defined tickets. This hands-on work built the foundational understanding that eventually allowed them to take on more complex challenges.
The first sentence was enough to convey everything you needed to know, but it kept on adding words in that AI cadence. The entire post is filled with this style of writing, which, even if it is not AI, is extremely annoying to read.
Here's another example from the blog:
> Here is something that gets lost in all the excitement about AI productivity: most software engineers became engineers because they love writing code.
> Not managing code. Not reviewing code. Not supervising systems that produce code. Writing it. The act of thinking through a problem, designing a solution, and expressing it precisely in a language that makes a machine do exactly what you intended. That is what drew most of us to this profession. It is a creative act, a form of craftsmanship, and for many engineers, the most satisfying part of their day.
can just be:
> Most software engineers became engineers because they love writing code. It is a creative act, a form of craftsmanship, and for many engineers, the most satisfying part of their day.
Clarity is something that is taught in every writing class but AI generated text always seems to have this weird cadance as follows: The sound is loud. Not a whimper, not a roar, a simple sound that is very loud. And that's why... blah blah blah.
You have to care about your readers if you're writing something seriously. Throwing just a bunch of text that all mean the same thing in your writing is one of the bigger sins you can do, and that's why most people hate reading AI writing.
The part you'd like to remove ("Not managing code...") may be not required to convey the objective meaning of the sentence, but humans have emotions, too. I could have written stuff like that. To build up a bigger emotional picture.
> The act of thinking through a problem, designing a solution, and expressing it precisely in a language that makes a machine do exactly what you intended.
This sentence may not be relevant for whatever you experience to be the relevant message of the text. But it still says something the remaining paragraph does not. And also something I can relate to.
Also, as LLMs are statistical models, one has to assume that they write like this because their training data tells them to. Because humans write like this. Not when they do professional writing maybe, but when they just ramble. Not all blogs are written by professionals. I'd say most aren't. LLM training data consists mostly of humans rambling.
I also sometimes write long comments on the internet. And while I have no example to check, I feel like I do write such sentences, expanding on details to express more emotional context. Because I'm not a robot and I like writing a lot. I think it's a perfectly human thing to do. I find it sad that "writing more than absolutely needed" is now regarded as a sign of AI writing.
I would read the hell out of Joyce’s Perl 5 documentation, but only after six or seven beers.
I don’t think there will be a point in coming to this site if it’s just going to be slop on the front page all the time.
Maybe mods should consider a tag or flag for AI generated content submissions?
Reading AI code is very pleasant. It's well annotated and consistent - how I like to read code (although not how I write code LOL). Reading language/opinions is not meant to be this way. It becomes repetitive, boring, and feels super derivative. Why would you turn the main way we communicate with each other into a soulless, tedious, chore?
I think with coding it's because I care* about what the robot is doing. But, with communication, I care about what the person is thinking in their mind, not through the interpretation of the robot. Even if the person's mind isn't as strong. At least then I can size the person up - which is the other reason understanding each other is important and ruined when you put a robot in between.
Looks like something AI would say. Regardless of how it really was written
Admittedly it was so long and basic, I stopped halfway.
A surgeon (no coding experience) used Claude to write a web app to track certain things about procedures he had done. He deployed the app on a web hosting provided (PHP LAMP stack). He wanted to share it with other doctors, but wasn't sure if it was 'secure' or not. He asked me to read the code and visit the site and provide my opinion.
The code was pretty reasonable. The DB schema was good. And it worked as expected. However, he routinely zipped up the entire project and placed the zip files in the web root and he had no index file. So anyone who navigated to the website saw the backups named Jan-2026.backup, etc. and could download them.
The backups contained the entire DB, all the project secrets, DB connection strings, API credentials, AWS keys, etc.
He had no idea what an 'index' file was and why that was important. Last I heard he was going to ask Claude how to secure it.
No jobs get easier with automation - they always move a step up in abstraction level.
An accountant who was super proficient in adding numbers no longer can rely on those skills once calculator was invented.
This is the key. I haven't found that things have become harder. The hard parts are still hard, and those have been the most important and prominent parts of my job once I reached a certain level.
What I never enjoyed was looking up the cumbersome details of a framework, a programming language or an API. It's really BORING to figure out that tool X calls paging params page and pageSize while Y offset and limit. Many other examples can be added. For me, I feel at home in so many new programming languages and frameworks that I can really ship ideas. AI really helps with all the boring stuff.
AI makes using them a breeze.
I can actually build nice UIs as a traditional ML engineer (no more streamlit crap). People are using them and genuinely impressed by them
I can fly through Rust and C++ code, which used to take ages of debugging.
The main thing that is clear to me is that most of the ecosystem will likely converge toward Rust or C++ soon. Languages like Python or Ruby or even Go are just too slow and messy, why would you use them at all if you can write in Rust just as fast? I expect those languages to die off in the next several years
The scenario I'm somewhat worried about is that instead of 1 PM, 1 designer and 5 developers, there will be 1 PM, 1 designer and 1 developer. Even if tech employment stays stable or even slightly increases due to Jevons paradox, the share of software developers in tech employment will shrink.
Maybe this is not entirely true yet, but it most likely will be in the near future.
Can they really? Engineering is about keeping the whole picture in mind so that you know which lever to push and which to not push for a certain goal. Trying until you're lucky can get you to that goal, but it's costly and not sustainable. So you need someone that can work out a model for experimentation in a less costly manner.
Judgment in this case is about deciding which path to direct the project, tradeoffs is being aware that there are other paths that are better in some aspects. And responsibility is acknowledging that a bad decision will bear a personal cost.
Everyone does the above in their own domain. But I don't think I've ever see a manager wanting to do it in the engineering domain. It's more about pushing the engineer to accept the responsibility, but denying them the power of judgment.
Interestingly, most jobs don't incentivize working harder or smarter, because it just leads to more work, and then burn-out.
[1] https://en.wikipedia.org/wiki/Automation#Paradox_of_automati...
An SWE who bases their entire identity and career around writing code is not an engineer - they are a code monkey.
The entire point of hiring a Software ENGINEER is to help translate business requirements into technical requirements, and then implement the technical requirements into a tangible feature or product.
The only reason companies buy software is because the alternative means building in-house, and for most industries software is a cost-center not a revenue generator.
I don't pay (US specific) 200K-400K TCs for code monkeys, I pay that TC for Engineers.
And this does a disservice to the large portion of SWEs and former SWEs (like me) who have been in the industry because we are customer-outcome driven (how do we use code to solve a tangible customer need), not write pretty code.
Look, AI/ML and especially LLMs are powerful, but there does remain a degree of instability and non-determinism which will require human intervention to remediate.
That said, there is a lot of dev work in companies that is a cost-center, and those are the portions that will start getting vibe coded and deployed in product with little-to-no oversight (eg. a support portal for SMBs at an enterprise), but the equivalent feature would have already been an afterthought even without LLMs and probably given to a couple SWEs we'd be fine re-orging in a quarter anyhow.
> but there does remain a degree of instability and non-determinism which will require human intervention to remediate.
I agree.
I mean, it depends on the feature/product and how critical it is to the health of the business.
Like I mentioned in my edited comment, there is a lot of dev work in companies that is a cost-center, and those are the portions that will start getting vibe coded and deployed in product with little-to-no oversight (eg. a support portal for SMBs at an enterprise), but the equivalent feature would have already been an afterthought even without LLMs and probably given to a couple SWEs we'd be fine re-orging in a quarter anyhow because we cannot justify spending $750K a year (the backend cost of 3 FT SWEs for a company) on a customer form which nets $0 in revenue and is not directly tied with pipeline generation.
... most software engineers became engineers because they love writing code. Not managing code. Not reviewing code. Not supervising systems that produce code. Writing it. The act of thinking through a problem, designing a solution, and expressing it precisely in a language that makes a machine do exactly what you intended. That is what drew most of us to this profession. It is a creative act, a form of craftsmanship, and for many engineers, the most satisfying part of their day.
Actually surprised none of the other comments have picked up on this, as I don't think it's especially about AI. But the periods of my career when I've been actually writing code and solving complicated technical problems have been the most rewarding times in my life, and I'd frequently work on stuff outside work time just because I enjoyed it so much. But the other times when I was just maintaining other people's code, or working on really simple problems with cookie-cutter solutions, I get so demotivated that it's hard to even get started each day. 100%, I do this job for the challenges, not to just spend my days babysitting a fancy code generation tool.
So for me being able to have AI wrote certain things extremely fast with me just doing voice to text with my specific approach, is amazing.
I am all in on everything AI and have a discord server just for openclaw and specialized per repo assistants. It really feels like when I'm busy I can throw it an issue tracker number for things.
Then I will ssh via vs code or regular ssh which forwards my ssh key from 1password. My agents have read only repo access and I can push only when I ssh in. Super secure. Sorry for the tangent to the article but I have always loved coding now I love it even more.
In any case, I think we should start treating the majority of code as a commodity that will be thrown away sooner or later.
I wrote something about this here: https://chatbotkit.com/reflections/most-code-deserves-to-die - it was inspired by another conversation on HN.
It never was
> That is not an upgrade. That is a career identity crisis.
This is not X. It is Y.
> The trap is ...
> This gap matters ...
> This is not empowerment ...
> This is not a minor adjustment...
Your typical AI slop rhetorical phrasing.
Phrases like: "identity crisis", "burnout machine", "supervision paradox", "acceleration trap", "workload creep"
These sound analytical but are lightly defined. They function as named concepts without rigorous definition or empirical grounding.
There might be some good arguments in the article, but AI slop remains AI slop.
> AI is an in-context learner, not a standards enforcer.
> The AI is not judging your code. It is learning from it.
> Speed without structure is not speed. It is borrowed time.
> This is not about premature optimization or over-engineering. It is about giving the AI the patterns it needs to work effectively on your behalf.
> This is not a theoretical distinction. It is the single most important practical reality of working with AI coding tools in 2026.
Its not this, its that.
> But here is the part nobody wants to hear: the reverse is equally true.
> The result was transformative.
> Here is why.
If you want I can provide N=3 with the same AI pattern and phrases again.
But I have no issue with your argumentation whatsoever, it is just that I think there is more than sufficient evidence, and you think there is not.
THE MARKET WILL FILL THAT VOID
IT DOES NOT MAKE IT TRUE
it's all so fucking tiresome
Also, check out the dude's linkedin: https://www.linkedin.com/in/ivanturkovic/
Another little thing that resonated was a tweet that said "some will use it to learn everything and some so that they don't have to learn anything ". Of course it's not really a hard truth. It's questionable how much you can learn without really getting your hands dirty. But I do think people looking at it as a tool that helps then and/or makes them better will profit more than people looking to cut corners.