I think affiliate links are the most fair/ethical advertisement can be. If i go on a random carpentry or painting blog, i'd rather have affiliate links to product they use rather than random google ads.
I only used Windows at work and for games in a VM, so take that with a grain of salt:
Older Windows bugs seemed fair: mostly edge cases, weird UI interaction, or stuff that only came out under heavy workload (also, windows file system).
This past few year, the bugs are incomprehensible. I understand non-professional versions are considered as Beta since Win10, but what it felt like is that Home version are actually alpha, and windows pro seems more and more like a beta.
NT4 had many serious BSODs. SP6 was so problematic due to a critical bug in LSA that it was re-released as SP6a.
Windows bugs have moved more and more into the 'edge case' territory. Not that major issues don't crop up for "everyone" today, but BSODs used to be much more common. Part of that was due to the architecture, thus drivers, but the other side of it was core Windows functionality that just had bugs.
Explorer.exe is still the shell -- the shell is defined at HKEY_LOCAL_MACHINE\SOFTWARE\Microsoft\Windows NT\CurrentVersion\Winlogon\Shell, if you want to look (or replace it).
> Just the jump from Sonnet 3.5 to 3.7 to 4.5, and Opus 4.5 has been pretty massive in terms of holistic reasoning, deep knowledge as well as better procedural and architectural adherence.
I don't really agree. Aside from how it handled frontend code, changes in Sonnet did not truly impact my overall productivity (from Sonnet 3.7 to 4 to 4.5, i did not try 3.5). Opus 4.5/Codex 5.2 are when the changes truly happenned for me (and i'm still a bit distrustfull of Codex 5.2, but i use it basically to help me during PRs).
That's fine. Maybe you're holding it wrong, or maybe your work is too esoteric/niche/complex for newer models to be bigger productivity boosters. Some of mine certainly is, I get that. But for other stuff, these newer models are incredible productivity boosters.
I also chat with these models for long hours about deep, complicated STEM subjects and am very impressed with the level of holistic knowledge and wisdom compared to models a year ago. And the abstract math story has gotten sooooo much better.
If we take out most of frontend work, and the easy backend/Ops tasks where writing the code/config is 99% of the work, i think my overall productivity with the latest gen (basically Opus 4.5) improve by 15-20%. I also am _very_ sure that with the previous generation (Sonnet 4, sonnet 4.5, Codex 5.1), my team overall velocity decreased, even taking into account the frontend and the "easy" tasks. The amount of production bug we had to deal with this year is crazy. To much code is generated, and me and the other senior on my team just can't carefully review everything, we have to trust sometime (especially data structures).
The worse part is reading a PR, and catching a reintroduced bug that was fixed a few commit ago. The first time i almost lost my cool at work and said a negative thing to a coworker.
This would be my advice to juniors (and i mean basically: devs who don't yet understand the underlying business/architecture): use the AI to explain how stuff work, generate basic functions maybe, but write code logic/algorithm yourself until you are sure you understand what you're doing and why. Work and reflect on the data structures by yourself, even if generated by the AI, and ask for alternatives. Always ask for alternatives, it helps understanding.
You might not see huge productivity gains from AI, but you will improve first, and then productivity will improve very fast, from your brain first, then from AI.
> The worse part is reading a PR, and catching a reintroduced bug that was fixed a few commit ago. The first time i almost lost my cool at work and said a negative thing to a coworker.
Losing your cool is never a good idea, but this is absolutely a time when you should give negative feedback to that coworker.
Feedback is what reviews are for; in this case, this aspect of the feedback should neither be positive nor neutral.
>> Kernighan's Law - Debugging is twice as hard as writing the code in the first place. Therefore, if you write the code as cleverly as possible, you are, by definition, not smart enough to debug it.
Now question is..
is AI providing solutions smarter than the developer using it might have produced?
And perhaps more importantly, How much time it takes AI to write code and human to debug it, even if both are producing equally smart solutions.
Just to add to your advice to juniors working with AI:
* Force the AI to write tests for everything. Ensure those tests function. Writing boring unit tests used to be arduous. Now the machine can do it for you. There's no excuse for a code regression making it's way into a PR because you actually ran the tests before you did the commit, right? Right? RIGHT?
* Force the AI to write documentation and properly comment code, then (this is the tricky part) you actually read what it said it was doing and ensure that this is what you wanted it to do before you commit.
Just doing these two things will vastly improve the quality and prevent most of the dumb regressions that are common with AI generated code. Even if you're too busy/lazy to read every line of code the AI outputs just ensuring that it passes the tests and that the comments/docs describe the behavior you asked for will get you 90% of the way there.
Sometimes the AI is all too good at writing tests.
I agree with the idea, I do it too, but you need to make sure the test don't just validate the incorrect behavior or that the code is not updated to pass the test in a way that actually "misses the point".
I've had this happen to me on one or two tests every time
Even more important, those tests need to be useful. Often unit tests are simply testing the code works as written which is generally doing more harm than good.
To give some further advice to juniors: if somebody is telling you writing unit tests is boring, they haven’t learned how to write good tests. There appears to be a large intersection between devs who think testing is a dull task and devs who see a self proclaimed speed up from AI. I don’t think this is a coincidence.
Writing useful tests is just as important as writing app code, and should be reviewed with equal scrutiny.
For some reason Gemini seems to be worse at it than Claude lately. Since mostly moving to 3 I've had it go back and change the tests rather than fixing the bug on what seems to be a regular basis. It's like it's gotten smart enough to "cheat" more. You really do still have to pay attention that the tests are valid.
Yep. It's incredibly annoying that obviously these AI companies are turning the "IQ knob" on these models up and down without warning or recourse. First OpenAI, then Anthropic and now Google. I'm guessing it's a cost optimization. OpenAI even said that part out loud.
Of course, for customers it is just one more reason you need to be looking at every AI outputs. Just because they did something perfect yesterday doesn't mean they won't totally screw up the exact same thing today. Or you could say it's one more advantage of local models: you control the knobs.
I had a colleague, senior software developer with masters degree in CS who said: why should I write tests if I can write a new feature to close sprint scope faster?
The irony is when company did lay off him due to covid the actual velocity of the team increased.
If I can reassure you, if your project is complex enough and involve heavy data manipulation, a 30% improvement using Opus/Gemini 3/codex 5.2 seems like a good result. I think on complex tasks, Opus 4.5 improves my output by around 20-25%.
And since it's way, way less wrong than sonnet4, it might also improve my whole team velocity.
I won't lie, AI coding has been a net negative for the 'lazy devs' on my team who don't delves into their own generated code (by 'lazy devs' here I mean the subset of devs who do the work but often don't bother to truly understand the logic behind what they used/did. They are very good coworkers, add velue and are not really lazy, but I don't see another term for that).
The DDG(X) was the destroyer the US navy wanted to build no? I thought it was a nice concept on what a modern destroyer should do m, what was your issue with it (and it's cancelled now? For sure?)
> The new Trump-class battleships will replace the Navy's previous plans to develop a new class of destroyer, the DDG(X). However, the sea service intends to incorporate the capabilities it had planned to employ on that platform into the new Trump-class ships.
Two month ago, i'd say 5%. With Opus4.5, Gemini 3 and gpt 5.2, it's now 20%, maybe 50% if we only talk about personal code output.
If we talk about my whole team output, I'd say the impact on code production is like 80-100%, but the impact in velocity is between 10% and -25%. So many bug in production, security holes, so many poor models definition making it to production DB only for me and the other true senior to have to fix it.
We are seniors, we read your PRs and try our best to do it thoroughly, but with AI multiplying your code output, and writing quite convincing solutions, it's way harder, so please: if an AI have written the code in your PR, verify the tests are not superficial, verify everything works, think for yourself about what the model is used for and if it can be improved before release. Re-verify the tests (especially if the AI had issues writing/passing them). And do it once more. Please (hopefully one of my coworkers will read this).
This person speak about "The Postmodern Condition" like they read it, but as they didn't, the criticism comes out as dead wrong and quite dumb.
Don't read this person. I'm not saying you have to read the whole book to use a quote, but talking this confidently while doing what we call in French a "contresens" is a sign that the author have no business talking about serious books or philosophy.
It's like Ayn Rand who believed Kant "critique of pure reason" is about criticism of the scientific method and now a huge part of US authors think the same while never having read Kant.
Didn't it happened once? Southern democrat great electors voted for the republican Vice president, because the democrat vice-president had a non-white wife, and this was forbidden under US law?
Virginia’s 23 Democratic electors (Southern) refused to vote for Democratic VP candidate Richard Mentor Johnson due to his open common-law relationship with Julia Chinn, an enslaved woman of mixed race (octoroon). Interracial marriage was illegal under anti-miscegenation laws.
They voted for Van Buren (president) but switched VP votes to William Smith (another Democrat), denying Johnson a majority. The Senate elected Johnson anyway.
I will remark that no one disputed OP when he remarked that the US executive power is also appointed, not elected, and that weirdly no one make the same point about how undemocratic it is. It does rs feel like OP is right about ideologues only being pedantic when it serves their points.
reply