Use of IA for PR - my POV

user1234 user1234 at 12.de
Tue Feb 10 20:16:06 UTC 2026


On Tuesday, 10 February 2026 at 17:54:29 UTC, Vladimir Panteleev 
wrote:
> On Tuesday, 10 February 2026 at 17:38:40 UTC, matheus wrote:
>> Interesting, since I'm not using AI I'd like to know, in this 
>> case you have LLM locally and you point to D source folder and 
>> It learns from that database and do everything from there?
>
> The main way I use LLMs is with Claude Code. Here's how it 
> works:
>
> 1. You open the directory with your project in a terminal
> 2. You run `claude`
> 3. This opens a TUI that looks like a chat interface. You type 
> your question or request for what you want the bot to do.
> 4. The bot looks at your code. If it's too big to fit into its 
> context (a limited window of how much it can see at a time), it 
> will search just for relevant bits.
> 5. If the problem is big, it will first write a plan for how it 
> aims to accomplish its goal, for you to read and approve.
> 6. It does the thing. It can edit files and run commands in 
> order to run your test suite (or at least check that the code 
> compiles). By default it will ask before every edit or command. 
> Many people run it in a sandbox and disable the prompts, so 
> that it can work by itself but still doesn't accidentally 
> delete your entire computer.
> 7. Sometimes the bot can automatically write down what it 
> learned in a memory file. It will read this file automatically 
> the next time you ask it to do something in that project. There 
> isn't really a lot of "learning" other than something like this.
>
> Before/aside that, I also have a spare GPU which I'm using to 
> run an autocomplete model. It's nice when writing code by hand. 
> For that I use https://github.com/CyberShadow/company-llama + 
> llama.cpp.
>
>> So I wonder if usually programming languages have restrictions 
>> to ensure bad code don't mess with anything, but on the other 
>> hand AI keep getting better and learns how to avoid bad code, 
>> what's the point of having all these languages? Or in fact, 
>> could AI write a better programming language by itself?
>
> Agentic coding actually works better the stricter the language! 
> This is because then the compiler can check if the code is 
> correct immediately, and if it's not, the agent sees the error 
> right away and can fix it before stopping. So, I think we will 
> see more strictly typed languages or languages with built-in 
> theorem proving become more popular. There are often too 
> frustrating or time consuming for humans to use for every day 
> programming, but it doesn't matter when the code is being 
> written by AI.
>
> I am seeing this too with testing and Nix. Writing integration 
> tests with Nix is usually a lot of work, but once it's written 
> then it's rock-solid proof that your thing works and that 
> everyone can verify that it works. You can even script entire 
> VMs that can run any software for integration tests, and these 
> VM tests run without any problems on any Linux machine 
> including GitHub Actions. So, I've since been adding Nix based 
> integration tests to all my projects (including this forum, 
> which now has Nix/Playwright based end-to-end tests).

I feel so old-fashioned when I read this. The worst is that I've 
been warned, pre-Covid time (so 2019), while chating, that this 
will the next big thing. You seem to have the method. I've heard 
recently that apparently another notable D user, "Feep", is 
heavily using AI agents.

I'm so done.


More information about the Digitalmars-d mailing list