Getting work done with AI
I like to give it a go to the Advent of Code programming contest problems. They are usually funny and entertaining, sometimes too entertaining. But this year, I was busy with other matters, and I could not see much of it, so it dawned on me that perhaps I could use that as an excuse to give it a go to some of the newest LLMs that some people were raving about.
Coping and pasting the text of an AoC problem did not require much attention, so armed with LMStudio software and a couple of Qwen2.5 LLMs, I started testing. The 8B model dashes (at around 24 tokens/sec on my PC equipped with an Nvidia RTX 4060 Ti with 16 GB of VRAM). Unfortunately, the output is usually not good enough.
However, when using the 32B LLM, the results, though slow at around 2.5 tokens/sec, are usually correct and produce Python source code that runs perfectly and delivers the proper solution, at least for the first question of each problem, in just a few minutes. Even with that slow generation rate, the output is created much faster than I would be able to write the code, even if I knew what the correct answer looked like from the beginning.
Indeed, the model could not produce a working solution for some of the problems, though a bit of coaching from my side helped it succeed even in these cases. For example, one problem was taking forever to calculate. Still, when I hinted that the solution code should include memoization to prevent repeated calculations, the model created another version that crunched the right solution within seconds.
Comments