For coding AI, it could make sense to specialize models on architecture, functional/array split from loopy solutions, or just asking 4 separate small models, and then using a judge model to pick the best parts of each.
I haven’t watched the video yet, but I have to say: running a personal LLM on my computer using products like gpt4all produces some really awesome results im very happy with.
I can totally envision a future where everyone can easily run their own local AI in the next ten years.
Yep, I’m actually getting pretty comparable results running devstral as an agent considered to a big model like Claude.
You have to specify the problem better, but the tradeoff is I can run it locally.
I’m still in the experiment phase of local models for agents, and they’re accomplishing tasks instead of just coding.
I could see a connection of these agents passing tasks around being highly effective.