

Nice! The idea that you can bias the model during training in such a way that it doesn’t focus on specific information is really fascinating.


Nice! The idea that you can bias the model during training in such a way that it doesn’t focus on specific information is really fascinating.


2FA is pretty much a must nowadays for anything you need to be even remotely secure. Maybe you can use a model to interrogate another model. 🤣


honestly that’s quite possible


I’m still kinda sad F# never caught on myself.


Ultimately, these things aren’t concrete plans, it’s just a conversation starter. The people who published it aren’t building anything, but it does provide a starting point for things to think of those of us who do build things. The parts I thought were meaningful were in the list at the end:
I think these are all good things to strive for.


Emacs is great though


The goals they state seemed perfectly reasonable to me. I don’t really see any contradiction with hyper-personalized computing and having thriving communities. I think it would be great if you could easily tailor your computer towards your workflow. It doesn’t mean that I’m not able to have shared interests with other people who have different flows.
In fact, I think the way modern applications are build is fundamentally wrong precisely because they couple the logic of the app with the UI. This is the reason we can’t compose apps the way we do command line utils. If apps were broken up into a service and frontend component by default, you’d be able to chain these services together to build highly customized workflows on top of that.
And that’s precisely the kind of thing AI tools are actually decent at doing. You can throw a bunch of API endpoints at it and have it build a UI using them that does what you want, or if it’s good enough you might not even need a UI, you can literally just type what you want and it’ll figure it out.


I think it’s very plausible that within a year or two we’ll have models that can run locally and do serious coding comparable to what Claude can do right now. There are two parallel tracks here too. One is models themselves improving, but the other is the tooling and the workflow. The thing I made for storing state outside the context is a good example of what’s possible and is still relatively unexplored. The less context the model needs to do its work the more viable small models become.


pretty much


Oh yeah, TDD works great with LLMs I find. They just have to make all the pieces fit, and since there’s a spec for what the code should be doing there’s no way to cheat. Either the tests pass or they don’t.


oh haha that one’s not me


Haha I don’t think I had a youtube link there. Yeah, using MCPs is handy, tools like crush can actually figure them out on their own. And thanks!


maybe find yourself a hobby other than opining on things you have no clue about dude


Let me know how it works out. I’ve been dogfooding it locally for a few days, but always good to hear reports from other people with these sorts of things. :)


The main approach right now is just doing reinforcement learning and minimizing operational context for the model. Another more expensive approach is using a quorum as seen here. You have several agents produce a solution and take the majority vote. While that sounds expensive, it might actually not be that much of a problem if you make smaller specialized models for particular domains. Another track people are pursuing is using neurosymbolic hybrid approach, where you could the LLM with a symbolic logic engine. This is my personal favorite route, and what I did with matryoshka is a variation on that. The LLM sits at the very top and it’s job is to analyze natural language user input, then form declarative queries to the logic engine, and evaluate the output. This way actual logic of solving the problem is handled in a deterministic way.


Thanks, honestly I’m really shocked it took people this long to realize that keeping state would be useful. The way MCP works by default is completely brain dead. And agree about RAGs, they’re good for biasing the model, but that’s about it. My thesis is that you don’t even need reinforcement learning, you can have a symbolic logic engine that the LLM drives instead. The job of the LLM is to parse noisy outside inputs like user queries, and to analyze the results to decide what to do next. The context of the LLM should focus on this, I have a task, and I make some declarative queries to the logic engine. The engine then takes over and does actual genuine reasoning to produce the result. The key failing of symbolic AI was the ontological problem where you had a combinatory explosion trying to create the ontologies for it to operate on. But with LLMs we can have it build the context for the logic engine to operate within on the fly. And you can even do code generation in this way!


At least would people try being creative or something with their critique. This is hexbear equivalent of libs sealioning into threads to tell you how China is authoritarian akchually on reddit!


Smaller models have been getting a lot more capable of late. A 32 billion param model can do quite a bit today, and there are a lot of things that can still be optimized going forward. If you look at capability improvements, it’s absolutely stunning, something you can run on a laptop today would’ve needed a whole data centre just a couple of years ago.
I actually like your way better, but thanks. :)