⚠️ This post links to an external website. ⚠️
Remember how I turned my old gaming laptop into a private LLM server with a Phoenix web interface? Maybe you remember how I was hindered by the fact that at the end of the day, my machine is a museum piece and barely runs models over 8GB?
Well, now I’m having the last laugh, because it turns out there’s a revolution occurring in the Large Language Model world, and it has the power to soup up the weak-sauce quantized (read: lobotomized) models that my aging system can barely run into true AGI (or at least, make them a bit better.)
As always, if you don’t care about the why and just want the how, I’ve put a link to the complete code down at the bottom of the article. If you are interested in the why…let’s talk about the Model Context Protocol (MCP).
continue reading on revelry.co
If this post was enjoyable or useful for you, please share it! If you have comments, questions, or feedback, you can email my personal email. To get new posts, subscribe use the RSS feed.