Software Engineers offer a necessary dose of reality, providing a grounded counterpoint to this pervasive enthusiasm. This cycle strongly echoes the dot-com speculative bubble, and its inevitable correction promises to deliver a severe global economic shock. This is precisely why AI must be treated as an integrating tool - a sophisticated component within the delivery mechanism - not as a monolithic, all-encompassing solution. When the AI house of cards collapses, those who mastered the tool will maintain stability, while those who relied solely on the illusion will be buried in the debris.
Taking a step back, taking it offline and identifying the actual problem gives a more balanced and level-headed approach to using AI. Running Large Language Models (LLMs) locally or in-house offers unparalleled flexibility. It's possible to switch models as needed, create new models, RAG and Context train and manipulate System User Prompts to achieve the best performance for each specific task – without being locked into a single model for everything or being forced to pay for premium features.
In my Home Lab, I can do all of this; from basic questions, discussions, queries, even running uncensored models, to context queries and RAG - I have access to it all by configuring models for our specific needs; like my chess playing AI - KnightShift - and my XP coding buddy, Coder AI.
Coder AI is grounded in qwen2.5-coder:7b and is my go-to coding assistant. I have refined the System User Prompt to ensure the answers given are succinct, code samples are constructive and when suggested solutions are repetitive or failing regularly, a new approach is considered without prompting. I can also provide supporting documents to RAG train the model or query in context. This level of control ensures (or drastically reduces) stupidity and hallucinations mainstream models like ChatGPT and CoPilot regularly present.
LLM's are pre-trained with no external internet access. Whatever we ask of the local LLMs, they share no data with OpenAI, AI Datacentres or dirty data brokers!
As a Software Engineer, I may have complicated questions, tasks or need another set of eyes on a problem. I have LLM's configured that I can use specifically for this.
There are a stack of models available for download for different purposes. Running LLM's locally with Ollama, I can hot-swap these to use a LLM best suited to the task in hand.
Ability to use RAG (Retrieval-Augmented Generation) on any personal files/datasets knowing the data source never leaves the network.
Local LLMs reduce dependency on external services and remove connectivity issues, ensuring greater reliability and true offline use.
Run multiple LLM models on a single query, see the responses side-by-side and and select the best response from the results.
This approach also prioritises my privacy. It keeps my prompts and other data out of large AI datacenters and away from ruthless and untrustworthy data brokers, keeping personal information secure and under my control.
And, with Open WebUI's slick interface modeled like the popular ChatGPT UI, I can seamlessly interact with LLMs in a way that feels familiar and natural, regardless of the model selected.
And for fun, I've taken a qwen model and with some manipulation, I've developed a Chess Playing AI called KnightShift. I play on a physical board and KnightShift plays virtually. The question is, have I beaten it?
With local LLMs, you can bypass the cloud entirely and unlock a more personalised, efficient, and cost-effective approach to AI development. By running models locally on your machine, you gain complete control over your data and workflows, ensuring that your projects are secure, scalable, and tailored to your specific needs.
No longer will you be tied to expensive subscription fees or limited by cloud infrastructure. With local LLMs, you can train, test, and deploy models at a fraction of the cost, while also enjoying faster processing speeds and more accurate results due to reduced latency. It's an empowering experience that will revolutionise your approach to AI-driven projects.
Whether you're a seasoned developer or just starting out with LLMs, local deployment offers a unique opportunity to take control of your data and workflows. Say goodbye to cloud-based limitations and hello to a more personalized, flexible, and cost-effective way of working with AI.
- Written by Llama3:8b in Open Web UI.
Use Open WebUI to upload personal or confidential documents and select your preferred LLM(s) to fine-tune their knowledge on these. Here's a summary of a famous fairy tale.
RAG allows you to ask any question about your provided dataset. They can be a vague or as complex as you wish. Use the power of the LLM to construct the answer you want.
Furthermore, the questions you ask with RAG do not have to be specific, the LLM is able to work out what you're asking and use the provided datasets to answer you.