An AI ‘Nerd Knob’ Each Community Engineer Ought to Know

on

|

views

and

comments


Alright, my buddies, I’m again with one other publish based mostly on my learnings and exploration of AI and the way it’ll match into our work as community engineers. In as we speak’s publish, I wish to share the primary (of what is going to seemingly be many) “nerd knobs” that I feel all of us ought to pay attention to and the way they are going to impression our use of AI and AI instruments. I can already sense the thrill within the room. In any case, there’s not a lot a community engineer likes greater than tweaking a nerd knob within the community to fine-tune efficiency. And that’s precisely what we’ll be doing right here. Wonderful-tuning our AI instruments to assist us be simpler.

First up, the requisite disclaimer or two.

  1. There are SO MANY nerd knobs in AI. (Shocker, I do know.) So, if you happen to all like this sort of weblog publish, I’d be completely happy to return in different posts the place we take a look at different “knobs” and settings in AI and the way they work. Nicely, I’d be completely happy to return as soon as I perceive them, at the very least. 🙂
  2. Altering any of the settings in your AI instruments can have dramatic results on outcomes. This consists of growing the useful resource consumption of the AI mannequin, in addition to growing hallucinations and reducing the accuracy of the knowledge that comes again out of your prompts. Contemplate yourselves warned. As with all issues AI, go forth and discover and experiment. However accomplish that in a protected, lab surroundings.

For as we speak’s experiment, I’m as soon as once more utilizing LMStudio operating regionally on my laptop computer slightly than a public or cloud-hosted AI mannequin. For extra particulars on why I like LMStudio, take a look at my final weblog, Making a NetAI Playground for Agentic AI Experimentation.

Sufficient of the setup, let’s get into it!

The impression of working reminiscence dimension, a.ok.a. “context”

Let me set a scene for you.

You’re in the course of troubleshooting a community problem. Somebody reported, or observed, instability at some extent in your community, and also you’ve been assigned the joyful activity of attending to the underside of it. You captured some logs and related debug data, and the time has come to undergo all of it to determine what it means. However you’ve additionally been utilizing AI instruments to be extra productive, 10x your work, impress your boss, you understand all of the issues which are happening proper now.

So, you determine to see if AI can assist you’re employed by the information sooner and get to the foundation of the difficulty.

You fireplace up your native AI assistant. (Sure, native—as a result of who is aware of what’s within the debug messages? Greatest to maintain all of it protected in your laptop computer.)

You inform it what you’re as much as, and paste within the log messages.

Asking an AI assistant to help debug a network issue.
Asking AI to help with troubleshooting

After getting 120 or so strains of logs into the chat, you hit enter, kick up your ft, attain on your Arnold Palmer for a refreshing drink, and anticipate the AI magic to occur. However earlier than you possibly can take a sip of that iced tea and lemonade goodness, you see this has instantly popped up on the display screen:

AI Failure! Context length issue
AI Failure! “The AI has nothing to say”

Oh my.

“The AI has nothing to say.”!?! How may that be?

Did you discover a query so troublesome that AI can’t deal with it?

No, that’s not the issue. Take a look at the useful error message that LMStudio has kicked again:

“Attempting to maintain the primary 4994 tokens when context the overflows. Nevertheless, the mannequin is loaded with context size of solely 4096 tokens, which isn’t sufficient. Attempt to load the mannequin with a bigger context size, or present shorter enter.”

And we’ve gotten to the foundation of this completely scripted storyline and demonstration. Each AI software on the market has a restrict to how a lot “working reminiscence” it has. The technical time period for this working reminiscence is “context size.” Should you attempt to ship extra information to an AI software than can match into the context size, you’ll hit this error, or one thing prefer it.

The error message signifies that the mannequin was “loaded with context size of solely 4096 tokens.” What’s a “token,” you marvel? Answering that could possibly be a subject of a completely completely different weblog publish, however for now, simply know that “tokens” are the unit of dimension for the context size. And the very first thing that’s executed whenever you ship a immediate to an AI software is that the immediate is transformed into “tokens”.

So what can we do? Nicely, the message provides us two potential choices: we will enhance the context size of the mannequin, or we will present shorter enter. Typically it isn’t an enormous deal to offer shorter enter. However different occasions, like after we are coping with massive log recordsdata, that choice isn’t sensible—the entire information is necessary.

Time to show the knob!

It’s that first choice, to load the mannequin with a bigger context size, that’s our nerd knob. Let’s flip it.

From inside LMStudio, head over to “My Fashions” and click on to open up the configuration settings interface for the mannequin.

Accessing Model Settings
Accessing Mannequin Settings

You’ll get an opportunity to view all of the knobs that AI fashions have. And as I discussed, there are numerous them.

Default configuration settings
Default configuration settings

However the one we care about proper now’s the Context Size. We are able to see that the default size for this mannequin is 4096 tokens. However it helps as much as 8192 tokens. Let’s max it out!

Maxing out the Context Length
Maxing out the Context Size

LMStudio offers a useful warning and possible motive for why the mannequin doesn’t default to the max. The context size takes reminiscence and assets. And elevating it to “a excessive worth” can impression efficiency and utilization. So if this mannequin had a max size of 40,960 tokens (the Qwen3 mannequin I take advantage of typically has that prime of a max), you may not wish to simply max it out immediately. As an alternative, enhance it by a little bit at a time to search out the candy spot: a context size sufficiently big for the job, however not outsized.

As community engineers, we’re used to fine-tuning knobs for timers, body sizes, and so many different issues. That is proper up our alley!

When you’ve up to date your context size, you’ll must “Eject” and “Reload” the mannequin for the setting to take impact. However as soon as that’s executed, it’s time to reap the benefits of the change we’ve made!

The extra context length allows the AI to analyze the data
AI absolutely analyzes the logs

And take a look at that, with the bigger context window, the AI assistant was capable of undergo the logs and provides us a pleasant write-up about what they present.

I significantly just like the shade it threw my manner: “…think about in search of help from … a professional community engineer.” Nicely performed, AI. Nicely performed.

However bruised ego apart, we will proceed the AI assisted troubleshooting with one thing like this.

AI helps put a timeline of the problem together
The AI Assistant places a timeline collectively

And we’re off to the races. We’ve been capable of leverage our AI assistant to:

  1. Course of a big quantity of log and debug information to establish potential points
  2. Develop a timeline of the issue (that can be tremendous helpful within the assist desk ticket and root trigger evaluation paperwork)
  3. Determine some subsequent steps we will do in our troubleshooting efforts.

All tales should finish…

And so you have got it, our first AI Nerd Knob—Context Size. Let’s overview what we realized:

  1. AI fashions have a “working reminiscence” that’s known as “context size.”
  2. Context Size is measured in “tokens.”
  3. Oftentimes occasions an AI mannequin will help a better context size than the default setting.
  4. Rising the context size would require extra assets, so make modifications slowly, don’t simply max it out utterly.

Now, relying on what AI software you’re utilizing, it’s possible you’ll NOT be capable of regulate the context size. Should you’re utilizing a public AI like ChatGPT, Gemini, or Claude, the context size will rely upon the subscription and fashions you have got entry to. Nevertheless, there most positively IS a context size that may issue into how a lot “working reminiscence” the AI software has. And being conscious of that truth, and its impression on how you should use AI, is necessary. Even when the knob in query is behind a lock and key. 🙂

Should you loved this look underneath the hood of AI and want to find out about extra choices, please let me know within the feedback: Do you have got a favourite “knob” you want to show? Share it with all of us. Till subsequent time!

PS… Should you’d wish to be taught extra about utilizing LMStudio, my buddy Jason Belk put a free tutorial collectively known as Run Your Personal LLM Domestically For Free and with Ease that may get you began in a short time. Test it out!

 

Join Cisco U. | Be a part of the  Cisco Studying Community as we speak without cost.

Be taught with Cisco

X | Threads | Fb | LinkedIn | Instagram | YouTube

Use  #CiscoU and #CiscoCert to hitch the dialog.

Learn subsequent:

Making a NetAI Playground for Agentic AI Experimentation

Take an AI Break and Let the Agent Heal the Community

Share:



Share this
Tags

Must-read

15 Greatest Seitan Recipes – Sharon Palmer, The Plant Powered Dietitian

Discover 15 superb seitan recipes which might be plant-based, vegan, and protein-packed. These wholesome meals from prime meals bloggers are good for meatless...

New Bereavement Value Calculator reveals that grief is pricey – The Well being Care Weblog

by EMMA PAYNE Grief is pricey. Along with the numerous human impacts, analysis reveals that bereavement results in a 20%-30% improve in well being...

Recent articles

More like this

LEAVE A REPLY

Please enter your comment!
Please enter your name here