Growth Cyber Newsletter

Open Source LLM Guardrails

Edition: June 26, 2025

Read time: 5 minutes

There’s lots of places you can get LLM Guardrails.

In my op-ed “Are LLM Guardrails a Commodity?” that was recently published in the Summer issue of AI Cyber Magazine I mentioned that LLM Guardrails can come from eval products and AI Security Runtime products.

They can also come from open source. AI Engineer Skylar Payne showed us how to use Haize Labs’ j1-micro reward model to create custom guardrails. We downloaded the model from hugging face and ran it locally with Ollama.

With the model running locally, we were able to use it within a python script to create guardrails. In a function we tell the model to check for PII in the message.

check_pii

We can then use this to check for PII in the input and output of the AI application.

check_io

j1-micro is a powerful and tiny model that be used to roll your own guardrails if you’re up for self hosting the model. Benefits of a smaller self hosted model are reduced latency, data privacy, and potential cost savings as well. The tradeoff of course is it’s more technical work on your end to set up.

If you’d like to learn more about this make sure to follow Skylar, Leonard (CEO of Haize Labs), and myself on X.

When you're ready here's 3 ways I can help you