Plutonic Rainbows

Guardrail

I built Guardrail Gateway as an AI safety platform to make interactions with Large Language Models more secure and transparent. It adds a layer of content filtering, policy enforcement, and audit logging between applications and providers such as OpenAI. The system runs on a FastAPI backend with a React frontend, acting as an intelligent proxy that checks every request and response against a set of customizable safety policies before it reaches the model.

The core of the platform is a policy engine that uses regex-based rules with adjustable severity levels and actions like blocking, warning, or redacting content. Right now, I’ve implemented two main policy sets: one for detecting and redacting personally identifiable information, and another for identifying prompt injections or attempts to extract system prompts. Every event is logged for traceability and compliance.

Developers (including myself) can test and tune policies through a web interface, which includes tools for validating configurations, managing policies, and reviewing audit logs. The system uses SQLite for development and PostgreSQL for production, with JWT authentication for secure access and UUID support across databases. Typical requests — from scanning to response logging — complete in about two seconds, with most scans finishing in under 50 ms.

I designed Guardrail Gateway to run quietly in the background, using Python’s asyncio loop on a high port (58001) to minimize interference with other services. It’s written for Python 3.13 and built to scale horizontally thanks to its stateless API design. The frontend, built in React with TypeScript and Vite, includes full documentation for both developers and AI agents.

Search is here

I've finally added a search functionality to my blog, after many months of deliberating over styling and performance impact. After considering various options, I implemented a lightweight client-side search that lets readers quickly find posts by typing keywords into the search box now positioned in the header. The search looks through post titles and content excerpts, highlighting matching terms and displaying up to 10 results in a dropdown. It's nothing revolutionary, but it works well — searches execute in under 10 milliseconds once the index is loaded, and the whole implementation adds just 5KB of JavaScript and CSS to the initial page load.

Search uses a lazy-loading mechanism. Rather than forcing every visitor to download the 143KB search index (containing data for all 629 posts), the index only loads when someone actually clicks or tabs into the search box. This means most visitors who come to read a specific post aren't penalized with extra download time they'll never use. When someone does focus the search input, the index loads in the background while they're typing their query — if they've already entered text by the time it loads, the search runs automatically. It's a simple optimization, but it keeps the blog fast for everyone while still providing instant search for those who need it. The entire search feature added less than half a second to my build time, which felt like a reasonable trade-off for the functionality gained.

Sonnet 4.5

Anthropic release the new model. Pricing remains the same as Claude Sonnet 4, at $3/$15 per million tokens.

Google Gemini

Apologies for the oversight in my previous post — I should have included Google Gemini in the discussion. Gemini is a key player in the current AI landscape, offering a versatile suite of models that combine strong reasoning, coding, and multimodal capabilities. Leaving it out may have given the impression that it isn’t relevant alongside GPT, Claude, Grok, and Qwen, but in reality, it deserves recognition as one of the most significant entrants shaping the competitive field.

Four AI Heavyweights Shaping the Future

Artificial intelligence has become more than just a buzzword — it’s becoming a daily partner in how we work, create, and even play. From writing code to generating ideas and powering conversations, today’s leading models each bring their own personality and strengths. Let’s take a quick look at four of the most exciting names in the space right now: GPT-Codex, Claude, Grok, and Qwen.

GPT-Codex is the coder’s dream assistant. Developed by OpenAI, it bridges natural language and programming, making it possible to describe your goals in plain English and have them turned into functional code. Whether you’re debugging, migrating projects, or building prototypes, Codex feels like an extra teammate who never gets tired of problem-solving.

Claude, from Anthropic, stands out for its thoughtful and safe design. Instead of just pushing raw power, it focuses on clarity, alignment, and long-form reasoning. This makes it an excellent choice for complex projects, structured workflows, and conversations where nuance matters. With Claude Code, developers in particular are finding new ways to work faster while staying organized.

Grok and Qwen represent the new wave of AI challengers. Grok, from xAI, has built its identity around speed, wit, and humor, making interactions more engaging without sacrificing intelligence. Qwen, from Alibaba Cloud, is all about versatility, offering a wide range of model sizes that excel at multilingual tasks, coding, and even image editing. Both are proof that the AI landscape is getting broader and more dynamic every day.

As these models continue to evolve, the takeaway is clear: there’s no single best AI, only the best fit for your goals. Codex shines in coding, Claude thrives in thoughtful reasoning, Grok brings personality to problem-solving, and Qwen pushes the boundaries of scale and adaptability. Together, they highlight an exciting future where we can choose from a diverse toolkit of digital partners — each designed to help us think, create, and build in new ways.