LLMs
Implementation strategies for Large Language Models, focusing on practical business applications and AI integration solutions.
LLMs Posts
Token Guard: Keeping Your Agent Context Lean in CI
Token Guard is a GitHub Action that counts tokens in your agent context files and enforces limits in CI. Here’s why we check agent context into our repos, and why keeping it lean matters for team collaboration.
How to Get Guaranteed JSON from LLMs with Structured Outputs
Tired of parsing flaky JSON from LLM responses? OpenAI’s Structured Outputs feature guarantees your responses match your schema exactly. Here’s how to use it with Pydantic, when to choose it over function calling, and the gotchas you’ll encounter in production.
How to Choose an LLM When Every Model Claims State of the Art
Benchmark scores don’t tell you whether a model will work for your business. Here’s how to evaluate LLMs on the three axes that actually matter: quality, throughput, and cost.
Pydantic.ai: Building Smarter, Type-Safe AI Agents
The team that brought type safety to Python web development with Pydantic has just unveiled their take on AI development: Pydantic.ai. This new framework reimagines how we build AI applications by bringing Pydantic’s legendary validation capabilities to the world of Large Language Models.
Benchmarking AI: Evaluating Large Language Models (LLMs)
Large Language Models like GPT-4 are revolutionizing AI, but their power demands rigorous assessment. How do we ensure these marvels perform as intended? Welcome to the crucial world of LLM evaluation.
Other Categories
Tell us about your project and how Cuttlesoft can help. Schedule a consultation with one of our experts today.
Contact Us



