What's New in Mellea 0.4.0 + Granite Libraries Release

Published: (March 20, 2026 at 10:14 AM EDT)
3 min read

Source: Hugging Face Blog


⚠️ Collection Error: Content refinement error: Error: 429 429 Too Many Requests: you (bkperio) have reached your weekly usage limit, upgrade for higher limits: https://ollama.com/upgrade


Back to Articles

We have released Mellea 0.4.0 alongside three Granite Libraries: granitelib-rag-r1.0,granitelib-core-r1.0,granitelib-guardian-r1.0. Together, these releases make it easier to build structured, verifiable, and safety-aware AI workflows on top of IBM Granite models.

Mellea is an open-source Python library for writing generative programs — replacing probabilistic prompt behavior with structured, maintainable AI workflows. Unlike general-purpose orchestration frameworks, Mellea is designed to make LLM-based programs maintainable and predictable through constrained decoding, structured repair loops, and composable pipelines (New to Mellea? Start with our introductory blog and meet the team)

	Mellea 0.4.0

Mellea 0.4.0 is the latest release of an open-source research project initiated and developed by IBM Research. Building on 0.3.0 foundational libraries and workflow primitives, 0.4.0 expands the library’s integration surface and introduces new architectural patterns for structuring generative workflows.

What’s included:

  • Native integration with the Granite Libraries, offering a standardized API that relies on constrained decoding to guarantee schema correctness.

  • Instruct-validate-repair pattern via rejection sampling strategies

  • Observability hooks for event-driven callbacks to monitor and track workflows

See full list of Mellea 0.4.0 features and updates here

	What Are the Granite Libraries

Simply put, a Granite Library is a collection of specialized model adapters designed to perform well-defined operations on portions of an input chain or conversation. Rather than relying on general-purpose prompting, each specialized model is fine-tuned for a specific task such as query rewriting, hallucination detection, or policy compliance checking. The use of specialized adapters allows us to increase the accuracy of each task at a modest parameter count cost, and without disrupting the base model’s capabilities.

Three libraries are being released today for the granite-4.0-micro model, each targeting a distinct set of pipeline tasks and comprised of a collection of LoRA adapters:

  • Granitelib-core-r1.0: targeting the step of requirements validation in Mellea’s instruct-validate-repair loop

  • Granitelib-rag-r1.0: targeting a variety of tasks in agentic RAG pipelines covering pre-retrieval, post-retrieval, and post-generation

  • Granitelib-guardian-r1.0: specialized models for safety, factuality, and policy compliance

Mellea 0.4.0 is the latest release of an open-source research project initiated and developed by IBM Research. Building on the foundational libraries and workflow primitives introduced in 0.3.0, this release expands the integration surface and introduces new architectural patterns for structuring generative workflows.

	Getting Started



	


	Mellea
0 views
Back to Blog

Related posts

Read more »

OpenAI to acquire Astral

OpenAI Accelerates Codex growth to power the next generation of Python developer tools Today we’re announcing that OpenAI will acquire Astralhttps://astral.sh/,...

How to Measure AI Value

AI value the wrong way. Instead of asking “What new capabilities does this unlock?”, the conversation quickly turns into questions such as: How many hours can w...