arrow_back Back to Projects
Python FastAPI LangChain LLM RAG

Meridian Policy Intelligence

A secure, full-stack RAG application built to synthesize corporate policies with Llama 3.3 and LangChain.

Meridian Policy Intelligence screenshot

Meridian Policy Intelligence : RAG Application & Editorial UI

📌 Project Overview

Meridian Policy Intelligence is a full-stack Retrieval-Augmented Generation (RAG) application built to securely query, synthesize, and cite internal corporate policy documents. Bridging advanced AI orchestration with high-end editorial design, the project moves beyond generic “AI chat” interfaces to deliver a stunning, magazine-inspired reading experience.

Technologies Used: Python, FastAPI, LangChain (LCEL), ChromaDB, HuggingFace Embeddings, Groq (Llama 3.3), HTML/CSS/JS (Vanilla).


🚀 Key Features & Architecture

meridian

1. Robust AI Orchestration (LangChain & FastAPI)

Replaced traditional manual retrieval scripts with LangChain Expression Language (LCEL), creating a highly readable, declarative pipeline (retriever | prompt | llm | parser). The backend is powered by FastAPI, providing native asynchronous support, Pydantic data validation, and automatic OpenAPI documentation.

2. High-Fidelity RAG Pipeline

  • Ingestion: Parses complex markdown documents using LangChain’s MarkdownHeaderTextSplitter to preserve structural metadata.
  • Embeddings & Vector Store: Utilizes HuggingFaceEmbeddings running locally on CPU, securely storing document vectors in a persistent ChromaDB instance.
  • Generation: Leverages the Llama 3.3 model (via Groq) for rapid, intelligent response generation. A custom JSON Output Parser was engineered to ensure strict structured data responses, securely separating the synthesized answer from its source citations.

3. “Anti-Slop” Avant-Garde Frontend

Rather than settling for standard dark-mode “chat bubble” interfaces, the frontend was completely designed from the ground up as a High-End Brutalist/Editorial Light Mode experience:

  • Typography & Layout: Features high-contrast pairings (Bodoni Moda & Epilogue) and continuous manuscript-style reading.
  • Magazine Tropes: Intelligent responses utilize massive drop caps and article-style formatting instead of chat bubbles.
  • Atmosphere: Uses SVG noise textures and structural, brutalist input components to ensure a premium, non-generic aesthetic.

📊 Evaluation & Metrics

The pipeline was rigorously evaluated against a 25-question test suite measuring Groundedness, Citation Accuracy, and Latency. By optimizing chunk size (800 overlap 100) and increasing Top-K retrieval (7), the system achieved:

  • Citation Accuracy: 96.0%
  • Groundedness: 93.5%

đź’ˇ What I Learned

  • Prompt Hardening for RAG: Gained deep experience in preventing LLM hallucinations by writing strict conversational boundaries that force the model to cite exact retrieved metadata.
  • Declarative AI Pipelines: Mastered transitioning legacy Python logic into clean, piped LCEL structures.
  • Design as a Differentiator: Learned that AI tools do not have to look like AI tools. By applying classical graphic design and editorial aesthetics to LLM outputs, user trust and engagement are significantly increased.