Modern research agents excel at retrieving and summarizing information but falter when the truth is contested. My work rethinks how large language models are trained to reason, introducing IMMUNE-Bench, a framework that teaches AI systems to navigate misinformation through verification rather than trust.
IMMUNE-Bench builds synthetic, evolving worlds—complete with conflicting claims, temporal shifts, and probabilistically modeled biases. Within these worlds, agents learn to weigh evidence, cross-reference sources, and calibrate their confidence instead of parroting dominant narratives. The system’s three-tier architecture orchestrates document generation, summarization, and temporal evolution, enabling thousands of interconnected events to unfold coherently over simulated time.
A reinforcement learning framework drives the training process, rewarding factual accuracy, synthesis quality, and epistemic humility. Agents are incentivized to compare sources, recognize fabrication, and abstain when evidence is inconclusive. This emphasis on how conclusions are reached—rather than simply what is correct—cultivates genuine reasoning resilience.
By moving beyond clean, consensus-driven datasets, this research aims to create AI systems capable of critical inquiry—models that can reason under uncertainty, question their own assumptions, and remain grounded in verification even when truth itself becomes adversarial.
Goal: Engineered an end-to-end RAG system to enhance the factual accuracy of the Llama2-7b model by augmenting it with external, verified context.
Contributions: Designed and built a complete system, including an embedding database and an efficient retriever.
Technologies: Python, UAE-Large-V1, FAISS, prompt engineering.
Goal: Fine-tuned an open-source GPT-2 model on mutliples datasets to create a conversational assistant capable of answering specialized questions.
Contributions: Lead the data collection and the fine-tuning of the model.
Technologies: Python, Hugging Face, NLP.
Goal: Improved the efficiency and performance of NLP models for internal data classification.
Contributions:
Engineered a data preprocessing pipeline that reduced processing time by 99.3%, from 35 minutes to just 15 seconds.
Led a custom model fine tuning based on ModernBERT with open documents (CC0) from the Federal Aviation Administration (FAA).
Designed and implemented a training pipeline that improved NLP model performance by 15%.
Designed and implemented a Streamlit application for key users to provide feedback on the model's performance with explainability through SHAP values.
Technologies: Python, AWS, GitLab CI/CD.