Hi, I'm Anurag Mishra

Graduate Student and Researcher working at the intersection of explainable AI, ethics, and deep learning.

I'm a graduate student and researcher at RIT, focusing on making AI systems more transparent and ethical. My work spans deepfake detection, bias mitigation in language models, and neural machine translation.

Currently, I'm exploring the intersection of explainable AI and adversarial robustness, developing techniques to make AI systems both interpretable and resilient.

Research Focus

→ Explainable AI (XAI) and Mechanistic Interpretability

Advancing research in explainability through mechanistic interpretability techniques, uncovering how models process and represent information. Developing frameworks to enhance the transparency and trustworthiness of AI systems across diverse applications.

→ Detecting and Debiasing Large Language Models

Investigating biases in LLMs by analyzing their training data and architectures. Designing robust debiasing methodologies to ensure fair and ethical AI, with a focus on mitigating adverse impacts in real-world deployments.

→ CUDA Kernel Development for High-Performance AI

Harnessing the power of CUDA for writing optimized kernels to accelerate computations in deep learning workflows. Exploring applications of CUDA in scaling and enhancing the efficiency of large language models for high-performance AI engineering.

→ Deepfake Detection and Explainable AI

Building multi-modal deepfake detection systems using innovative approaches in neuro-symbolic AI and explainable AI. Enhancing detection accuracy and transparency with layer-wise relevance propagation to tackle emerging challenges in multimedia authenticity.

Current Research Work

DeFake Project

Machine Learning Research Assistant • Oct 2024 - Present

Leading research with Dr. Matthew Wright on explainable AI for deepfake detection. Developing novel approaches to make detection systems transparent and interpretable through multi-modal analysis and saliency mapping.

LLM Security Research

Research Assistant @ RIT • Oct 2024 - Present

Collaborating with Dr. Bartosz Krawczyk to study adversarial vulnerabilities in large language models. Developing frameworks for bias detection and mitigation in AI-generated content.