Open to Opportunities

Vignesh Balamurugan M.B

|

AI/ML Engineer with hands-on experience in deep learning, NLP, LLM orchestration, RAG systems, and production-grade deployment. Currently pursuing B.Tech in Artificial Intelligence & Data Science at IIIT Sri City.

Building Intelligent Systems at Scale

Vignesh Balamurugan M.B

AI/ML Engineer with hands-on experience in deep learning, NLP, computer vision, LLM orchestration, and full-stack development. Proven track record through a research internship at NIT Trichy and production-grade open-source projects.

Passionate about building intelligent systems that solve real-world problems at scale — from multi-hop fact verification pipelines to multi-agent LLM orchestration frameworks.

LLM Engineering RAG Systems Agentic AI Multi-Agent Orchestration Deep Learning Computer Vision NLP MLOps
8+
Projects
1
Internship
2
Research Ongoing

B.Tech — Artificial Intelligence & Data Science

IIIT Sri City | 2023 – 2027 | CGPA: 8.8

CBSE XII — 94%

Edison G Agoram Memorial School | 2021 – 2023 | Head Boy AY 2022–23

CBSE X — 94.8%

Athena Global School | 2019 – 2021 | House Captain

Professional Experience

Research and industry experience in AI/ML systems.

Summer Research Intern

National Institute of Technology, Trichy
May 2025 – Jul 2025 · Trichy, Tamil Nadu
  • Worked on Human Activity Detection and Recognition using Vision Transformers, Computer Vision, and LSTMs.
  • Contributed to a research project on activity recognition currently under review for publication.
  • Assisted Prof. M Sridevi in captioning model creation from video input.
  • Delivered 18% accuracy improvement and minimized inference latency by 22% through optimized preprocessing and lightweight fine-tuning strategies.

Research & Ongoing Work

Human Activity Detection & Recognition

NIT Trichy — Vision Transformers, CV, LSTMs
2025 · Paper Draft Ready for Submission
  • Research paper on multi-modal human activity recognition using Vision Transformers for spatial feature extraction and LSTMs for temporal sequence modeling; paper draft completed and ready for submission.

STRIDE — Intelligent Transportation System

Python, OpenCV, Computer Vision
Research Ongoing
  • Developing a real-time traffic analysis system with dynamic ROI detection using a 2-tier parallel virtual lattice (8×8 grid) where each cell independently computes density, speed, and directional flow.
  • Targets dynamic lane region identification and vehicle direction prediction for adaptive traffic signal control.

Featured Projects

Production-grade AI/ML systems, research implementations, and full-stack applications.

STRIDE — Intelligent Transportation System

13-stage architecture leveraging a 2-tier parallel virtual lattice (8×8 ROI grid) for localized density, velocity, and directional traffic-flow estimation. Attained 0.70 IoU accuracy and 0.69 directional-flow accuracy.

OpenCVOptical FlowPythonDynamic Lattice

Crowd Stampede Risk Predictor

Dual-head CANNet architecture (ResNet-34 FPN backbone) for simultaneous crowd density estimation and person localization from aerial surveillance video. Temporal grid-based risk engine detecting rapid crowd accumulation and inward movement.

PyTorchResNet-34CANNetOpenCVDroneCrowd

Chennai Night Safety Intelligence

Integrated 12+ geospatial datasets (CCTV, police coverage, streetlights, crime heatmaps). Trained a spatial-context MLP with 9-channel neighborhood features and built a route-risk comparison engine yielding 10.5% peak risk reduction.

QGISgeopandasrasterioSpatial MLPOSRM

LinkSnip — URL Shortener & Product Intelligence

Full-stack Flask app with URL shortener (custom aliases, expiry, QR codes, click analytics) and a product intelligence engine scraping Amazon/Flipkart with anti-bot evasion and NLP sentiment analysis. Auto-categorization across 10 product types.

FlaskSQLiteBeautifulSoupChart.jsNLP

RL for CRISPR Gene-Editing Optimization

Modeled CRISPR gene-editing as a 2D grid MDP and solved using Value Iteration, Policy Iteration, and tabular SARSA with ε-greedy exploration. Extended with a Dyna-style drift-aware hybrid agent for faster post-drift recovery.

SARSAPolicy IterationMDPsPythonNumPy

Timetable Fragmentation & Student Fatigue Analysis

Cross-sectional survey data collection with regex-based extraction, median imputation, IQR outlier capping, Min-Max scaling. EDA using Pearson correlation heatmaps, regression plots, histogram distributions, and scatter plots.

PandasSeabornMatplotlibPython

Technical Expertise

Technologies and tools I use to build intelligent systems at scale.

🧠 AI & Machine Learning

PyTorchTensorFlowScikit-LearnOpenCVHuggingFace TransformersFAISSDeep LearningNLPComputer VisionReinforcement LearningVision Transformers

🤖 LLM & Agentic AI

LangChainLangGraphRAG SystemsLLM OrchestrationMulti-Agent PipelinesPrompt EngineeringLoRA Fine-Tuning

☁️ Cloud & MLOps

DockerKubernetesGitHub ActionsMLflowDVCAWSPrometheusGrafana

⚙️ Backend Development

FastAPIFlaskNode.jsREST APIsSQLiteRedisChromaDB

📊 Data & GIS

PandasNumPyQGISgeopandasrasterioSQLSeabornMatplotlibPlotly

💻 Languages & Tools

PythonJavaCSQLJavaScriptGitLinuxPostmanJupyter

My Resume

View or download my latest resume below.

Certifications & Achievements

📜

Supervised ML: Regression & Classification

Coursera — DeepLearning.AI · 99.83%

Verify Certificate
📜

Advanced Learning Algorithms

Stanford / DeepLearning.AI · 100%

Verify Certificate
📊

Data Analytics Job Simulation

Deloitte via Forage · January 2026

96.5%ile
JEE Mains
Head Boy
AY 2022–23
94%
CBSE XII
GGJ 2024
Unity Game Shipped

Research Insights

Deep dives into my ongoing research and technical explorations.

April 2026 Computer Vision ITS

STRIDE: Training-Free Lane Detection from Fixed-Camera Surveillance

How do you detect lanes on roads that have no painted markings, from overhead surveillance cameras, without any training data? Our answer: let the vehicles draw the lanes themselves.

The Problem

Lane detection is a fundamental requirement for intelligent transportation systems (ITS). Yet existing solutions demand either expensive per-viewpoint deep learning models or depend on fragile trajectory-level analysis. Most lane detection research focuses on forward-facing ego-vehicle cameras with visible lane markings — but the real need is in oblique overhead surveillance cameras viewing roads that may lack any painted markings.

Our Approach: STRIDE

STRIDE (Spatio-Temporal Rule-based Identification and Direction Extraction) is a 13-stage rule-based pipeline that transforms raw vehicle detections into labeled lane regions with calibrated direction estimates — requiring zero training data. The key insight: in fixed-camera settings, vehicles themselves trace out lane boundaries through repeated traversal patterns.

Key Technical Innovations

  • Confidence-Weighted Direction Accumulation: Instead of naive averaging, each vehicle detection contributes directional evidence weighted by detection confidence, building robust directional histograms over time.
  • Directional Coherence Map: We use the Rayleigh statistic to quantify how consistently vehicles move in each spatial region, filtering noise from parking or U-turning vehicles.
  • Edge-Preserving Guided-Filter Smoothing: Traditional Gaussian smoothing bleeds lane boundaries together. Our guided-filter approach preserves sharp density transitions between adjacent lanes.
  • Von Mises Mixture Model (VMM): For direction-based lane separation, we employ VMM clustering — the circular-statistics equivalent of Gaussian mixture models — yielding statistically principled angular clustering.

Results

Evaluated on 47 video sequences from the UA-DETRAC dataset, with manually created lane-level ground truth (the first published dataset of its kind). STRIDE achieves:

  • Mean Region IoU: 0.599 — outperforming classical baselines
  • Mean Direction Error: 6.00 degrees
  • Deployment Consistency: Top-3 finish in 83% of scenes
  • Statistically significant improvement (p = 7.2 x 10^-4)

Why It Matters

STRIDE demonstrates that principled rule-based systems can deliver competitive performance for infrastructure-side perception tasks without any training data. This is particularly valuable for rapid deployment across thousands of cameras in smart city networks, where per-camera fine-tuning is impractical.

Get in Touch

Interested in collaborating on AI/ML projects or discussing opportunities? Let's connect.