Knowledge Hub
Research papers, technical deep-dives, and industry perspectives.
Multi-Modal Deepfake Detection System
Mani Chourasiya, Chetan Khapedia, Diksha Bharawa
IJSRNSC Vol. 12(4), Aug 2024
A 7-model heterogeneous ensemble — one PyTorch audio-visual transformer (Pinpoint: ResNet18 + MFCC Conv1D/GRU + 8-head gated cross-attention) and six TensorFlow CNNs (EfficientNet-B4, ResNet-50 ×2, VGG-16 ×2, InceptionV3) — unified behind a single weighted-voting loader with named ensemble groups (`default`, `fast`, `single`, `maximum_accuracy`, `visual_only`). The contribution is reaching 72.3% accuracy on FaceForensics++ c23 with reviewer-facing per-frame heatmaps and mel spectrograms, packaged as a 3-container Docker stack with a one-click AWS CloudFormation deploy.
Scalable ML Pipeline for Precision Agriculture
Chetan Khapedia
An end-to-end architecture for fleets of agricultural edge devices: multi-modal sensor ingestion, sub-100 ms on-device inference on Coral TPUs, and a Kubeflow-orchestrated retraining loop with drift-aware model promotion. The focus is the system contract between edge and cloud, not just the model.
A Practical Guide to Quantization for Edge AI
Jun 2025
What actually changes when you go from FP16 to INT8: accuracy, latency, memory, and the failure modes nobody warns you about. Side-by-side numbers from Jetson Nano and Coral TPU, post-training vs. quantization-aware training, and the deployment checklist that separates a working edge model from a broken one.
Building a Self-Hosted Production ML Platform
Aug 2025
Architecture notes — not a tutorial. How experiment tracking, the model registry, the inference layer, and observability fit together on K3s; where the seams are; and which tradeoffs are worth making when one engineer has to operate the whole stack. GitOps promotion, canary rollout, and the SLOs that make it boring.
Real-Time Control with ROS 2 on Constrained Hardware
Nov 2025
Where the latency actually comes from in a ROS 2 control loop — DDS, executor model, scheduling, and IPC — and what it takes to hold sub-1 ms jitter on a Raspberry Pi 5 with a PREEMPT_RT kernel. Measurements, not promises: cyclictest baselines, executor tuning, and a reproducible setup for deterministic robotic control.
Generating Reliable Quiz Questions with a Local LLM
Feb 2026
What it actually takes to make a 7B local model produce exactly N questions, in the right mix, in valid JSON, every time. A custom Ollama Modelfile (`quiz-master`) on top of qwen2.5-coder:7b, a length-aware token budget (~400 tokens per question), a schema-locked prompt with type-mix balancing, and a multi-provider fallback chain (Ollama → Gemini → OpenAI) so the platform stays online when the M4 host is offline. Plus the failure modes — silent truncation, type collapse, off-by-one counts — that any production prompt pipeline has to handle.