Skip to content
View Y-R-A-V-R-5's full-sized avatar

Block or report Y-R-A-V-R-5

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Y-R-A-V-R-5/README.md

👋 Hi, I'm Adithya Vardhan Reddy

ML Engineer | Computer Vision | Multimodal Retrieval | Architecture-Level R&D


🚀 Overview

ML Engineer focused on designing efficient, interpretable ML systems under real-world constraints.

  • Build CPU-efficient pipelines with measurable latency–accuracy trade-offs
  • Study model behavior (failure modes, drift, sensitivity) beyond benchmark scores
  • Develop architecture-level modifications (not just training optimizations)
  • Create modular, reproducible experiment systems for controlled evaluation
  • Translate research ideas into deployable, constraint-aware systems

⚙️ Core Work Domains

YOLO Architecture Engineering

  • Cross-version design (v8 → v11 hybrids) with backbone–head decoupling
  • Custom modules integrated via shape-consistent YAML pipelines
  • Focus on tiny-object sensitivity & feature preservation
  • Profiling FLOPs, latency, and structural efficiency (CPU-bound)

Multi-Object Tracking (MOT)

  • Implemented SORT / DeepSORT / OCSORT under unified pipelines
  • Analyze ID-switches, track continuity, and failure cases
  • Develop proxy metrics beyond MOTA for real-world reliability
  • Visual diagnostics: trajectory behavior & temporal consistency

Multimodal Retrieval

  • CLIP-based embedding systems (image–text alignment)
  • ANN search using NNDescent / FAISS
  • UMAP clustering for structure discovery
  • Metadata-aware filtered semantic retrieval

Experimental ML Systems

  • Fully reproducible pipelines (config-driven experiments)
  • Structured ablations for architecture validation
  • Controlled studies on noise, resolution, and scaling effects
  • CPU-first debugging for true bottleneck identification

🧩 Featured Projects

VisionTracker

  • YOLOv11 + MOT benchmarking system
  • Unified evaluation across SORT / DeepSORT / OCSORT
  • Metrics: ID-switches, fragmentation, track lifetime
  • Includes trajectory visualization & sequence diagnostics
  • Built for real-world tracker comparison

🔗 https://github.com/Y-R-A-V-R-5/VisionTracker


YOLO-Tweaks

  • Cross-generation YOLO architecture experimentation
  • Backbone–head hybridization via modular configs
  • Focus: tiny-object detection & efficiency trade-offs
  • Benchmarks FLOPs vs CPU latency vs accuracy
  • Rapid testing of custom structural variants

🔗 https://github.com/Y-R-A-V-R-5/YOLO-Tweaks


🏗️ Industrial & Private Work

SerpensGate-YOLOv8

  • Custom backbone for plant disease detection
  • Modules: SerpensELAN, DySnake, CBAM
  • Optimized CPU-only inference pipeline
  • Tuned for domain-specific feature patterns
  • Evaluated under lighting & texture variability

Multimodal Retrieval System

  • CLIP (ViT-B/32) embedding pipeline
  • ANN indexing via NNDescent
  • UMAP-based clustering & visualization
  • Metadata-conditioned semantic search
  • Designed for low-latency large-scale retrieval

🛠 Technical Stack

ML & Vision: PyTorch, YOLO (v8–v11), OpenCV
Retrieval: CLIP, FAISS, NNDescent
Engineering: Python, Git, modular ML systems
Analysis: Matplotlib, Seaborn


🎯 Research Directions

  • Long-context vision architectures
  • Efficient MOT without heavy re-identification
  • Multimodal fusion beyond embedding similarity
  • CPU-first architectural optimization
  • Model sensitivity to noise, scale, and distribution shift

📈 Research Approach

  • Ablation-first design to isolate architectural impact
  • Track stability, variance, and failure patterns
  • Ensure reproducibility across seeds & configs
  • Prefer controlled experiments over large opaque runs
  • Focus on understanding model behavior, not just outputs

🌐 Connect

GitHub: https://github.com/Y-R-A-V-R-5
LinkedIn: https://www.linkedin.com/in/yravr/
Email: adithyavardhanreddy2003@gmail.com

Pinned Loading

  1. FragileML FragileML Public

    Examines model fragility across capacity, fidelity, stability, representation, and temporal axes. Isolated and multi-axis experiments reveal sensitivity to noise, drift, and randomness. Metrics pri…

    Jupyter Notebook

  2. layerscale layerscale Public

    Studies width & depth scaling to identify diminishing returns, hardware-aware sweet spots, and practical capacity limits. Measures marginal utility of larger models on real hardware, emphasizing em…

    Jupyter Notebook

  3. hybrids hybrids Public

    Explores cross-generation architecture splicing to uncover latent YOLO assumptions. Mini-projects include: C3K2↔CIB↔SCDown block compatibility, topology propagation sanity checks, CPU cost vs small…

    Jupyter Notebook 2

  4. resomap resomap Public

    Maps resolution sensitivity to detect when higher visual detail actually improves performance. Evaluates trade-offs between AP_small vs AP_medium, memory, latency, and interactions with capacity. F…

    Python 1

  5. DownScaleXR DownScaleXR Public

    Architecture-level study of how early CNN downsampling choices affect bias, generalization, and CPU inference behavior under constrained settings.

    Jupyter Notebook 2

  6. VisionTracker VisionTracker Public

    VisionTracker benchmarks multi-object tracking using YOLOv11l with SORT, DeepSORT, and OCSORT. It provides heuristic metrics like track lifetime, ID switches, bounding box volatility, and trajector…

    Jupyter Notebook 2