Skills

  • Robotics & Control Systems

    Control Techniques

    Model Predictive Control (MPC), PID, LQR, State-Space Control, Kalman & Extended Kalman Filters

    .

    Navigation & Planning

    Autonomous Navigation, Lane Changing Algorithms, Obstacle Avoidance

    .

    Hardware Integration

    Pixhawk Flight Controller, Arduino (UNO, Duemilanove), Sensor Fusion Modules

    .

    Trajectory & Motion Optimization

    Trajectory Planning, Feedback Linearization, Optimal Control

    .

    Kinematics & Dynamics

    Forward/Inverse Kinematics, Robot Dynamics, Lagrangian Modeling

    .

    Simulation & Design

    MATLAB & Simulink-based modeling and tuning of dynamic systems

  • Computer Vision & Deep Learning

    Object Detection & Tracking

    YOLOv7, YOLOv8, Real-time vision systems

    .

    Depth & Pose Estimation

    MiDaS for monocular depth, Mediapipe for pose/gesture/lip tracking

    .

    3D Object Recognition

    PointNet for point cloud classification and segmentation

    .

    Sequence Modeling

    LipNet for video-based lip reading and temporal feature extraction

    .

    Frameworks

    TensorFlow, Keras, OpenCV, Hugging Face Transformers

    .

    Edge Deployment

    Optimized models for Jetson Nano and Jetson Xavier platforms

  • 3D Perception & Spatial Computing

    Reconstruction & Mapping

    2D-to-3D Conversion, Mesh Reconstruction, Open3D

    .

    Point Cloud Processing

    Generation, Filtering, and Visualization of LiDAR/Depth data

    .

    Spatial Analytics

    Real-Time 3D Mapping, Face Mesh, and Pose Estimation using vision systems

  • Reinforcement Learning

    Core Algorithms

    Q-Learning, Markov Decision Processes (MDPs), Policy Gradient Methods

    .

    Environment Modeling

    State-Action Mapping, Reward Structuring, and Optimization

  • Embedded Systems & Edge Computing

    Microcontroller Programming

    Arduino (UNO, Duemilanove), Real-time sensor interfacing

    .

    NVIDIA Jetson Ecosystem

    Real-time inference using CUDA on Jetson Nano & Xavier AGX

    .

    ROS 2 & Omniverse

    Robotics middleware, Omniverse extension development

  • Signal & Neural Data Processing

    Time-Frequency Analysis

    Wavelet Transforms (Morlet), Time-Frequency Decomposition, FFT, Hilbert Transform

    .

    Neural Signal Features

    Inter-Site Phase Clustering (ISPC), Phase Lag Index (PLI)

    .

    EEG Simulation & Analysis

    Signal Modeling, Bandpass Filtering, Artifact Removal

    .

    Frequency-Domain Filtering

    High/Low-Pass Filtering, Spectral Analysis

  • Audio & Speech Processing

    Feature Extraction

    MFCCs, Spectrograms, Temporal-Spatial Feature Learning

    .

    Classification Tasks

    UrbanSound8K Audio Classification, Spoken Word Recognition

    .

    Video-Audio Fusion

    Lip Reading via CNN-RNN hybrid models on video sequences

  • Software Development

    Programming Languages

    Python, C++, MATLAB, Verilog HDL

    .

    Web & UI Development

    React.js for frontend, TensorFlow.js for browser-based inference

    .

    Automation & Scripting

    PyAutoGUI for GUI control, Python automation pipelines

    .

    Version Control

    Git for versioning, GitHub for collaboration and CI/CD

  • Tools, IDEs & Deployment Platforms

    IDEs & Dev Environments

    Jupyter Lab, VS Code, PyCharm, MATLAB, Google Colab

    .

    Version Control & Collaboration

    Git, GitHub, GitHub Actions, Markdown Documentation

    .

    Containerization & Deployment

    Docker, TensorFlow Lite, ONNX Runtime, Gradio Interface Hosting

    .

    Simulation & Visualization

    Real-Time Animations, Matplotlib, Open3D, COVID-19 Spread Simulations, EEG Signal Visualization

    .

    Cloud & Robotics Tools

    NVIDIA Omniverse, NVIDIA Jetson Ecosystem, Git-based CI/CD, Python Automation, Intro to Cloud Platforms (AWS, GCP)

  • Soft Skills

    Analytical Thinking & Problem Solving

    Demonstrated through physics-based simulations, ML pipelines, and embedded control systems

    .

    Team Collaboration & Communication

    Proven via collaborative internships, GitHub repos, and contributions to open-source tools

    .

    Documentation & Technical Writing

    Maintained wikis, Jupyter notebooks, and project blogs; emphasized reproducibility

    .

    Adaptability & Rapid Learning

    Gained experience in 25+ project domains: GANs, SLAM, LiFi, RL, 3D Vision, and more

    .

    Research Mindset & Scientific Curiosity

    Designed solvers for chaotic systems, EEG signal decoding, and scientific model validation

    .

    Ethical Robotics & Safety-Aware Design

    Bias-sensitive ML model development, fail-safe robotics logic, and transparent benchmarking

Projects

NeuroPsych Trading Assistant: A Neuromorphic Multi-Agent System with Brain-Computer Interface for Computational Psychiatry in Financial Markets

Description:

The NeuroPsych Trading Assistant represents a groundbreaking convergence of neuromorphic computing, computational psychiatry, robotics, and electronic systems design to address the critical mental health crisis among retail traders. This project develops a comprehensive ecosystem that monitors, predicts, and intervenes in real-time to prevent emotion-driven trading losses and mental health deterioration.

My system employs cutting-edge neuromorphic hardware design, EEG-based brain-computer interfaces, computer vision, multi-agent AI coordination, and robotic companions to create the world's first comprehensive mental health support system for high-stress financial decision-making.

Learn More (Coming Soon)

View (Coming Soon)

Internship Semester (UG Final Semester) Project

Problem Statement:

Aerial vehicles struggle with real-time, low-latency object detection due to small object sizes, computational constraints, and dynamic environments. This project addresses the gap by deploying an edge-optimized YOLOv7 model to enable accurate, real-time detection on drones without cloud dependency.

Summary:

Developed a real-time aerial object detection system using YOLOv7, trained on a custom dataset with NVIDIA Jetson AGX Xavier. Deployed on the "Tunga" aerial vehicle (NVIDIA Jetson Nano + Pixhawk) to enable edge-computing for dynamic environments. Achieved 89% mAP, 22 FPS inference speed, and 95% real-world detection accuracy, optimizing resource usage by 40% compared to baseline models. Demonstrated scalability for aerial surveillance and disaster response applications.

Learn More

Time-Frequency Analysis of Neural Signals Using Complex Morlet Wavelets

Problem Statement:

EEG signal analysis faces challenges in balancing temporal and spectral resolution while mitigating edge artifacts. This project addressed these limitations by implementing complex Morlet wavelets to enable high-precision time-frequency decomposition and comparing it with traditional filter-Hilbert methods for robust neural oscillation characterization.

Summary:

This project leveraged complex Morlet wavelet convolution to analyze EEG data, extracting time-frequency power and phase dynamics from single and multi-trial neural signals. Key tasks included comparing wavelet convolution with filter-Hilbert methods (achieving 25% higher spectral precision), quantifying edge artifacts in transient signals, and visualizing multi-channel spectral activity. Results revealed robust inter-trial phase coherence (ITPC > 0.4) in high-frequency bands and identified boundary artifacts in non-stationary data. The study demonstrated wavelet convolution’s superiority in resolving frequency-specific neural oscillations, offering insights for improved EEG signal processing in cognitive and clinical neuroscience applications.

Learn More

View

Comparative Analysis of Pathfinding Algorithms (A* ,BFS & DFS)

Problem Statement:

Existing pathfinding algorithms vary in efficiency and optimality for maze navigation. This project compares DFS, BFS, and A* to determine their effectiveness in minimizing path length, search time, and heuristic impact on performance.

Summary:

This project implements and evaluates DFS, BFS, and A* algorithms in maze navigation using PyAmaze to visualize pathfinding and measure efficiency. Results show BFS guarantees the shortest path, while A* (with Manhattan heuristic) reduces search time by 25–40% compared to DFS. The Manhattan heuristic outperformed Euclidean in 70% of test cases, achieving shorter search paths. Metrics include path length (BFS/A: 100% optimal vs. DFS: 60% longer paths) and search efficiency (A explored 35% fewer nodes than BFS). The study highlights trade-offs between completeness, speed, and heuristic choice for real-world navigation systems.

Learn More

View

Autonomous Lane Changing Control System (Lateral Dynamics Simulation)

Problem Statement:

Traditional autonomous lane-changing systems struggle to balance dynamic constraints and real-time adaptability. This project addresses the challenge of designing a computationally efficient control framework using MPC to ensure safe, smooth lateral maneuvers while tracking time-varying trajectories under hardware limitations.

Summary:

This MATLAB project designs a Model Predictive Control (MPC) system for autonomous vehicles to execute precise lane-changing maneuvers. The MPC algorithm dynamically optimizes steering inputs over a receding horizon to track a reference trajectory, balancing computational efficiency and control accuracy. The simulation framework includes trajectory generation, state-space modeling, and iterative quadratic optimization (quadprog) to resolve constraints. Results demonstrate >95% trajectory tracking accuracy, steering angles within ±0.35 rad limits, and stable lateral deviation (less than 0.15m) under varying conditions. Visualization of states (yaw rate, position) and inputs validates the controller’s robustness for real-time applications, with adaptive horizon tuning reducing solve times by 20%. The project highlights MPC’s capability to handle nonlinear vehicle dynamics while ensuring safety-critical performance.

Learn More

View

Speed Estimation and Vehicle Tracking System

Problem Statement:

Traditional traffic monitoring systems lack accuracy in real-time speed estimation and struggle with occlusions in dense traffic. This project addresses these gaps by automating vehicle detection, tracking, and speed calculation using YOLOv8 and computer vision to enable efficient, scalable traffic analysis.

Summary:

This project leverages YOLOv8's state-of-the-art object detection to identify and track vehicles in real-time video streams. A custom tracking algorithm assigns persistent IDs to vehicles, enabling precise speed calculation as they cross two predefined lines. The system achieved 85%+ detection accuracy, tracked 100+ vehicles simultaneously, and computed speeds with less than 10% error relative to ground truth. Results were visualized using OpenCV, displaying bounding boxes, dynamic speed labels (in km/h), and traffic metrics (e.g., 45 vehicles moving downward vs. 32 upward). Designed for scalable traffic analysis, this solution demonstrates robust performance in real-world scenarios, offering insights for urban planning and congestion management.

Learn More

View

Real-Time Depth Estimation with MiDaS

Problem Statement:

Traditional depth sensing relies on specialized hardware (e.g., LiDAR, stereo cameras), which is costly and computationally intensive. This project addresses this gap by implementing a real-time, GPU-accelerated monocular depth estimation system using lightweight MiDaS models to democratize 3D perception from standard 2D webcams.

Summary:

This project implements a real-time monocular depth estimation system using PyTorch and MiDaS models (DPT_Large, DPT_Hybrid, MiDaS_small) to infer 3D structure from 2D webcam input. Leveraging GPU acceleration (NVIDIA RTX 3060), it processes 20–30 FPS with optimized latency, balancing accuracy and speed: DPT_Large achieved ±5% relative depth error but slower inference (~15 FPS), while MiDaS_small prioritized speed (~30 FPS) for real-time applications. The pipeline integrates OpenCV for live video capture, PyTorch for model inference, and color-mapped depth visualization, demonstrating a hardware-efficient alternative to traditional depth sensors like LiDAR.

Learn More

View

Lip Read To Text

Problem Statement:

Existing visual speech recognition systems struggle to accurately transcribe spoken words from lip movements due to variable lighting, speaker differences, and lack of temporal alignment. This project addresses these challenges by developing a deep learning model (3D CNN + Bidirectional LSTM) to automate silent speech interpretation with robust spatiotemporal feature extraction and CTC-based alignment-free training.

Summary:

This project develops a LipNet-based lip reading model using TensorFlow/Keras, integrating 3D CNNs and Bidirectional LSTMs to analyze spatiotemporal lip movements. The system achieves 18% word error rate (WER) on test data, trained with CTC loss for alignment-free text prediction. Preprocessing includes frame normalization and lip ROI extraction, while evaluation shows 83% accuracy on short phrases. Future enhancements target larger datasets and transformer-based architectures for improved robustness.

Learn More

View

Real Time Hand Gesture Recognition For Cursor Control

Problem Statement:

Traditional input devices limit mobility and accessibility. This project addresses this by designing a vision-based system to control cursor movements and clicks through hand gestures, eliminating physical hardware dependency.

Summary:

Developed a real-time hand gesture recognition system using MediaPipe and PyAutoGUI to enable touchless cursor control. The solution processes webcam input to detect 21 hand landmarks, achieving 95% gesture accuracy, maps finger movements to screen coordinates with less than 50ms latency, and triggers mouse clicks via pinch detection (index-thumb distance <20px). Tested at 30 FPS, it offers intuitive, low-latency hands-free navigation for enhanced accessibility.

Learn More

View

Optimized Ludo with Q-Learning

Problem Statement:

Traditional rule-based Ludo AI lacks adaptability to dynamic game states and opponent strategies. This project addresses the need for an autonomous agent that learns optimal moves through trial and error, using Q-learning to balance short-term rewards and long-term winning strategies.

Summary:

This project implements a Q-learning-based AI agent to master the strategic board game Ludo. By defining a state-action space and iteratively updating a Q-table using rewards, the agent learns optimal moves through exploration (epsilon-greedy policy) and exploitation. The AI was trained over 10,000 simulated games using Python and Ludopy, achieving an 85% win rate against rule-based opponents and a 40% reduction in average move decision time. Key innovations include dynamic reward tuning for safe piece advancement and blocking adversaries. The modular design, powered by NumPy for efficient Q-table updates, enabled the agent to adapt to complex board states, demonstrating a 92% success rate in avoiding captures. Results validate reinforcement learning’s potential for dynamic strategy games.

Learn More

View

Deep Audio Classifier

Problem Statement:

Urban sound recognition is challenging due to overlapping acoustic patterns and environmental noise. This project solves this by developing an MFCC-driven neural network to classify sounds accurately, enabling scalable noise monitoring and urban analytics.

Summary:

This project tackles urban sound classification using the UrbanSound8K dataset by extracting Mel-Frequency Cepstral Coefficients (MFCC) to train a neural network. The model achieved 85% test accuracy and was trained in under 1 hour, efficiently distinguishing 10 sound classes (e.g., drilling, sirens, street music) despite background noise. By combining dropout regularization and Adam optimization, the system demonstrates robust performance for real-world applications like noise pollution monitoring and smart city infrastructure.

Learn More

View

Engineering (UG) Group Project (Capstone Project)

Problem Statement:

Traditional RF-based wireless networks face spectrum congestion, security vulnerabilities, and interference issues. This project addresses these limitations by prototyping a LiFi system that leverages visible light for high-bandwidth, low-risk, and energy-efficient PC-to-PC communication.

Summary:

Designed and implemented a LiFi system enabling secure, high-speed data transmission between PCs using modulated LED light and photodetectors. Developed hardware (transceiver circuits) and software (encoding/decoding protocols), achieving a 15 Mbps data rate, less than 10⁻⁵ bit error rate, and 2-meter transmission range. Demonstrated LiFi’s viability as a low-latency, energy-efficient alternative to congested RF networks (WiFi), with potential applications in EMI-sensitive environments. Validated through real-time text/file transfers, offering a scalable foundation for future light-based communication systems.

Learn More

Smart Dustbin

Problem Statement:

Manual dustbin lids pose hygiene risks due to frequent contact and often remain open, causing odor and spillage. This project addresses these issues by automating lid operation using sensors, ensuring touchless disposal and timely closure.

Summary:

Designed an Arduino UNO-based smart dustbin with an ultrasonic sensor to enable touchless, hygienic waste disposal. The system detects proximity (up to 50 cm) and opens automatically, reducing physical contact and spillage. Achieved 95% detection accuracy, 0.5-second response time, and 60% improvement in user convenience, validated through 100+ test cycles. Ideal for public spaces to promote cleanliness and operational efficiency.

Learn More
See More

Certifications

ROS 2

The Robot Operating System (ROS) is an open-source framework that helps researchers and developers build and reuse code between robotics applications.

Applied Control Systems

This module will introduce the student to key topics within control and signal processing, developing understanding through a combination of theoretical content and practical application.

Neural Signal Processing

Neural signals consist of recordings of potentials that are presumably generated by mixing some underlying components of brain activity.

Mastering Microcontroller

Microcontroller is a compressed micro computer manufactured to control the functions of embedded systems in office machines, robots, home appliances, motor vehicles, and a number of other gadgets.

Jetson Nano Boot Camp

The Jetson Nano module is a small AI computer that gives you the performance and power efficiency to take on modern AI workloads, run multiple neural networks in parallel, and process data from several high-resolution sensors simultaneously.

CUDA Programming in C++

CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.

VLSI SoC Design using Verilog HDL

VLSI design involves creating integrated circuits by combining thousands to billions of transistors on a single chip, enabling the development of complex electronic systems.

AI

Artificial Intelligence (AI), the ability of a digital computer or computer-controlled robot to perform tasks commonly associated with intelligent beings.

Modern Computer Vision

Computer vision is a field of computer science that focuses on enabling computers to identify and understand objects and people in images and videos.

Disaster Risk Monitoring using Satellite Imagery (NVIDIA)

One remarkable instance of satellite imagery in disaster management was during the 2011 Japan earthquake and tsunami. Satellite data helped assess the extent of the damage, guiding rescue efforts and aid distribution.

Python

Python is a high-level, general-purpose programming language. Its design philosophy emphasizes code readability with the use of significant indentation.

Introduction to Cloud Computing

Cloud computing is the on-demand availability of computer system resources, especially data storage and computing power, without direct active management by the user.

Flight Dynamics with Tensors

Flight dynamics is shifting from vectors to tensors in order to adapt to the ever-increasing computer power available for solving complex aerospace problems.

Model, Simulate and Control a Drone in MATLAB & SIMULINK

Design and Simulate the Aerodynamics of Propellers in MATLAB

Develop ,Customize ,and Publish in Omniverse with Extensions

NVIDIA Omniverse is a scalable, multi-GPU real-time development platform for building and operating metaverse apps.

My Bin

My learning through Research

Research from Youtube, Internet Articles and Research papers.

Contact Me

ksrujan_be19@thapar.edu

kt.srujan@gmail.com

+91 9100725768