Skills

  • AI/ML

    Supervised & Unsupervised Learning

    (Classification, Clustering, Dimensionality Reduction)

    .

    Deep Learning Architectures

    CNNs, RNNs, LSTMs, Transformers, GANs, Autoencoders

    .

    Reinforcement Learning

    Q-Learning, Deep Q Networks (DQN), PPO, A2C

    .

    Natural Language Processing (NLP)

    BERT, GPT-based LLMs, Tokenization, Named Entity Recognition, Sentiment Analysis, Text-to-Speech (TTS), Speech-to-Text, Prompt Engineering

    .

    Transfer Learning & Model Compression

    PEFT, LoRA, QLoRA, Knowledge Distillation, Fine-Tuning HuggingFace Transformers

    .

    Classical ML Algorithms

    SVM, KNN, Random Forest, Logistic/Linear Regression, Decision Trees

    .

    MLOps

  • Generative & Predictive Modeling

    Time Series Forecasting

    ARIMA, SARIMAX, NeuralProphet, Rolling Window Analysis

    .

    Generative Modeling

    GANs, ESRGAN, Diffusion Models (Stable Diffusion, Latent Diffusion), LangChain, LLM, RAG

    .

    Pattern & Anomaly Detection

    Isolation Forest, k-Means Clustering, Sensor-based Anomalies

    .

    Semantic Similarity & Embeddings

    Word2Vec, GloVe, BERT Embeddings, Sentence Transformers

  • Computer Vision & 3D Reconstruction

    Supervised & Unsupervised Learning

    (Classification, Clustering, Dimensionality Reduction)

    .

    Deep Learning Architectures

    CNNs, RNNs, LSTMs, Transformers, GANs, Autoencoders

    .

    Reinforcement Learning

    Q-Learning, Deep Q Networks (DQN), PPO, A2C

    .

    Natural Language Processing (NLP)

    BERT, GPT-based LLMs, Tokenization, Named Entity Recognition, Sentiment Analysis, Text-to-Speech (TTS), Speech-to-Text, Prompt Engineering

    .

    Transfer Learning & Model Compression

    PEFT, LoRA, QLoRA, Knowledge Distillation, Fine-Tuning HuggingFace Transformers

    .

    Classical ML Algorithms

    SVM, KNN, Random Forest, Logistic/Linear Regression, Decision Trees

  • Data Handling & Optimization

    Data Cleaning & Preprocessing

    Handling Missing Data, Label Encoding, One-Hot Encoding, CI/CD Pipeline,ETL Pipeline(Apache Airflow)

    .

    Feature Engineering & Selection

    PCA, Statistical Feature Extraction, Mutual Information

    .

    Model Evaluation & Tuning

    Cross-Validation, ROC-AUC, Grid Search, Bayesian Optimization

    .

    Optimization Techniques

    Gradient Descent, Adam, Particle Swarm Optimization (PSO)

    .

    Big Data & GPU Computing

    RAPIDS.ai (cuDF, cuML), CUDA Acceleration, Multi-GPU Training

    .

    Classical ML Algorithms

    SVM, KNN, Random Forest, Logistic/Linear Regression, Decision Trees

  • Frameworks & Libraries

    ML/DL Frameworks

    TensorFlow, PyTorch, Keras, Fastai, Scikit-learn

    .

    Data & Viz

    Pandas, NumPy, Matplotlib, Seaborn, Plotly

    .

    Vision/NLP

    OpenCV, HuggingFace Transformers, Gensim, NLTK, spaCy

    .

    GPU-Accelerated Libraries

    RAPIDS (cuDF, cuML, cuGraph), CUDA Programming using

    .

    Simulation & Animation

    Omniverse Isaac Sim, Matplotlib Animations, OpenAI Gym

  • Tools, IDEs & Platforms

    IDEs

    Jupyter Lab, VS Code, PyCharm, Google Colab

    .

    Version Control

    Git, GitHub ,GitHub Actions

    .

    Virtualization & Deployment

    Docker, TensorFlow Lite

    .

    Cloud Platforms

    AWS (SageMaker, EC2, S3), Google Cloud, Hugging Face Hub, Flask, FastAPI, Astronomer

    .

    Experiment Tracking

    TensorBoard, Weights & Biases, Grafana, PostgreSQL

    .

    Linux ,Shell Scripting & Coding

    Ubuntu, bash/zsh scripting, Cursor AI(For Coding Assistance)

  • Soft Skills

    Problem Solving & Critical Thinking

    Proven by diverse AI applications across vision, NLP, and forecasting

    .

    Team Collaboration

    Experience in collaborative research projects and internships

    .

    Communication

    Technical blogging, documentation, and public GitHub repos/h1>

    .

    Research Mindset

    Scholarly projects (e.g., Exoplanet detection, 3D protein structure)

    .

    Ethical AI Awareness

    Bias mitigation, fairness, and data privacy best practices

    .

    Adaptability & Learning Agility

    Rapid prototyping, continuous learning of emerging AI tech

Projects

NeuroPsych Trading Assistant: A Neuromorphic Multi-Agent System with Brain-Computer Interface for Computational Psychiatry in Financial Markets

Description:

The NeuroPsych Trading Assistant represents a groundbreaking convergence of neuromorphic computing, computational psychiatry, robotics, and electronic systems design to address the critical mental health crisis among retail traders. This project develops a comprehensive ecosystem that monitors, predicts, and intervenes in real-time to prevent emotion-driven trading losses and mental health deterioration.

My system employs cutting-edge neuromorphic hardware design, EEG-based brain-computer interfaces, computer vision, multi-agent AI coordination, and robotic companions to create the world's first comprehensive mental health support system for high-stress financial decision-making.

Learn More (Coming Soon)

View (Coming Soon)

Internship Semester (UG Final Semester) Project

Problem Statement:

Aerial vehicles struggle with real-time, low-latency object detection due to small object sizes, computational constraints, and dynamic environments. This project addresses the gap by deploying an edge-optimized YOLOv7 model to enable accurate, real-time detection on drones without cloud dependency.

Summary:

Developed a real-time aerial object detection system using YOLOv7, trained on a custom dataset with NVIDIA Jetson AGX Xavier. Deployed on the "Tunga" aerial vehicle (NVIDIA Jetson Nano + Pixhawk) to enable edge-computing for dynamic environments. Achieved 89% mAP, 22 FPS inference speed, and 95% real-world detection accuracy, optimizing resource usage by 40% compared to baseline models. Demonstrated scalability for aerial surveillance and disaster response applications.

Learn More

News Researcher AI Agents

Problem Statement:

The rapid evolution of AI in robotics and healthcare creates an overwhelming volume of innovations, but manual research and content creation struggle to deliver timely, accurate, and engaging articles, hindering stakeholders from accessing actionable insights. This project automates the end-to-end process using AI agents to research, analyze, and generate high-quality technical content at scale.

Summary:

This AI-driven system automates and coordinates technical content creation using the CrewAI framework, integrating advanced AI models (Gemini-1.5 Flash) for research and narrative generation. The research agent identifies breakthroughs in robotics and healthcare with 85% accuracy, while the writing agent produces 40+ articles daily, combining technical depth with audience-friendly formatting. Collaborative workflows reduce manual effort by 70%, enabling rapid, scalable dissemination of cutting-edge insights.

Learn More

View

Stock Trading Reinforcement Learning

Problem Statement:

Traditional algorithmic trading strategies struggle to adapt to dynamic market regimes, leading to suboptimal decisions during volatility shifts or black swan events. This project addresses this gap by developing reinforcement learning models that autonomously learn and evolve strategies from historical data, optimizing for risk-adjusted returns in unpredictable financial environments.

Summary:

This project develops and evaluates adaptive trading strategies using reinforcement learning (RL) with historical market data. We implemented A2C (LSTM-based) and PPO (MLP-based) algorithms in a simulated gym-anytrading environment, achieving a 18.5% cumulative return (A2C) and 12.2% return (PPO) while maintaining Sharpe ratios of 1.32 and 1.65. Performance was validated through intuitive visualizations of trades against price trends, revealing A2C’s strength in trend capture and PPO’s resilience in volatility.

Learn More

View

Fine-Tuning Llama-2-7b With LORA And QLoRA

Problem Statement:

Traditional fine-tuning of large models like Llama-2-7b demands prohibitive computational resources. This project addresses the challenge of efficiently adapting such models using QLoRA, enabling cost-effective customization without sacrificing performance.

Summary:

This project demonstrates fine-tuning the Llama-2-7b model using QLoRA, a 4-bit quantization method, to optimize memory usage while preserving performance. Leveraging the guanaco-llama2-1k dataset, the model was trained with LoRA configurations (rank=64, alpha=16) for 1 epoch, achieving 40% faster training and 50% reduced GPU memory consumption compared to full fine-tuning. Post-training evaluation via text-generation pipelines confirmed coherent outputs (e.g., 98% accuracy on domain-specific prompts). The model was compressed to 6.8GB and deployed to Hugging Face Hub, showcasing efficient adaptation of large language models for resource-constrained environments.

Learn More

View

RAG-Powered QA System with LLAMA3

Problem Statement:

Traditional chatbots struggle with domain-specific queries due to static knowledge bases. This project solves dynamic information retrieval by creating a RAG system that combines real-time document processing with LLAMA3's reasoning, enabling accurate answers from user-provided technical content.

Summary:

Built a Retrieval-Augmented Generation (RAG) system using LLAMA3-70B and NVIDIA embeddings to enable context-aware question answering over custom documents. The system processes PDFs via LangChain text splitting, generates vector embeddings with NVIDIA models, and retrieves context using FAISS for low-latency semantic search. A Streamlit interface allows users to upload documents, trigger embeddings (processing 30+ document chunks in less than 2 sec), and query the AI (avg. response time: 1.8 sec). Achieved 89% accuracy in contextual QA benchmarks through optimized chunking strategies and hybrid retrieval. This end-to-end pipeline demonstrates efficient handling of domain-specific knowledge without model retraining.

Learn More

View

Sentiment Analysis Of Yelp Reviews

Problem Statement:

Manual analysis of vast customer reviews is time-intensive and prone to subjective bias. This project addresses the challenge by automating sentiment classification using BERT to deliver scalable, objective insights from Yelp reviews.

Summary:

This project automates sentiment classification of customer reviews from Yelp using web scraping and a pre-trained BERT model. By extracting reviews from a Yelp page, preprocessing text data, and leveraging NLP techniques, the system assigns sentiment scores (1-5) to quantify customer opinions. Results showed 85% accuracy in categorizing sentiments (positive, neutral, negative), with 200+ reviews processed and visualized in a structured DataFrame. The solution enables businesses to efficiently gauge customer satisfaction trends and identify actionable insights from unstructured feedback.

Learn More

View

Brain Tumor Classification

Problem Statement:

Manual diagnosis of brain tumors from MRI scans is time-consuming and prone to human error. This project addresses the need for an automated, accurate classification system using machine learning to improve diagnostic efficiency and reliability.

Summary:

This project demonstrates the development of a machine learning system to classify brain tumors in MRI images into "no tumor" and "pituitary tumor" categories. It involves preprocessing MRI scans (resizing, grayscale conversion, normalization) and training Logistic Regression and SVM models. The system achieved 97.14% accuracy with Logistic Regression and 95.51% with SVM, validated on a test dataset. Misclassification analysis highlighted robustness, with only 38 mislabeled samples out of 879. The final implementation includes a user-friendly interface for rapid tumor detection, emphasizing its potential to assist medical diagnostics. Future extensions could incorporate deep learning and multi-class tumor classification.

Learn More

View

Neuron Segmentation

Problem Statement:

Manual neuron segmentation in microscopy images is time-consuming, subjective, and error-prone. This project addresses the need for an automated, scalable solution using SAM to achieve pixel-accurate neuron delineation, improving reproducibility in neurobiological studies.

Summary:

This project leverages Meta's Segment Anything Model (SAM) to automate neuron segmentation in microscopy images, accelerating neurobiological analysis. Using PyTorch and GPU acceleration (NVIDIA RTX 3060), SAM generates precise segmentation masks with a 0.92 mean IoU (Intersection over Union) and processes images 40% faster than manual annotation. The model adapts to diverse neuron morphologies via customizable parameters like pred_iou_thresh=0.9 and min_mask_region_area=100, validated through quantitative metrics and visual inspection. The solution demonstrates SAM’s versatility for biomedical imaging tasks while maintaining computational efficiency for scalable research applications.

Learn More

View

Lip Read To Text

Problem Statement:

Existing visual speech recognition systems struggle to accurately transcribe spoken words from lip movements due to variable lighting, speaker differences, and lack of temporal alignment. This project addresses these challenges by developing a deep learning model (3D CNN + Bidirectional LSTM) to automate silent speech interpretation with robust spatiotemporal feature extraction and CTC-based alignment-free training.

Summary:

This project develops a LipNet-based lip reading model using TensorFlow/Keras, integrating 3D CNNs and Bidirectional LSTMs to analyze spatiotemporal lip movements. The system achieves 18% word error rate (WER) on test data, trained with CTC loss for alignment-free text prediction. Preprocessing includes frame normalization and lip ROI extraction, while evaluation shows 83% accuracy on short phrases. Future enhancements target larger datasets and transformer-based architectures for improved robustness.

Learn More

View

Weather Prediction With NeuralProphet

Problem Statement:

Rising climate variability in Williamtown demands accurate temperature forecasts to mitigate risks for agriculture and infrastructure. This project addresses the gap in localized, long-term predictions by leveraging NeuralProphet to model historical weather patterns and generate actionable forecasts.

Summary:

This project forecasts temperature trends in Williamtown, Australia, using NeuralProphet, a hybrid time-series model combining neural networks and classical forecasting. Historical weather data (2007–2015) was preprocessed to isolate daily 3 PM temperatures, trained over 1,000 epochs to capture seasonal patterns and long-term trends. The model achieved robust performance (visualized forecasts aligned closely with historical trends) and predicted temperatures for 3+ years ahead, with residuals indicating consistent accuracy. Components like seasonality, trend, and uncertainty intervals were analyzed to validate reliability. The model was serialized with pickle for scalable deployment, demonstrating adaptability for climate analytics in agriculture, urban planning, and disaster preparedness.

Learn More

View

Optimizing Census Data Analysis with GPU-Accelerated ML

Problem Statement:

Traditional CPU-based processing of large datasets like the U.S. Census suffers from computational bottlenecks, while model selection for imbalanced categorical data remains inefficient. This project solves these by implementing GPU-accelerated preprocessing with RAPIDS and systematically comparing ML models to identify the optimal approach for income prediction.

Summary:

This project harnessed RAPIDS' GPU acceleration to preprocess and analyze the U.S. Census dataset, addressing scalability challenges in handling 45,000+ entries with categorical features. Categorical encoding, standardization, and GPU-optimized workflows reduced preprocessing time by 60% compared to CPU methods. Four models—Logistic Regression, K-Nearest Neighbors, Random Forest, and SVM—were evaluated, with Random Forest achieving the highest accuracy (85.2%) and SVM demonstrating the best precision (88%) for ">50K" income prediction. Results highlighted a 15% performance gap between the best (Random Forest) and weakest (KNN) models, validated via confusion matrices. The end-to-end pipeline showcased RAPIDS' ability to accelerate data science workflows while maintaining interpretability, proving its value for real-world demographic analysis.

Learn More

View

Rock-Paper-Scissors Hand Gesture Classifier with FasterViT

Problem Statement:

Traditional CNN-based gesture classifiers struggle with balancing speed and accuracy for real-time applications. This project addresses this gap by implementing FasterViT, a vision transformer optimized for computational efficiency, to achieve sub-20ms inference times with >98% accuracy in dynamic hand gesture recognition.

Summary:

This project leverages the FasterViT architecture to build a real-time hand gesture classifier for rock-paper-scissors games. The dataset was enhanced using advanced preprocessing (random resizing, cropping, normalization) and augmentation techniques to improve model generalization. The fine-tuned FasterViT model achieved 98.2% validation accuracy in just 5 training epochs, optimized via GPU acceleration and adaptive learning rates. The deployed model demonstrates rapid inference times (less than 15ms/image) on an RTX 3060 GPU, enabling seamless real-time predictions. Practical integration was validated by overlaying predictions on test images, showcasing robust performance across diverse lighting and gesture variations.

Learn More

View
See More

Certifications

Python

Python is a high-level, general-purpose programming language. Its design philosophy emphasizes code readability with the use of significant indentation.

AI

Artificial Intelligence (AI), the ability of a digital computer or computer-controlled robot to perform tasks commonly associated with intelligent beings.

Tensor Flow

TensorFlow is an open-source machine learning framework developed by Google for building and deploying machine learning models.

PyTorch

PyTorch is a machine learning library based on the Torch library, used for applications such as computer vision and natural language processing, originally developed by Meta AI and now part of the Linux Foundation umbrella.

Modern Computer Vision

Computer vision is a field of computer science that focuses on enabling computers to identify and understand objects and people in images and videos.

Introduction to Cloud Computing

Cloud computing is the on-demand availability of computer system resources, especially data storage and computing power, without direct active management by the user.

Prompt Engineering for AI

Prompt engineering is the process where you guide generative artificial intelligence (generative AI) solutions to generate desired outputs.

Intro to AI Agents: Build an Army of Digital Workers with AI

In intelligence and artificial intelligence, an intelligent agent is an agent acting in an intelligent manner. It perceives its environment, takes actions autonomously in order to achieve goals, and may improve its performance with learning or acquiring knowledge.

MLOps

2025 Bootcamp: Generative AI, LLM Apps, AI Agents, Cursor AI

AI Application Boost with Nvidia Rapids Acceleration

RAPIDS is a suite of open-source software libraries and APIs for executing data science pipelines entirely on GPUs—and can reduce training times from days to minutes.

Disaster Risk Monitoring using Satellite Imagery (NVIDIA)

One remarkable instance of satellite imagery in disaster management was during the 2011 Japan earthquake and tsunami. Satellite data helped assess the extent of the damage, guiding rescue efforts and aid distribution.

Develop ,Customize ,and Publish in Omniverse with Extensions

NVIDIA Omniverse is a scalable, multi-GPU real-time development platform for building and operating metaverse apps.

My Bin

My learning through Research

Research from Youtube, Internet Articles and Research papers.

Contact Me

ksrujan_be19@thapar.edu

kt.srujan@gmail.com

+91 9100725768