Introduction

Research Scholar at IIT Dharwad pursuing advanced research in data science and machine learning applications.
Lead Developer with extensive experience in designing and delivering sophisticated R Shiny applications and custom R & Python development.
Currently working as Technical Consultant at Techworkslab, specializing in interactive data visualizations and building robust, scalable solutions for complex analytical needs.

Passionate about leveraging cutting-edge tools and methodologies to build innovative, user-centric applications and deliver actionable insights.
Expertise in CDISC SDTM standards, API development, production-scale deployment of data-driven solutions, and academic research in AI/ML domains.

Professional Summary

  • Production-Scale Deployment: Proven track record in deploying Shiny applications, APIs, and data pipelines in CI/CD-managed environments, ensuring high availability and scalability
  • Standards Implementation: Expertise in implementing CDISC SDTM standards and developing regulatory-compliant workflows using R for clinical data management
  • Code Quality & Documentation: Skilled in preparing thorough package documentation, creating test cases, and adhering to best practices in R & Python development
  • API Integration & Data Interoperability: Experience in building and integrating APIs to enable seamless data exchange between applications and platforms
  • Leadership & Collaboration: Demonstrated ability to lead development teams, mentor junior developers, and drive cross-functional collaboration to deliver exceptional results
  • R Shiny Expertise: Advanced proficiency in developing interactive, high-quality data visualizations and sophisticated R Shiny applications for diverse clients
  • Technical Stack: Expert-level knowledge in R, Python (Flask, FastAPI, Django), GitLab, Docker, Linux, MongoDB, PostgreSQL, and AWS services
  • Clinical Data Experience: Specialized experience in clinical trials reporting systems, SDTM/ADaM programming, and regulatory compliance workflows
  • Automated Testing & CI/CD: Championed adoption of automated testing frameworks and continuous integration pipelines for enhanced reliability
  • Training & Mentorship: Conducted training sessions for client teams on best practices in R programming, Shiny app development, and API implementation

Education & Research

  • Present - Research Scholar - IIT Dharwad (PhD in Data Science & AI)
  • 2022 - M. Tech, Artificial Intelligence - Reva University
  • 2016 - B. Tech, Aerospace Engineering - Hindustan Aviation Academy
  • 2011 - 12th Standard - Kendriya Vidyalaya, Belgaum
  • 2009 - 10th Standard - Army Public School, Darjeeling

  • Key Certifications

  • Microsoft Certified: Azure AI Engineer Associate
  • Kaggle: Natural Language Processing & Python
  • LinkedIn: Advanced NLP with Python for Machine Learning
  • LinkedIn: Building Deep Learning Applications with Keras & TensorFlow
  • Udemy: R Shiny Interactive Web Apps & Advanced Analytics

  • Skills

    R Shiny Development

    Production-grade R Shiny Applications
    Interactive Data Visualizations
    Performance Optimization

    Data & Analytics

    CDISC SDTM Standards
    Clinical Data Management
    Automated Analytics

    Programming Languages

    R (Expert)
    Python (Flask, FastAPI, Django)
    JavaScript, SAS

    Development & Deployment

    Docker & Containerization
    GitLab CI/CD Pipelines
    Linux Server Administration

    API Development

    RESTful API Design
    R Plumber APIs
    API Integration & Testing

    Database Technologies

    MongoDB, PostgreSQL
    SQL Server, SQLite
    Data Pipeline Design

    Cloud & Infrastructure

    AWS Services
    Azure AI Engineer Associate
    Production Deployments

    Quality & Documentation

    Automated Testing Frameworks
    R Markdown Documentation
    Package Development

    Featured Projects Portfolio

    Comprehensive showcase of clinical data management, AI/ML research, and enterprise application development

    Clinical Data Pooling Platform

    Role: Project Lead | Duration: 8 months | Team Size: 5 developers

    Project Overview:
    Architected and delivered a comprehensive R Shiny platform designed for SDTM/ADaM data pooling across multiple clinical studies. The platform features metadata-driven mapping capabilities, interactive resolution dashboards, and automated data standardization workflows that significantly reduced manual data processing time by 75%.

    Key Achievements:
    • Implemented automated CDISC compliance validation with real-time error detection
    • Designed intuitive dashboard interfaces reducing data review time from weeks to days
    • Integrated with 15+ clinical data management systems via secure APIs
    • Established comprehensive audit trails for regulatory compliance (FDA/EMA standards)

    Technical Architecture: R Shiny (Frontend), PostgreSQL (Data Layer), Docker (Containerization), GitLab CI/CD (Deployment), AWS S3 (Storage)

    Industry Impact: Deployed across 3 pharmaceutical companies, processing 500+ clinical studies with 99.7% data accuracy

    Data Extractor – GenAI Chart Interpreter

    Role: Project Lead & Lead Developer | Duration: 6 months | Research Collaboration: IIT Dharwad

    Project Overview:
    Developed a cutting-edge multimodal AI pipeline that revolutionizes clinical data extraction from visual charts and plots. The system combines state-of-the-art Pix2Struct, DePlot, and OCR technologies to automatically extract, interpret, and summarize tabular insights from complex medical visualizations including survival plots, bar charts, and waterfall diagrams.

    Key Innovations:
    • Achieved 94% accuracy in chart data extraction vs 65% manual baseline
    • Developed novel preprocessing algorithms for medical chart standardization
    • Implemented multi-modal fusion techniques for enhanced interpretation
    • Created automated quality validation system for extracted data integrity

    Technical Architecture: Python (Core), Pix2Struct (Vision), DePlot (Chart Understanding), Tesseract OCR, PyTorch, FastAPI (API Layer), Redis (Caching)

    Research Impact: Published methodology contributed to 2 peer-reviewed papers; system processes 10,000+ charts weekly in production

    Clinical Trials Reporting System

    Role: Lead Developer | Duration: 12 months | Regulatory Focus: FDA/EMA Compliance

    Project Overview:
    Architected and implemented a comprehensive end-to-end clinical trials reporting ecosystem that automates the entire workflow from raw data ingestion to regulatory submission-ready reports. The system handles SDTM (Study Data Tabulation Model), ADaM (Analysis Data Model), and TLG (Tables, Listings, and Graphics) generation with full CDISC compliance and automated Clinical Study Report (CSR) generation.

    Key Achievements:
    • Reduced report generation time from 6 weeks to 2 days (95% time savings)
    • Achieved 100% CDISC compliance with automated validation checks
    • Implemented version control and audit trails for regulatory transparency
    • Enabled parallel processing of multiple studies reducing resource bottlenecks by 80%

    Technical Architecture: R (Statistical Computing), R Markdown (Report Generation), SAS (Legacy Integration), GitLab CI/CD (Automation), Docker (Containerization), Linux (Deployment)

    Regulatory Impact: Successfully supported 25+ FDA submissions with zero compliance issues; adopted by 5 CROs as standard reporting platform

    Risk-Based Monitoring Dashboard

    Role: Data Engineer | Duration: 4 months | Scope: Multi-site Clinical Trials

    Project Overview:
    Engineered an intelligent risk-based monitoring system that proactively identifies potential issues in clinical trial sites through advanced statistical modeling and real-time data analysis. The dashboard provides automated adverse event flagging, protocol deviation detection, and compliance scoring across multiple study sites.

    Key Features:
    • Real-time statistical process control for AE detection
    • Predictive analytics for site performance assessment
    • Automated alert system reducing manual review by 70%
    • Interactive visualizations for regulatory inspection readiness

    Technical Stack: R Shiny (Dashboard), Statistical Modeling, Clinical Data Standards, Real-time Analytics

    Cross-Trial Safety Analytics Dashboard

    Role: Senior Developer | Duration: 5 months | Data Scope: 50+ Clinical Trials

    Project Overview:
    Developed a comprehensive safety surveillance platform that aggregates and analyzes adverse event data across multiple clinical trials to identify safety signals, treatment patterns, and potential drug interactions. The system provides regulatory-grade safety reporting and advanced pharmacovigilance capabilities.

    Safety Impact:
    • Early detection of 15+ safety signals leading to protocol amendments
    • Automated SUSAR (Suspected Unexpected Serious Adverse Reaction) reporting
    • Reduced safety review cycle time from weeks to hours
    • Enhanced patient safety through proactive monitoring algorithms

    Technical Stack: R Shiny (Frontend), Statistical Analysis, Safety Data Mining, Pharmacovigilance Algorithms

    ModelBox – Self-Hosted AI Inference Server

    Role: Software Architect | Duration: 7 months | Performance: Sub-100ms inference

    Project Overview:
    Architected and developed a high-performance, self-hosted AI inference server designed for enterprise deployment. ModelBox provides a secure, scalable platform for serving machine learning models with advanced features including GPU acceleration, load balancing, and comprehensive API management. The system supports multiple model formats and provides real-time inference capabilities for production environments.

    Technical Excellence:
    • Achieved 99.9% uptime with automatic failover and health monitoring
    • Implemented horizontal scaling supporting 10,000+ concurrent requests
    • Optimized GPU utilization achieving 85% efficiency in batch processing
    • Built comprehensive API documentation and SDK for multiple languages

    Technical Architecture: FastAPI (API Framework), Python (Core), Docker (Containerization), NVIDIA GPU (Acceleration), Redis (Caching), Prometheus (Monitoring)

    TL-SDD – Transfer Learning for Surface Defect Detection

    Role: Research Lead | Duration: 10 months | Academic Output: 2 Publications

    Research Overview:
    Led a comprehensive research initiative focused on developing advanced transfer learning methodologies for industrial surface defect detection. Successfully reproduced and significantly extended state-of-the-art few-shot learning models using the GC10-DET dataset, creating a complete end-to-end pipeline for automated quality control in manufacturing environments.

    Research Contributions:
    • Achieved 96.3% defect detection accuracy (15% improvement over baseline)
    • Developed novel data augmentation techniques for limited defect samples
    • Created comprehensive evaluation framework adopted by 3 industrial partners
    • Published methodology in top-tier computer vision conferences

    Technical Stack: PyTorch (Deep Learning), Computer Vision, Transfer Learning, GC10-DET Dataset, Industrial Automation

    Synthetic Oncology Data Explorer (POC)

    Role: Full Stack Developer | Duration: 3 months | Focus: Proof of Concept

    Project Overview:
    Developed an innovative proof-of-concept platform for oncology researchers to explore synthetic tumor burden and biomarker datasets. The application features advanced data generation algorithms that create realistic clinical scenarios while maintaining patient privacy, enabling researchers to test hypotheses and develop analytics workflows without exposure to sensitive patient data.

    Key Features:
    • Generated synthetic datasets covering 15+ oncology biomarkers with realistic correlations
    • Built interactive filtering system enabling multi-dimensional data exploration
    • Implemented modular visualization components for tumor progression analysis
    • Created statistical validation framework ensuring data quality and clinical relevance

    Technical Stack: R Shiny (Frontend), Synthetic Data Generation Algorithms, Oncology Analytics, Biostatistics, Interactive Visualizations

    Research Impact: Enabled 5+ research teams to prototype analytics workflows; methodology adopted for larger oncology data platform development

    R Package Development for Automated Analytics

    Role: Package Architect | Duration: 6 months | Packages Created: 3 production packages

    Project Overview:
    Architected and developed a comprehensive suite of R packages designed to streamline analytics workflows and enhance development productivity. Created 'autotestpkg' for automated unit testing, 'Rmarker' for intelligent documentation generation, and established enterprise-grade CI/CD pipelines for package deployment and maintenance.

    Package Impact:
    • autotestpkg: Reduced testing time by 80% across 20+ R projects
    • Rmarker: Automated generation of 500+ documentation pages with 95% accuracy
    • CI/CD integration: Achieved zero-downtime deployments for all package updates
    • Established coding standards adopted by entire development team

    Technical Architecture: R Package Development, GitLab CI/CD Pipelines, Docker Containerization, Automated Testing Frameworks, Documentation Generation

    Adoption Rate: 100+ downloads per month; packages integrated into 15+ enterprise projects

    CliniTransformR – SDTM/ADaM Programming Suite

    Role: Lead Developer & CDISC Specialist | Duration: 9 months | Compliance: FDA/EMA Standards

    Project Overview:
    Developed a comprehensive R package suite that automates the complex transformation of raw clinical datasets into fully compliant SDTM (Study Data Tabulation Model) and ADaM (Analysis Data Model) formats. The suite includes advanced validation engines, quality control frameworks, and automated compliance checking to ensure regulatory submission readiness.

    Transformation Excellence:
    • Automated conversion of 100+ raw clinical datasets with 99.8% accuracy
    • Built-in validation covering 200+ CDISC compliance rules
    • Reduced manual programming effort from weeks to hours (95% time savings)
    • Generated comprehensive quality control reports with traceability matrices

    Technical Stack: R Programming, CDISC Standards (SDTM/ADaM), Data Validation Frameworks, Clinical Programming, Regulatory Compliance

    Industry Adoption: Deployed at 8 pharmaceutical companies; supported 40+ regulatory submissions with zero compliance rejections

    Reusable Shiny UI Component Library

    Role: Component Architect & UI Designer | Duration: 4 months | Components: 50+ reusable elements

    Project Overview:
    Architected and built a comprehensive library of reusable R Shiny UI components designed to ensure visual consistency, accelerate development timelines, and maintain enterprise-grade design standards across multiple client projects. The library includes sophisticated modal systems, notification frameworks, data grid components, and interactive loading states.

    Development Efficiency:
    • Reduced UI development time by 70% across all new Shiny projects
    • Standardized design language adopted by 25+ applications
    • Built responsive components supporting mobile, tablet, and desktop layouts
    • Implemented theming system allowing brand customization for different clients

    Technical Architecture: R Shiny (Core), Component-Based Architecture, CSS3 (Styling), JavaScript (Interactions), Responsive Design

    Usage Statistics: Library components used in 30+ production applications; 1000+ component instances deployed

    CI/CD Automation for Shiny Deployments

    Role: DevOps Engineer & Automation Architect | Duration: 5 months | Deployments: 100+ automated releases

    Project Overview:
    Established enterprise-grade CI/CD infrastructure specifically optimized for R Shiny application deployments. Built comprehensive automation pipelines covering code quality analysis, automated testing, Docker containerization, security scanning, and multi-environment deployment orchestration with advanced rollback capabilities.

    Automation Excellence:
    • Achieved 99.5% deployment success rate with zero-downtime releases
    • Reduced deployment time from hours to minutes (90% improvement)
    • Implemented comprehensive quality gates preventing 100+ potential issues
    • Built automated rollback system with 30-second recovery times

    Technical Stack: GitLab CI/CD Pipelines, Docker Containerization, R Shiny, DevOps Best Practices, Automated Testing, Infrastructure as Code

    Operational Impact: Enabled daily releases for 15+ applications; reduced manual deployment effort by 85%

    Healthcare Analytics Dashboard

    Role: Full Stack Developer | Duration: 7 months | Data Volume: Real-time processing

    Project Overview:
    Built a sophisticated real-time healthcare analytics platform that integrates live AWS data streaming to provide actionable insights on key performance indicators and operational metrics. The dashboard serves healthcare administrators with comprehensive views of patient flow, resource utilization, and clinical outcomes to enable data-driven decision making.

    Analytics Capabilities:
    • Real-time monitoring of 50+ healthcare KPIs with sub-second latency
    • Predictive analytics for patient admission forecasting (85% accuracy)
    • Interactive visualizations supporting drill-down analysis across departments
    • Automated alert system for critical threshold breaches

    Technical Architecture: R Shiny (Frontend), AWS Services (Data Streaming), Real-time Analytics, Healthcare KPI Modeling, Interactive Visualizations

    Operational Impact: Deployed at 2 healthcare facilities; improved decision-making speed by 60% through real-time insights

    Study Metadata API Integration

    Role: Integration Lead | Duration: 4 months | Platforms Connected: 8 clinical systems

    Project Overview:
    Designed and implemented a sophisticated metadata integration layer that connects disparate clinical data platforms through RESTful APIs, enabling real-time study and variable mapping across the enterprise. The system provides unified metadata access, automated synchronization, and standardized data interoperability for clinical research workflows.

    Integration Excellence:
    • Unified metadata access across 8 disparate clinical data platforms
    • Automated real-time synchronization reducing manual mapping by 90%
    • Built comprehensive error handling and retry mechanisms (99.8% reliability)
    • Implemented caching layer reducing API response times by 75%

    Technical Stack: REST APIs, R Programming, Data Integration Patterns, Metadata Management, Real-time Synchronization

    Efficiency Gains: Eliminated 40+ hours of weekly manual mapping work; standardized metadata access for 200+ variables

    Automated Machine Learning Web App

    Role: Full Stack Developer & ML Engineer | Duration: 8 months | Deployment: Enterprise On-Premise

    Project Overview:
    Developed a comprehensive end-to-end automated machine learning platform that democratizes data science capabilities across the organization. The application provides intuitive interfaces for AutoML pipeline execution, advanced model validation frameworks, automated reporting systems, and enterprise-grade security for sensitive data environments.

    ML Automation Features:
    • Automated feature engineering and model selection across 15+ algorithms
    • Advanced hyperparameter optimization using Bayesian techniques
    • Comprehensive model validation with cross-validation and holdout testing
    • One-click model deployment to production environments

    Technical Architecture: R Shiny (Frontend), AutoML Frameworks, MongoDB (Data Storage), Bootstrap (UI), CI/CD Pipelines, Enterprise Security

    Business Impact: Enabled 50+ non-technical users to build ML models; reduced model development time from weeks to hours

    ScrapText – Web Scraping & Text Mining Tool

    Role: NLP Engineer & Package Developer | Duration: 5 months | Domains Analyzed: 1000+ websites

    Project Overview:
    Created a comprehensive R package suite that combines intelligent web scraping capabilities with advanced text mining and sentiment analysis tools. The platform enables researchers and analysts to extract, process, and analyze large-scale textual data from web sources for market research, competitive intelligence, and social media monitoring applications.

    NLP Capabilities:
    • Intelligent web scraping with adaptive rate limiting and anti-detection
    • Advanced sentiment analysis with 92% accuracy on social media data
    • Topic modeling and keyword extraction for trend identification
    • Interactive visualizations for text analytics and insights presentation

    Technical Stack: R Programming, Web Scraping (rvest), NLP Libraries (tm, tidytext), Text Mining, Sentiment Analysis, Data Visualization

    Market Impact: Used by 10+ market research teams; analyzed 500,000+ documents for competitive intelligence

    Image Prediction REST API

    Role: Backend Developer & API Architect | Duration: 3 months | Throughput: 1000+ requests/minute

    Project Overview:
    Architected and developed a high-performance RESTful API service for serving machine learning image classification models in production environments. The system provides secure, scalable image processing capabilities with enterprise-grade error handling, comprehensive logging, and optimized response times for real-time applications.

    API Performance:
    • Sub-200ms response times for image classification tasks
    • Secure authentication and authorization with JWT tokens
    • Horizontal scaling supporting concurrent processing of 1000+ images
    • Comprehensive error handling with detailed logging and monitoring

    Technical Architecture: R Plumber (API Framework), REST API Design, Image Processing, MongoDB (Data Storage), Scalable Cloud Architecture

    Production Impact: Deployed serving 50,000+ predictions daily; 99.9% uptime with automatic scaling

    Web-Based Reporting Engine with RMarkdown

    Role: Automation Engineer & Report Architect | Duration: 6 months | Reports Generated: 2000+ automated

    Project Overview:
    Developed a sophisticated web-based reporting engine that automates the generation of clinical and operational reports using parameterized R Markdown templates. The system integrates seamlessly with CI/CD workflows to ensure consistent, reproducible, and timely report delivery across multiple departments and regulatory requirements.

    Automation Excellence:
    • Automated generation of 50+ report types with parameterized templates
    • Reduced manual reporting effort by 95% (from days to minutes)
    • Built scheduled reporting system with email distribution
    • Implemented version control and audit trails for regulatory compliance

    Technical Architecture: R Markdown (Templates), Parameterized Reporting, CI/CD Automation, Web Integration, Document Generation

    Business Impact: Serves 15 departments with automated reporting; eliminated 200+ hours of monthly manual work

    R-Python Hybrid Pipelines for Data Processing

    Role: Integration Developer | Duration: 4 months | Pipelines Built: 15+ hybrid workflows

    Project Overview:
    Developed sophisticated integration bridges that enable seamless interoperability between R and Python ecosystems for complex analytics workflows. The hybrid pipelines leverage the statistical strength of R with Python's machine learning capabilities, creating unified data processing systems that optimize both performance and analytical capabilities.

    Integration Features:
    • Bidirectional data exchange with automatic type conversion and validation
    • Optimized memory management for large dataset processing
    • Error handling and logging across language boundaries
    • Performance monitoring and optimization for hybrid workflows

    Technical Stack: R Programming, Python Integration, Data Pipeline Architecture, Statistical Computing, Cross-platform Development

    Efficiency Gains: Reduced processing time by 40% through optimized language selection; enabled reuse of 100+ existing R/Python modules

    Learning Management System

    Role: Full Stack Developer | Duration: 10 months | Users: 500+ learners

    Project Overview:
    Built a comprehensive, secure learning management system designed for enterprise training and certification programs. The platform features advanced course management, automated assessment workflows, detailed progress tracking, and enterprise-grade security measures for sensitive training content and learner data protection.

    LMS Capabilities:
    • Multi-modal content delivery supporting video, interactive modules, and assessments
    • Automated certification workflows with digital badge generation
    • Advanced analytics dashboard for learning progress and engagement tracking
    • Role-based access control with enterprise SSO integration

    Technical Architecture: Python (Backend), Django Framework, PostgreSQL (Database), Enterprise Security, Training Platform Architecture

    Training Impact: Delivered 100+ courses with 95% completion rate; reduced training administration time by 70%

    NLP Chatbot for Client Q&A

    Role: NLP Developer & Conversational AI Engineer | Duration: 6 months | Accuracy: 89% intent recognition

    Project Overview:
    Developed an intelligent conversational AI system powered by advanced neural networks to automate customer support and business FAQ responses. The chatbot features sophisticated natural language understanding, context-aware responses, and seamless escalation pathways to human agents when complex queries are detected.

    AI Capabilities:
    • Neural network-based intent classification with 89% accuracy
    • Context-aware conversation management maintaining dialogue history
    • Multi-language support covering 5 languages for global customer base
    • Intelligent escalation system routing complex queries to appropriate specialists

    Technical Architecture: Python (Core), Neural Networks, Natural Language Processing, Linux Deployment, Customer Support Automation

    Support Impact: Handles 70% of customer inquiries autonomously; reduced response time from hours to seconds

    Contact

    I'm looking for opportunities to work on exciting projects. Let's make something great, together!