Skip to content
View Abhishek-Kumar-Rai5's full-sized avatar
  • Cluster Innovation Centre, University of Delhi
  • Delhi
  • 08:45 (UTC +05:30)
  • LinkedIn in/abhik-rai

Highlights

  • Pro

Block or report Abhishek-Kumar-Rai5

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Hi, I'm Abhishek Kumar Rai

I work on backend and data-oriented systems, mostly focused on building pipelines that take raw or unstructured inputs and turn them into usable, reliable outputs.

Over time, I’ve worked on projects involving API-based systems, structured data processing, and deployment pipelines. A lot of my work involves designing how data flows through a system, making sure each stage is predictable, and debugging things when they inevitably break.

More recently, I’ve been spending time improving how these systems handle semi-structured data, thinking more carefully about validation layers, and making deployments more consistent across environments.

I’m particularly interested in how data processing systems, backend services, and machine learning components can be combined into workflows that are simple, maintainable, and production-ready.


Technical Stack

🧠 Languages

Python (backend systems, data pipelines, ML workflows) • SQL (querying, transformation, validation) • JavaScript (API interaction, tooling)

⚙️ Backend & API Engineering

FastAPIFlask

Designing RESTful services with clear request–response contracts • Input validation • Error handling
Middleware integration and request lifecycle management
Structuring backend services for modularity and maintainability

🗂️ Data Engineering & Processing

Data ingestion pipelines for structured and semi-structured inputs
ETL workflows with transformation and validation stages • File processing (CSV, JSON, logs)
Schema handling and structured data transformations
Designing data flows with clear stage separation (ingestion → processing → output)
Logging and monitoring for pipeline visibility and debugging

🤖 Machine Learning Systems

End-to-end workflows (data → training → evaluation → inference)
Designing API-based inference pipelines
Experiment structuring and reproducible training setups
Integration of ML components into backend systems
Handling model inputs/outputs with consistent data interfaces

🚀 Infrastructure & Deployment

Docker • CI/CD Pipelines • Linux

Containerization for environment consistency
Automated testing, validation, and deployment workflows
Deployment strategies focused on reproducibility and stability

🧩 System Design & Reliability

Separation of concerns across data, processing, and serving layers
Designing modular architectures for scalability and maintainability
Error handling, failure recovery, and debugging strategies
Observability through logging and traceability
Ensuring reproducibility across environments and deployments

🛠️ Workflow & Tooling

GitGitHub Actions • Command-line tooling

Version control and collaboration workflows
CI/CD automation and pipeline orchestration
System-level operations and debugging
Structured development workflows for consistent iteration and testing ---

Key Projects

ML Deployment Framework — Containerized inference system

  • Designed a backend service for serving machine learning models via API
  • Implemented reproducible deployment workflows using Docker
  • Structured prediction pipelines for consistent and reliable inference
  • Built with modular components to simplify integration and scaling

Secured CI/CD Pipelines — Automated build and deployment workflow

  • Developed CI/CD pipelines for automated testing, validation, and deployment
  • Integrated security and validation checks into the deployment lifecycle
  • Reduced manual intervention through consistent automation workflows
  • Focused on reliability and repeatable system behavior

Email Classification Pipeline — Structured text processing system

  • Built a modular pipeline for data ingestion, preprocessing, and classification
  • Designed workflows for consistent training and evaluation
  • Ensured reproducibility across data processing stages
  • Structured the system for easy extension and maintenance

MLOps Car Price Pipeline — End-to-end ML system with API serving

  • Developed an end-to-end pipeline from data processing to model serving
  • Exposed real-time prediction functionality via API
  • Integrated logging and monitoring for system visibility
  • Designed for deployment-ready usage with modular workflow components

How I Approach Systems

  • Break problems into clear stages: ingestion → processing → serving
  • Separate data transformation, validation, and delivery layers
  • Design systems to fail gracefully and remain debuggable
  • Prioritize clarity, modularity, and long-term maintainability

Engineering Principles

  • Build systems that are maintainable and observable
  • Prefer simple architectures that scale cleanly
  • Keep components modular and testable
  • Focus on reliability and reproducibility

Currently Working On

  • Improving reliability of backend and data processing systems
  • Building more robust data ingestion and validation workflows
  • Refining deployment and automation practices

Connect

LinkedIn GitHub Email

Popular repositories Loading

  1. Matlab-Fused-Flipbook-Animation Matlab-Fused-Flipbook-Animation Public

    Computational graphics project in MATLAB demonstrating mathematical modeling of frame-by-frame motion (flipbook simulation).

    MATLAB

  2. Email-classification-pipeline Email-classification-pipeline Public

    Modular NLP pipeline for spam detection and sentiment classification using TF-IDF features and classical ML models.

    Jupyter Notebook

  3. ML-deployment-framework ML-deployment-framework Public

    Modular ML deployment framework integrating FastAPI backend with Streamlit frontend for model serving.

    Python

  4. Secured-CI-CD-Pipelines Secured-CI-CD-Pipelines Public

    DevSecOps-based CI/CD pipeline integrating security scanning, automated testing, and secure deployment workflows.

    TypeScript

  5. Mlops-car-price-pipeline Mlops-car-price-pipeline Public

    End-to-end MLOps pipeline for car price prediction with training, monitoring, and FastAPI-based real-time serving.

    Python

  6. dataloom dataloom Public

    Forked from c2siorg/dataloom

    Project is to design and implement a web-based GUI for data wrangling, aimed at simplifying the process of managing and transforming tabular datasets. This application will serve as a graphical int…

    JavaScript