Chandini Saisri Uppuganti

CHANDINI SAISRI UPPUGANTI

I'm a

About Me

I'm a graduate student in Software Engineering (Data Science) at San Jose State University with 4+ years of experience as a Data Engineer. I specialize in building scalable data pipelines, deploying ML/LLM-based solutions, and working across cloud ecosystems like AWS and Azure. My work spans legal AI tools, product ranking engines, and Retrieval-Augmented Generation systems—focused on delivering impactful, intelligent solutions.

Education

San Jose State University, San Jose, CA

Master of Science in Computer Software Engineering in Data Science

GPA - 3.8/4

Aug 2024 – May 2026

Sagi RamaKrishnam Raju Engineering College, India

Bachelor of Technology in Computer Science and Engineering

GPA - 3.9/4

June 2016 – Sept 2020

Technical Skills

Languages

Python
SQL
Java
R
PL/SQL
PySpark
Shell

Operating Systems

Linux
Windows

Other Tools

Docker
Kubernetes
Git
Tableau
Power BI
SSIS
IICS
Informatica
Ansible
Jupyter

Projects

AI based Legal Brief Argument Linker

Python, MiniLM, BERT, HuggingFace Transformers, Unsloth

Developed a pairwise classification model using fine-tuned MiniLM and BERT via HuggingFace Transformers and Unsloth to link argument-counterargument sections in legal briefs. Built preprocessing pipelines for JSON datasets and deployed an interactive Streamlit UI to visualize and explore predictions for legal research assistance.

PersonaRank: A Product Scoring Engine

Python, TensorFlow, PyTorch, NLP, Graph Analysis, scikit-learn

Designed a hybrid product scoring system using NLP embeddings (BERT, Word2Vec) and graph-based similarity to personalize search relevance. Built and evaluated ranking models (Logistic Regression, SVM, Gradient Boosting), and optimized preprocessing pipelines to enhance model training by 30% and user engagement by 25%.

LLM-RAG (Retrieval-Augmented Generation)

LLM (OpenAI, Gemini), Python

Pioneered development of a sophisticated Retrieval-Augmented Generation (RAG) model, integrating LLM with external databases, boosting poem retrieval precision by 45% and generating highly meaningful and contextually rich poetic content.

Sales Forecasting and Optimization System

Python, scikit-learn, Pandas, TensorFlow

Developed a Sales Forecasting and Optimization System using ARIMA, SARIMA, and LSTM, improving forecast accuracy by 20% through EDA, trend analysis, and feature engineering (lag variables, moving averages, economic indicators). Optimized models with cross-validation and hyperparameter tuning, reducing RMSE and MAE for precise sales predictions.

Nexus Agent: MCP & A2A Communication System

AWS, Model Context Protocol, Agent-to-Agent Communication, Python

Developed a multi-agent system using Model Context Protocol (MCP) and Agent-to-Agent (A2A) communication patterns. The system leverages AWS services to create a scalable architecture for distributed AI agents, enabling efficient context sharing and collaborative decision-making through innovative communication protocols.

Experience

Senior Data Engineer: Informatica, Hyderabad, India

Jul 2022 – Jul 2024
  • Architected and automated robust ETL pipelines to collect, clean, and validate data from diverse on-premise and cloud databases using Informatica and AWS Glue, boosting processing efficiency by 30% and ensuring high data accuracy
  • Performed in-depth data analysis on petabyte-scale datasets using complex SQL queries to identify trends and patterns. Built and maintained dynamic Tableau dashboards to effectively communicate insights to stakeholders
  • Developed reliable, automated data migration workflows using Informatica PowerCenter and custom Python ETL scripts (version controlled in GitHub), orchestrated with Apache Airflow for scalable scheduling and monitoring, achieving 99% data quality accuracy

Data Engineer: Wipro, Bangalore, India

Jul 2020 – Jul 2022
  • Designed and implemented a 5-zone architecture for migrating data from SAP ECC & SOAP to Azure Data Lake Storage and Synapse Analytics, supporting efficient database design and performance optimization
  • Built and maintained secure, structured ETL pipelines using IICS, improving data integrity by 25% and ensuring seamless integration into analytical databases
  • Implemented automated data quality checks and validation processes to uphold data accuracy and consistency across storage layers, while processing large volumes of big data using Hadoop

Achievements

Extracurricular Activities