University of Texas at Dallas
Information Technology and Management (ITM)
A M.S IT graduate from the University of Texas at Dallas
I'm a software engineer specialized in data domain with over four+ years of experience, working with global companies like Infosys, Amdocs, and Briston Infotech.
Currently, I'm actively looking for an opportunity in data- driven domains such as data engineering, data or solution architect, ETL/DWH developement, and ETL Testing.
Information Technology and Management (ITM)
Computer Science and Engineering (CSE)
Senior Secondary School (+2)
I'm an innovative, independent, and deadline-driven Data Engineer and Architect with over four years of hands-on experience designing, developing, and testing user-centered enterprise data warehouse solutions.
I've worked across diverse domains, including telecommunications, health insurance, and retail, bringing expertise in data engineering, solution architecture, etl processing, data pipeline streamlining, and data warehousing to every project.
Data Engineering, Data Architect, Solution, Architect, Data Analyst, Data Science, Big Data, ETL Development, ETL Testing, Business Analyst, SDE
AWS Cloud Solution Architecture, Big Data, Business Data Warehousing, Advance Statistics for Data Science, Business Analytics with R, Database Foundation for Business Analytics, Predictive Analytics for Data Science, and Prescriptive Analytics.
A comprehensive solution that streamlines vehicle damage assessment using AI technology. This innovative system allows users to upload images of damaged vehicles and receive instant analysis of damage severity and repair options through a portable device.
A comprehensive platform for generating and managing digital license keys with seamless API integration. This solution simplifies the software licensing process for developers and businesses, ensuring secure key distribution and validation.
A robust web-based application designed to validate and test data across various sources including flat files, relational databases, and APIs. ETLQC ensures data accuracy and reliability throughout ETL/ELT processes, making it an essential tool for data engineers and quality assurance teams.
A comprehensive analysis of car auction data encompassing both electric and non-electric vehicles from various manufacturers. This project leverages Python and Object-Oriented Programming principles to extract valuable insights from automotive market data.
A collection of end-to-end ETL/Data Engineering solutions implemented using Microsoft Azure services. This repository showcases expertise in cloud-based data processing, from minor tasks to full-scale applications, demonstrating versatile skills in modern data architecture.
A comprehensive educational repository featuring a carefully curated series of SQL exercises designed to take users from beginner to advanced level. Each exercise includes detailed schemas, challenging questions, and thoroughly explained solutions to build strong SQL fundamentals.
Stay tuned for insightful articles and tutorials on data engineering, cloud technologies, and more!
Feel free to reach out for opportunities, collaborations, or just to say hello!
Here are some common questions about my background, skills, and how we can work together.
My expertise centers around building scalable data pipelines, data warehousing solutions, and ETL/ELT processes. I specialize in cloud platforms like Azure and AWS, with strong skills in Python, SQL, Spark, and modern data tools like Databricks, ADF, and Airflow. I'm particularly strong in designing data architectures that balance performance, cost, and maintainability.
I believe data quality is the foundation of any successful data initiative. My approach includes implementing robust validation rules, automated testing pipelines, and monitoring systems to catch issues early. For governance, I work to establish clear data ownership, lineage tracking, and documentation practices. I've developed custom data quality frameworks that integrate with ETL processes to ensure consistency and reliability throughout the data lifecycle.
I've designed and implemented several real-time data processing systems using technologies like Apache Kafka, Azure Event Hubs, and Stream Analytics. One notable project involved creating a real-time customer analytics platform that processed millions of events per hour with sub-second latency. This system used a combination of stream processing for immediate insights and batch processing for historical analysis, providing business users with both real-time dashboards and comprehensive reporting capabilities.
Large-scale data migrations require careful planning and execution. My approach involves thorough source system analysis, detailed mapping documentation, and creating a robust testing strategy before any migration begins. I typically implement the migration in phases, starting with a proof of concept followed by incremental migrations when possible. Throughout the process, I use automated validation to verify data integrity and completeness. I also design fallback mechanisms and maintain parallel systems during the transition period to minimize business disruption.
I welcome guest contributions on topics related to data engineering, cloud technologies, analytics, or software development! To contribute, simply reach out through the contact form with the subject "Blog Contribution" and include a brief outline of your proposed topic. The ideal length is 1000-2000 words, and I encourage practical, hands-on content that provides value to readers. You'll receive full attribution for your work, and it's a great way to share your knowledge with the community while gaining exposure for your expertise.
The blog welcomes a wide range of technical topics, including but not limited to:
The most valuable contributions share practical insights, provide code examples when relevant, and offer actionable takeaways for readers.
I particularly enjoy working with Azure Databricks, Python, and modern data pipeline orchestration tools like Apache Airflow. I find Databricks especially powerful for its unified analytics platform that combines the best of data engineering and data science capabilities. On the AWS side, I'm excited about the capabilities of services like Glue, Redshift, and Step Functions for building serverless data workflows. I'm also increasingly interested in the intersection of data engineering with MLOps, and how we can build better pipelines to support model training and deployment.
Yes, I'm selectively available for freelance consulting on data engineering projects, particularly those involving complex data architectures, performance optimization, or cloud migrations. I can provide services ranging from architecture review and technical guidance to hands-on implementation and team mentoring. If you have a project in mind, please reach out through the contact form with details about your needs and timeline, and we can discuss how I might be able to help.