Summary
Overview
Work History
Education
Skills
Websites
Medium
Industryexpertise
Languages
Training
Timeline
Generic
GIRIJA SANKAR PANDA

GIRIJA SANKAR PANDA

Software Architect
Hyderabad

Summary

Experienced Solution Architect with 13 years of experience designing and implementing scalable data architectures across multi-cloud environments, including AWS and GCP. Demonstrated expertise in building and optimizing ETL pipelines, data lakes, and real-time streaming solutions using Python, PySpark, and Terraform. Proficient in leveraging cloud-native services such as AWS S3, Athena, EMR, Snowflake, BigQuery, and Google Cloud Storage to deliver efficient and cost-effective solutions. Adept at driving performance optimization, automation, and end-to-end data integration across large-scale systems through strong data engineering skills. Skilled in Kubernetes, Docker, and microservices architecture to ensure highly available, secure, and resilient infrastructure. Extensive experience in implementing Industry 4.0 initiatives in the manufacturing sector.

Overview

13
13
years of professional experience

Work History

Solution Architect (Data Engineering/ML) GCP

Micron Technology SMAI
10.2022 - Current
  • Led a cross-functional team in architecting and implementing real-time data processing solutions using Google Cloud Pub/Sub and MQTT protocols, resulting in a 40% reduction in downtime on the production line
  • Designed and optimized data pipelines using Google Cloud DataProc and BigQuery, enabling efficient analysis of large datasets and reducing query times by 50%
  • Implemented machine learning models utilizing YOLO for defect detection and quality control, leading to a 30% improvement in product quality
  • Improved solution efficiency by designing and implementing scalable architectures for enterprise clients.
  • Conducted research to evaluate systems design and process efficiency.
  • Worked with customers or prospective customers to develop integrated solutions and lead detailed architectural dialogues to facilitate delivery of comprehensive solution.
  • Facilitated seamless communication between technical teams and non-technical stakeholders by effectively translating complex concepts into understandable terms for all parties involved in projects.
  • Worked closely with product teams to define and prioritize partner feature requests
  • Optimized resource allocation across multiple projects by utilizing advanced project management tools and techniques for more efficient scheduling and task prioritization.
  • Collaborated with developers on creating user-friendly interfaces that simplified complex workflows within custom-built solutions.
  • Created and implemented innovative business solutions to support corporate objectives.
  • Accelerated project delivery with effective management of cross-functional teams, adhering to agile methodologies and best practices.
  • Managed project scope, schedule, status and documentation.
  • Collaborated with manufacturing engineers and data scientists to identify key business challenges and develop tailored machine learning solutions to address them
  • Conducted regular training sessions for the engineering team to enhance their understanding of data engineering and machine learning concepts
  • Implemented DDR4 and DDR5 memory technologies in server configurations, enhancing data processing capabilities and supporting large-scale manufacturing data analytics
  • Conducted performance optimization and troubleshooting for RAM, DRAM, and SSD-based systems, resolving complex technical issues to minimize production downtime

Data Engineering Lead /AWS Solutions Architect

EPAM Systems, Inc : Novartis
08.2021 - 08.2022
  • Improved AWS infrastructure efficiency by designing and implementing scalable solutions using Terraform to automate infrastructure deployment, resulting in a 25% reduction in setup time and fewer human errors.
  • Engineered and maintained ETL pipelines using PySpark, processing large-scale data from various sources and loading it into AWS S3 and Snowflake for analytical processing, achieving 99.9% data availability. Integrated AWS Athena to query data stored in S3, further optimizing data access for ad-hoc analysis.
  • Implemented real-time data streaming solutions using Apache Kafka and Spark Streaming, enabling timely analysis and visualization of streaming data for operational insights, leveraging AWS EMR to scale processing capabilities efficiently.
  • Collaborated with AWS architects to optimize storage and compute resources on S3 and EMR clusters, ensuring cost-effective and scalable data processing capabilities, which led to a 20% reduction in overall operational costs.
  • Led training sessions for junior data engineers, providing mentorship on best practices, coding standards, and the effective use of Databricks, PySpark, and cloud services such as Snowflake and AWS Athena.
  • Implemented data governance and quality checks, establishing best practices to maintain high-quality data within the data lake architecture, leading to a 25% improvement in data accuracy and compliance with industry regulations.
  • Conducted performance tuning and optimization of Spark jobs on AWS EMR, reducing processing time by 30% and lowering operational costs. Utilized AWS Athena for cost-effective querying and analysis of large datasets stored in S3.
  • Designed and deployed cloud-based infrastructure, using Terraform for provisioning, reducing overall costs while increasing performance and scalability. Streamlined data lake and Snowflake integration for more efficient data processing pipelines.
  • Introduced agile methodologies to improve project management efficiency within the team, resulting in faster project delivery cycles and improved collaboration across teams.
  • Developed scalable and maintainable data ingestion frameworks, leveraging AWS services such as S3, EMR, and Snowflake, resulting in streamlined data integration processes and increased data processing efficiency.
  • Automated repetitive tasks using Python scripts, saving time and resources for higher-priority projects, particularly around Snowflake and AWS infrastructure management.


Lead Software Engineer

Accenture Zurich Insurance
11.2017 - 08.2021
  • Led the development of end-to-end data integration solutions using Azure Data Factory, ensuring efficient data movement and transformation across both cloud and on-premises environments.
  • Designed and implemented data pipelines for diverse data sources, enabling seamless data flow and real-time analytics for business intelligence purposes.
  • Exposed database tables via FastAPI, creating RESTful endpoints to serve data for various consumers and applications. This enabled secure and efficient access to structured data, facilitating real-time analytics and decision-making.
  • Containerized the FastAPI application using Docker to ensure portability and consistency across development and production environments.
  • Deployed the containerized FastAPI application to Kubernetes, automating the deployment, scaling, and management of the microservices, leading to increased reliability, high availability, and simplified infrastructure management.
  • Collaborated with cross-functional teams to gather requirements, design scalable ETL architectures, and optimize data workflows, resulting in a 30% improvement in data processing efficiency.
  • Performed performance tuning and optimization of Azure Data Factory pipelines and Kubernetes-managed services, reducing processing time and improving overall system reliability.

Software Engineer

Accenture: BNP Paribas
11.2014 - 11.2017
  • transformation.
  • Wrote and automated shell scripts to streamline and monitor the movement of data from Hadoop Distributed File System (HDFS) to PostgreSQL, reducing manual intervention by 40%.
  • Used PostgreSQL to store and manage relational datasets for downstream applications, ensuring data integrity and consistency.
  • Managed version control for all data engineering projects using Git, ensuring collaborative development and proper versioning of codebase.
  • Conducted performance tuning and query optimization in Hive and PostgreSQL, reducing query execution times by 25%.
  • Integrated Sqoop for data transfer between PostgreSQL and Hadoop, improving data ingestion processes and ensuring real-time availability for analytics.
  • Collaborated with data analysts and business intelligence teams to provide clean, transformed data for insights and decision-making.
  • Automated routine tasks and data validation processes using shell scripting, improving overall operational efficiency.

Programmer Analyst

Cognizant Technologies :FDMS
12.2011 - 11.2014
  • Developed and maintained COBOL programs integrated with DB2 for high-volume transaction processing in the FastData Merchant Service.
  • Created and optimized shell scripts to automate repetitive tasks, enhancing system efficiency and reducing manual intervention.
  • Designed and managed DB2 databases, performing data extraction, transformation, and loading (ETL) processes to ensure accurate and timely data availability.
  • Collaborated with development teams to support merchant service applications, ensuring seamless integration with mainframe systems.
  • Conducted performance tuning and query optimization for DB2 to minimize response times and maximize system throughput.
  • Led database migration projects to upgrade and maintain legacy systems, ensuring data integrity and system stability throughout transitions.
  • Wrote and tested JCL scripts for batch processing, ensuring efficient execution of data processing jobs across the mainframe environment.
  • Provided production support, troubleshooting system issues, and implementing fixes to minimize downtime and improve system reliability.
  • Documented processes, configurations, and system changes for operational and audit purposes, ensuring compliance with internal and regulatory standards.

Education

Bachelor of Information Technology & Computer - Information Technology & Computer

College of Engineering
Bhubaneswar
08.2007 -

Skills

GCP Solution Architect/Dataproc/pubsub/Dataflow

Cloud Function/GCS/GKE/Big query

Apache Kafka/Databricks

AWS Solution Architect/AWS IAM/Glue/S3/Lamda/Step Fucntion

Python/Scala/SQL

Apache NiFi

Snowflake

Kubernetes/Docker

API Design/Serverless Architecture

Systems design

Big Data Solutions

Enterprise Architecture Design

Medium

https://medium.com/@girijasankarpanda

Industryexpertise

  • Manufacturing - I4.0
  • HealthCare
  • Banking and Financial Service
  • Card and Finance
  • Insurance

Languages

English (Native)
Bilingual or Proficient (C2)
Hindi (Intermediate)
Bilingual or Proficient (C2)
French (Intermediate)
Intermediate (B1)

Training

AWS Solution Certified 

GCP Trained

LLM/RAG

Timeline

Solution Architect (Data Engineering/ML) GCP

Micron Technology SMAI
10.2022 - Current

Data Engineering Lead /AWS Solutions Architect

EPAM Systems, Inc : Novartis
08.2021 - 08.2022

Lead Software Engineer

Accenture Zurich Insurance
11.2017 - 08.2021

Software Engineer

Accenture: BNP Paribas
11.2014 - 11.2017

Programmer Analyst

Cognizant Technologies :FDMS
12.2011 - 11.2014

Bachelor of Information Technology & Computer - Information Technology & Computer

College of Engineering
08.2007 -
GIRIJA SANKAR PANDASoftware Architect