Summary
Overview
Work History
Education
Skills
Prior Work Experience
Training
Skills Profile - Technical
Certification
Timeline
Generic

Vineel Kumar Raju Kanumuri

Senior Manager
Hyderabad

Summary

Around 18+ years of experience in all phases of SDLC including development, production support & maintenance projects. Experience in migrating data from on-prem to GCP. Experience in working with DBT for Automation (Handle large volumes of data and schedule data pipelines), Testing and Monitoring in GCP. Configure DBT to integrate with GCP. Experience in Google Cloud platform using resources like Dataflow, Pub/sub, GCS, Compute engine, GKE, Cloud composer (Airflow DAG’s), Cloud SDK, Data Fusion and Big Query. Experience working with Apache Beam pipelines. Extensive experience in writing Big Query complex queries. Extensive experience in data warehouse design and modeling for Star and Snowflake schemas.

Overview

7
7
years of professional experience
7
7
years of post-secondary education
1
1
Certification

Work History

Senior Manager

Capgemini Technology Services India
Hyderabad
05.2019 - Current
  • Migrate Data from MS Access databases to Big query and generated Looker Dashboards
  • Migrated data from different sources schemas using DBT into big query schema
  • Migrated data from different MS Access DB’s to Big query using Data Flow
  • Worked in creating big query design for multiple layers and transform data using Big query queries
  • Guided team and involved in preparing Data Flow custom based templates using Apache Beam pipelines
  • Involved working with client team on setting-up Dev Ops setup
  • Worked on creating Looker Dashboards for end users
  • Involved in the project from beginning stage of development

Senior Manager

Capgemini Technology Services India
09.2023 - 10.2024
  • Migrate data from different DB’s sources and NAS files into Big Query DBS-R Data Asset
  • Setup Juniper connectivity to extract data from different databases
  • Involved in creating Alteryx workflows to trigger once the data is available in source DB
  • Involved in creating Alteryx workflow once the source system process files in NAS drive
  • Created separate connectivity to trigger files from Juniper to Big Query landing area
  • Once data is available in landing area, we have different layers like, Cleansed, Delta, Enriched and then push into final Data Asset
  • Remove all the duplicates and clean the data based on business requirement and push into final Data Asset, have written queries in Big Query
  • Wrote refresh proc’s in Big Query and scheduled in Alteryx server
  • Created new Data Assets based on business requirements
  • Involved in ML Random Forest model and deployed files into prod and predicted next 3 days prediction report every day to stake holders

Senior Manager – Data Engineer

Capgemini India Private Limited
06.2022 - 08.2023
  • Build Automated infrastructure creating for DDS
  • Worked on violations and resolving them in HSBC in DDS project
  • Created dynamic terraform code to automate infrastructure requests
  • Created Airflow startup script in DDS environment
  • Defining data augmentation pipelines
  • Worked on Data Flow and Data Proc pipelines
  • Created Pub/Sub topics and subscribers for DDS project
  • Involved in maintaining IAM roles in SOT file
  • Worked on GitHub and created Jenkins pipelines for image promotion for Virtual machines
  • Worked on DSW (Data Science Workbench) creating a layer ADL (Analytics Data Layer) in Big Query using terraform
  • Pulled required data from CDL layer and migrated into ADL layer using BQ data transfers services and BQ queries for selected data based on DSW use case
  • Assigned BQ roles to DSW users using terraform

Senior Manager

Capgemini Technology Services India
04.2021 - 05.2022
  • Build Automated Infrastructure & Monitoring tool for I&D practice
  • Engaged in HSBC IHUB Global platform support
  • Leading team of 16 resources working globally
  • Working on violations and resolving them in HSBC in IHUB project
  • Experience on creating dynamic terraform code to automate infrastructure requests
  • Trigged dynamic Terraform to commit into GitHub repo and automatically trigger Jenkins build
  • Defining data augmentation pipelines
  • Worked on defining Pull requests, to authorize and run CI/CD Jenkins pipelines
  • Have created pipeline syntax in Jenkins builds
  • Have created multiple multi-branch pipelines in Jenkins, to normalize CI/CD process
  • Supported ESP (Event Streaming Pipeline) project for DSW
  • ESP team had many DSW use cases, where Data Flow and Airflow DAG’s jobs were used, and with the help of L3 support team we have resolved the job failure issues in production environment
  • DSW portal job failure notifications were monitored in stack driver logs and resolved the issue with business users
  • Based on any new DSW use case created, created Airflow DAGs for data processing and promoted to Prod environment

Manager – Data Engineer

Capgemini India Private Limited
03.2020 - 02.2021
  • Building automation and performance tuning for NOLA pipeline
  • Migrate NOLA Apache beam pipeline to Google Data Flow
  • Understand existing Apache Beam NOLA pipeline architecture
  • Build automation for NOLA pipeline
  • Used Apache Flink to run the beam pipelines in Windows environment
  • Configured checkpoints and save points in Apache Flink to monitor the commit points
  • Experience in Google Protocol Buffers to convert C++ code to python at runtime
  • Used Apache Flink to monitor the process of Backpressure of tasks running in pipeline
  • Defining data augmentation pipelines
  • Worked with dynamic handling of load in Apache Flink
  • Replaced Apache Flink and migrated app data into GKE clusters using CLI and Kubectl commands
  • Created and Managed Cluster type and auto scaling in new cluster using terraform
  • Managed Kubernetes clusters in dashboard to monitor the nodes strength
  • Worked on to resolve VPC connectivity issues in GKE
  • Migrated Apache beam pipeline to Google Data Flow
  • Migrated NOLA dispatch app to Kubernetes containers
  • Experience in ML K-Means algorithm
  • Experience in Krylov Subspace
  • Experience in building Jenkins pipelines

Manager – Data Engineer

Capgemini India Private Limited
05.2019 - 01.2020
  • Build Google data analytics dashboard on Google cloud platform
  • Use predictive analysis on tensor flow platform and build market analysis for future
  • Have experience working with Data Forms on accessibility for the users
  • Understanding business objectives and developing models that help to achieve them, along with metrics to track their progress
  • Analyzing the ML algorithms that could be used to solve a given problem and ranking them by their success probability
  • Exploring and visualizing data to gain an understanding of it, then identifying differences in data distribution that could affect performance when deploying the model in the real world
  • Verifying data quality, and/or ensuring it via data cleaning
  • Defining validation strategies
  • Defining the preprocessing or feature engineering to be done on a given dataset
  • Defining data augmentation pipelines
  • Training models and tuning their hyperparameters
  • Analyzing the errors of the model and designing strategies to overcome them

Data Engineer Architect

Spring ML Private Limited
07.2018 - 01.2019
  • Analyze IMVU Hadoop ecosystems and migrate Hive, Sqoop and Spark jobs to Google cloud platform
  • Created clusters in Data Proc and ran Hive, Sqoop and Spark jobs in Google cloud
  • Importing data from relational databases using Sqoop import into HDFS (Hadoop Distributed File System)
  • Running Sqoop on a Data Proc Hadoop cluster and accessing the built-in Cloud Storage connector which has used the Cloud Storage
  • Loaded the data into Big Query using the Cloud SDK bq command line tool
  • Created workflow template which automatically creates, submits and deletes the clusters
  • Configured 1 master node VM and 2 worker nodes
  • All 3 node VMs are type n1-standard-4 machines
  • Using Standard and legacy SQL, wrote queries on Big Query datasets to generate Analytical data
  • To generate live streaming, used web APIs to extract data and push into Bigtable with columnar families
  • Created Big query dataset on Bigtable and generated specific intervals of periodic data

Technical Architect – Data Analyst

ATMECS Technologies Private Limited
01.2018 - 06.2018
  • Analyze Google’s insights data and provide them visualization with-in cloud platform, which gives Google to analyze the Google Cloud Products which are performing better and also analyze the metrics of customer raised outages VS Google’s internal monitoring system
  • Worked onsite in Google Sunnyvale office, Google Cloud Platform (GCP) extension to Mountain View headquarters, California
  • Data analyses on different Google’s internal systems, like OMG, Post-mortem Data, customer data and Buganizer data
  • Analyzed data and came up with analyses on different systems to compare the metrics between Google’s detection system VS Customer raised outages
  • Prepared dashboards to visualize in Google’s internal PLX tool to analyze data including trends
  • Developed Map Reduce Programs those are running on the cluster
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Loading log data directly into HDFS using Flume
  • Experienced in managing and reviewing Hadoop log files
  • Analyzing data with Hive, Pig and Hadoop Streaming for Customer centric analysis dashboard
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig
  • Developed Map Reduce Programs for data analysis and data cleaning
  • Developed PIG Latin scripts for the analysis of semi structured data
  • Used Hive and created Hive tables and involved in data loading and writing Hive UDFs

Education

Bachelor of Engineering - Mechanical

University of Madras
Chennai, Tamil Nadu
09.1998 - 09.2002

Intermediate - Mathematics, Physics, Chemistry (MPC)

Board of Secondary Education
Hyderabad, Telangana
06.1995 - 09.1997

10th Standard - Mathematics, Physics, Chemistry

Central Board of Secondary Education
New Delhi, Delhi
03.1994 - 03.1995

Skills

Google clod Composer(Airflow, GKE, GCS, GCE, Data Flow, Big Query(SQL, Data form, DBT), Monitoring and Alerting Policies PUB/SUB, Data Fusion, Data Proc, Spanner))

Prior Work Experience

  • Eliptico IT Solutions, Technical Architect, 31/10/16, 30/11/17, Hyderabad, Telangana
  • Virtusa Consulting Services Private Limited, Associate Architect, 22/09/14, 14/09/16, Hyderabad, Telangana
  • Alliance Global Services Private Limited, Associate Architect, 25/11/13, 02/07/14, Hyderabad, Telangana
  • Value Momentum Software Services Private Limited, Technical Lead, 26/10/09, 18/10/13, Hyderabad, Telangana
  • Danlaw Technologies India Limited, Database Administrator, 05/06/06, 30/09/09, Hyderabad, Telangana

Training

  • Red Hat Enterprise Linux AS4.0 training, SQL Star International, Hyderabad, Telangana, 03/06
  • Oracle 9i DBA training, SQL Star International, Hyderabad, Telangana, 06/06
  • Google Spanner training, Google Inc., Sunnyvale, California, 02/18
  • Terraform course, Coursera, 06/23
  • Google Cloud Certified Professional Cloud Architect, 12/21

Skills Profile - Technical

  • Airflow, GKE, Bigtable, Stack Driver Monitoring, Google Data Flow, GCS, GKE, GCE, Terraform, Big Query, SQL, Dynamic SQL, PL/SQL, Google SQL, DREMEL SOL, Google PLX, Python3.7, Oracle 9i/10g, SQL Server 2008R2, 2012, HBase, ERWIN 7.1, Physical & Logical Data Modelling, De-Normalization techniques, Red Hat Enterprise Linux 4.1 AS, CentOS 6.3, UNIX, Change Data Capture (CDC), Slowly Changing Dimensions (SCD), Hive, Star Schema, Snowflake, DBT, Tableau 8.3, Google PLX, Looker, Data Fusion – Google ETL, MSBI 2005/2008/2012, Data Stage 7.5.1, Informatica, DBT, Oracle Enterprise manager, SQL
  • LOADER, Export/Import, Dream Coder, Business Intelligence Development Studio (BIDS), GitHub – Web hooks, Bit Bucket, Jenkins, Random Forest, Python 3.6, K-means, Krylov subspace

Certification

Google Cloud Certified Professional Architect

Timeline

Senior Manager

Capgemini Technology Services India
09.2023 - 10.2024

Senior Manager – Data Engineer

Capgemini India Private Limited
06.2022 - 08.2023

Google Cloud Certified Professional Architect

12-2021

Senior Manager

Capgemini Technology Services India
04.2021 - 05.2022

Manager – Data Engineer

Capgemini India Private Limited
03.2020 - 02.2021

Senior Manager

Capgemini Technology Services India
05.2019 - Current

Manager – Data Engineer

Capgemini India Private Limited
05.2019 - 01.2020

Data Engineer Architect

Spring ML Private Limited
07.2018 - 01.2019

Technical Architect – Data Analyst

ATMECS Technologies Private Limited
01.2018 - 06.2018

Bachelor of Engineering - Mechanical

University of Madras
09.1998 - 09.2002

Intermediate - Mathematics, Physics, Chemistry (MPC)

Board of Secondary Education
06.1995 - 09.1997

10th Standard - Mathematics, Physics, Chemistry

Central Board of Secondary Education
03.1994 - 03.1995
Vineel Kumar Raju KanumuriSenior Manager