A results-driven Data Engineer with over 8 years of hands-on experience specializing in designing and implementing scalable Big Data solutions. Expertise in leveraging cutting-edge technologies like Azure, Spark (Python), and Hadoop to deliver high-performance data pipelines, analytics
Manulife - Data Ingestion, Transformations and enhancements
I was responsible for ingesting data from global financial institutions using Azure Cloud Services, ensuring seamless integration of various financial data sources. My role involved curating this data to meet business requirements, focusing on risk management within the BFSI (Banking, Financial Services, and Insurance) domain. This included transforming raw data into structured, actionable insights to support risk assessments, regulatory compliance, and decision-making processes. The project aimed to streamline the management of financial data, improve data quality, and enhance the accuracy of risk analysis models for business-critical applications.
SunTrust Bank, 04/01/18 - 02/28/21, PySpark, Python, Hive, Shell Scripting, Impala, TWS, Oozie, Created Hive tables and worked on them using Hive QL, which will automatically invoke and run MapReduce jobs in the backend., Managing and scheduling batch Jobs on a Hadoop Cluster using Oozie., Experienced in loading and transforming large sets of structured, semi-structured and unstructured data., Assisted in monitoring Hadoop cluster using Cloudera Manager., Analysed large amounts of data sets to determine optimal way to aggregate and report on it., Involved in daily SCRUM meetings to discuss the development/progress of Sprints and was active in making scrum meetings more productive. Outage Prediction
Bank of America (BOFA), 03/01/16 - 03/31/18, Spark, Hive, Shell Scripting, Impala, TWS, Oozie, Responsible for building distributed data solutions using Hadoop., Analysed data using Hadoop components Hive., Experienced in loading and transforming large sets of structured, semi-structured and unstructured data Hadoop concepts., Created Hive tables to store data and written Hive queries., Involved in importing data from various data sources, performed transformations using Hive and loaded data into HDFS., Extracted the data from Sybase into HDFS using Sqoop.