Summary
Overview
Work History
Education
Skills
Timeline
Generic

Priti Mogili

Toronto,Canada

Summary

Having 5+ Years of Experience as a Cloud Data Engineer, Big Data Engineer, Hadoop Developer and Python/ETL Developer who undertakes complex assignments, meets tight deadlines, and delivers superior performance, practical knowledge in data Analytics and Optimization Applies string analytical skills to inform senior management of key trends identified in the data. Excellent domain knowledge on career portals and automobile industries, generating Power BI reports for clients in finance, capital markets and insurance while serving as an SME. Hands on experience in Amazon Web Services (AWS) provisioning and good knowledge of AWS services like EC2, S3, Bamboo, Glacier, ELB (Load Balancers), SQS, Autoscaling, RedShift, LakeFormation, Glue, Athena, Cloud Formation, KMS, Cloud Watch, CDK, SNS, EMR, EBS, RDS.

Overview

5
5
years of professional experience

Work History

Azure Cloud Data Engineer / SME

AutoNation
Toronto
11.2022 - Current
  • Experience as a Big Data Engineer with expertise in designing data-intensive applications using Hadoop
  • Integrated in Infrastructure Development and Operations involving Azure Cloud platforms, Firewall setup, Blob Storage, Resource Groups, Network etc…
  • Hands on experience with Data Engineering within Azure cloud using the services (Azure Data Factory, Azure data bricks, Azure SQL Database, AZURE Data Warehouse, Azure Data Lake Storage, Azure Blob Storage, Cosmos DB)
  • Create Notebooks using Databricks, Python, Scala and spark and capturing the data from Delta tables in Delta lakes
  • Worked in data migration from on premise data bases (MySQL, SQL Server and Oracle) into Azure cloud
  • Worked on loading the data into noSQL databases such as MongoDB
  • Created Azure Data Factory and managing policies for Data Factory and Utilized Blob storage for storage and backup on Azure
  • Extensive knowledge in migrating applications from internal data storage to Azure
  • Experience in building streaming applications in Azure Notebooks using Kafka and Spark
  • Hands on experience with Creation and Designing of Power BI data visualization utilizing cross tabs, maps scatterplots, pie, bar and density charts
  • Hands on experience with writing complex SQL Queries (Stored Procedures) and improving the performance of the SQL Queries
  • Used the various Complex data structures in collections Framework in python for building the Business Logic
  • Woked on the SQL databases and data-modeling techniques
  • Hands on experience with data warehouse concepts involved in implementing star schema, snowflake schema and Dimensional modeling, OLAP and OLTP concepts
  • Hands on experience with data modeling techniques and data warehouse concepts
  • Built and configured a virtual data center in the Azure cloud to support Enterprise Data Warehouse hosting including Virtual Private Cloud (VPC), Public and Private Subnets, Security Groups, Route Tables
  • Design and maintain well-structured relational database Schemes
  • Create automated ETL pipelines for a variety of raw manufacturing data sources
  • Research and develop new data storage architectures for growing data volumes
  • Expertise in Creating, Debugging, Scheduling and Monitoring jobs using Airflow for ETL batch processing to load into Snowflake for analytical processes
  • Worked on SnowSQL and Snowpipe
  • Converted Azure jobs to support snowflake functionality
  • Evaluate Snowflake design considerations for any change in application
  • Hands on experience with Creation and Designing of Power BI data visualization utilizing cross tabs, maps scatterplots, pie, bar and density charts
  • Experience in Developing Spark applications using Spark - SQL or data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns
  • Developed Python based API (RESTful Web Service) to track sales and perform sales analysis using Flask, SQLAlchemy and PostgreSQL
  • Developed and implemented ETL pipelines using Python, SQL, Spark and PySpark to ingest data and updates to relevant databases

AWS Cloud Data Engineer

Talentnet
Toronto
09.2021 - 09.2022
  • Implemented Data pipeline using AWS Cloud platform and data services like Lake Formation, Glue, Athena, EMR, Glue, RDS, Data lake, Redshift, Cloud Watch, SQS, SNS, IAM, S3, Cloud Formation
  • Worked on AWS Lambda to run the code in response to events, such as changes to data in an Amazon S3 bucket, Amazon DynamoDB table, HTTP requests using AWS API Gateway and invoked the code using API calls made using AWS SDKs
  • Worked on distributed frameworks such as Apache Spark and Presto in Amazon EMR, Redshift and interact with data in other AWS data stores such as Amazon S3 and Amazon DynamoDB
  • Experience in automatic deploying of applications with Amazon Lambda and Elastic Beanstalk
  • Used the various Complex data structures in collections Framework in python for building the Business Logic
  • Developed full backend system in Python and Django, Web Socket to deliver data to the web applications for dynamic pages
  • Built REST API using Django web framework to send data to the user interface in JSON format
  • Wrote Python routines to log into the websites and fetch data for selected options
  • Good Understanding of the Database Schema, Data Dictionary, and Data Models for testing the data in an ETL process
  • Worked on the SQL databases and data-modeling techniques
  • Connected PostgreSQL database to Django's database-abstraction API to query data and update the database
  • Created data access using SQL and PL/SQL stored procedures
  • Responsible for loading candidate’s data and event logs into HBase using python API
  • Created HBase tables to store variable data formats of input data coming from different portfolios
  • Involved in adding huge volumes of data in rows and columns to store data in HBase
  • Used Flume to collect the log data from different resources and transfer the data type to Hive tables using different SerDes to store in JSON, cl and Sequence file formats
  • Used Hive to find correlations between candidate’s browser logs in different sites and analyzed them
  • Created User accounts and given the users the access to the Hadoop Cluster
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
  • Developed the Pig UDF's to pre-process the data for analysis
  • Loaded files to Hive and HDFS from MongoDB Solr
  • Monitored Hadoop cluster job performance and performed capacity planning and managed nodes on Hadoop cluster
  • Responsible for using Oozie to control workflow

Data Engineer/Python Developer

MKRK
Hyderabad, INDIA
10.2019 - 09.2021
  • Generated ad-hoc, parameterized, drill down, and drill through reports utilizing SSRS
  • Written expressions to hide sensitive data at report level in SSRS
  • Developed reports in SSRS using different types of properties like chart controls, Filters, Interactive Sorting, SQL parameters etc
  • Developed medium to complex reports like detailed list report, Summary reports, Matrix reports, drill through reports, Sub-reports, ad-hoc reports to meet the business needs using SSRS
  • Extensively worked in designing various Data Visualization Dashboards, gathering customer requirements, writing test cases, preparing analytical reports to help users to identify critical KPI’s
  • Created cache refresh plans for frequently used reports in Report Manager
  • Worked extensively on T-SQL like creating table, view, Index, stored procedure, function, trigger etc
  • Created SSIS Packages using Lookup, Derived Columns, Condition Split, Term extraction, Aggregate, Execute SQL Task, Data Flow Task, Execute Package Task, lookup etc
  • To generate underlying data for the reports and to export cleaned data from Excel, Text file and CSV files to data warehouse
  • Created customized reports using Power BI
  • Extensively designed the packages and data mapping using Control flow task, Sequence container task, Dataflow Task, Execute SQL Task, Data conversion task, Derived Column task and Script Task in SSIS Designer
  • Used SQL Server Package Configuration technique to make Dynamic configurations
  • Used Text Files/ SQL Server Logging for all SSIS packages
  • Developed joins and subqueries for complex queries involving multiple tables
  • Developed triggers, stored procedures, and data manipulation using DDL and DML
  • Created stored procedures, views, tables and Index to generate SSRS Reports
  • Involved in System Design, Coding, creating table, views, SP’s, SSIS package, managed performance of SQL queries, developed complex reports as per client requirement, load the data from MS Excel to SQL Server Table and SQL Server table to MS Excel
  • Analyzing requirement and provide business / technical solution to client
  • Handled Performance Tuning and Optimization on SSIS, with strong analytical & troubleshooting skills for quick issue resolution
  • Utilized ETL process to Extract, Transform, Joiner update strategy and creating mappings/workflows to extract data from SQL Server and Flat File sources, Legacy systems and load into various Business Entities

Education

Bachelor of Engineering -

Jawaharlal Nehru Technological University
India
01.2019

Skills

  • Cloud Technologies: AWS (EC2, S3, Glue, Athena, EMR, Redshift, RDS, SNS, SQS, Lambda, DynamoDB, API Gateway, CloudFormation, CloudWatch, Lake Formation, ELB, Auto Scaling, KMS, Route 53, CDK), Azure (Data Factory, Databricks, SQL, Cosmos DB, Data Lake, Blob Storage, AD, Virtual Machines, Stream Analytics, Fabric, VPN Gateway, Web Jobs, Time Series Insights)
  • Big Data & Data Engineering: Hadoop, HDFS, MapReduce, Hive, Pig, Sqoop, Spark, Kafka, HBase, Oozie, Flume, NiFi
  • Databases & ETL: MongoDB, Cassandra, MySQL, PostgreSQL, SQL Server, Oracle, SSIS, Apache Airflow, Pentaho, Data Warehousing, Star Schema, Snowflake Schema, OLAP, OLTP, Dimensional Modeling
  • Programming & Scripting: Python (Flask, SQLAlchemy, Pandas, NumPy), Scala, Java, Linux/Unix Shell Scripting
  • DevOps & CI/CD: GitHub, Azure DevOps
  • Data Visualization & Reporting: Power BI, SSRS, Tableau
  • Other Skills: Data Modeling, Business Intelligence, Agile, Scrum, Performance Tuning, SQL Query Optimization

Timeline

Azure Cloud Data Engineer / SME

AutoNation
11.2022 - Current

AWS Cloud Data Engineer

Talentnet
09.2021 - 09.2022

Data Engineer/Python Developer

MKRK
10.2019 - 09.2021

Bachelor of Engineering -

Jawaharlal Nehru Technological University
Priti Mogili