Summary
Overview
Work History
Education
Skills
Timeline
Generic

DHARA NAROLA

Toronto,Canada

Summary

Over 7 years of experience in Big Data technologies including Hadoop, Spark, and Kafka, with a strong focus on system requirements analysis, design, and development. Proficient in processing large datasets across structured, semi-structured utilizing tools such as Spark and Scala for data management. Demonstrated expertise in Spark architecture and the Hadoop ecosystem, with hands-on experience in application development using Java and Scala. Strong understanding of SDLC phases and Agile methodologies, complemented by effective communication and teamwork skills. Innovative Big Data Developer known for high productivity and efficiency in task completion. Skilled in Hadoop ecosystems, Spark, and NoSQL databases which enable effective handling and analysis of large data sets. Excel in problem-solving, teamwork, and communication, leveraging these soft skills to collaborate effectively on projects and meet deadlines with precision.

Overview

11
11
years of professional experience

Work History

BigData Developer

RBC
Toronto, Canada
01.2021 - Current
  • Working in big team of AML Client screening – where we have one list of world check – All the clients will be scanned against that check file and relative sanctions, downstream application will consume the data and create alert to investigate further.
  • Analyzing mapping document received from business
  • Creating component for development
  • Strategy for git repo decoupled from other sources as component is shared across
  • Write complex SQL dataframes to make data more detailed for end user (as per BU ask)
  • Keep JIRA story up to date depends on development
  • Preparing the Implementation documents
  • Supporting the sandbox / prod after hour deployment
  • PIV after every deployment
  • PS team support and maintenance whenever needed
  • Project 2: Migration of AML Legacy Fintrac Reporting Application - When there are reasonable grounds to suspect that a transaction or an attempted transaction is related to the commission or attempted commission of a money laundering (ML) or terrorist financing (TF) offence, we must submit a suspicious transaction report (STR). This is a migration project of existing legacy application. The purpose of this project is to report to Fintrac all transaction which is exceeded from 10K Canadian amount.
  • Involved in AML IT Swift, LCTR program
  • Analyze large data sets in conjunction with data dictionaries, system documentation, and end user documentation to understand business context and translate into data requirements from an enterprise AML Program perspective
  • Perform data profiling and analysis of data sources to assess data availability and ensure AML data requirements are continually satisfied
  • Conducting daily standup meeting for tracking the progress of assigned task
  • Creating batch job by using internal big data framework (HINT) own by RBC
  • Using SparkSQL, Dataframe and dataset to apply business rule
  • Worked on Hadoop eco system
  • Comparing outputs new application with legacy application for equality
  • Troubleshooting the problem if we have any in lower environment
  • Creating supporting documents for PS team
  • GitHub is used for version control
  • Creating shell script for flexibility of running the jobs
  • Using iterative waterfall model for development of this new application and Jira is used for tracking tool

BigData/Spark Developer

National Bank of Canada
Toronto, Canada
01.2020 - 11.2020
  • Project: SDD Reconciliation - Design and implement Strategic Data Repository for Datalake and generating report
  • Reconciling the payment and outstanding balances between two payment processing systems on account level.
  • Creating the Spark Scala application for reading processing and storing into Hive.
  • Using SparkSQL, Dataframe and dataset for this application.
  • Creating control table for logging the status of an application.
  • Bitbucket is used for version control.
  • Creating the CI/CD pipeline using Teamcity and uDeploy.
  • Autosys is used for scheduling the job for automatic deployment.
  • Creating shell script for running the job by checking files have been loaded.
  • Agile methodology is general practice and Jira is used for tracking tool.
  • Project: The purpose of this project is to capture all data streams from different sources into our cloud stack based on technologies including Hadoop, Spark and Kafka. Also, we built new processing pipelines over transaction records, user profiles, files, and communication data ranging from emails, instant messages, social media feeds. Moreover, using Spark to enrich and transform data to internal data models powering search, data visualization and analytics.
  • Designed and implemented scalable infrastructure and platform for large amounts of data ingestion, aggregation, integration and analytics in Hadoop, including MapReduce, Spark, Hive, Pig and HBase
  • Written scripts to import, export and update the data between HDFS/hive and Relational databases
  • Created Flume configure file to collect, aggregate and store the web log data into HDFS.
  • Developed Pig Latin scripts to transform log data files and load into HDFS
  • Import data from difference sources like HDFS/ HBase into Spark RDD
  • Involved in converting Hive queries into Spark transformations using Spark RDD and Scala/Python
  • Created tables, then loaded and analyzed data using hive queries
  • Utilized Kafka to capture and process real time and near-real time streaming data

Bigdata Developer

Zenith
Toronto, Canada
05.2019 - 08.2019
  • Project: The purpose of this project is to provide data processing and analytic solutions including streaming data ingestion, log and relational databases integration, data transformation and data modelling
  • Loaded large sets of structured, semi-structured, and unstructured data with Kafka, Flume and Sqoop
  • Developed Sqoop commands to pull the data from Teradata
  • Developed Kafka consumer to receive and store real time data from Kafka to Amazon S3
  • Extracted data from MongoDB HBase through Sqoop and placed in HDFS for processing
  • Migrated MapReduce jobs into Spark RDD transformations using Scala
  • Installed and configured the spark cluster as well as integrating it with the existing Hadoop cluster
  • Loaded data into Spark RDD and do in memory data computation to generate the output response
  • Worked with Oozie and Zookeeper to manage job workflow and job coordination in the cluster

Hadoop Application Developer-I

ISquare Infotech
12.2015 - 08.2018
  • Project: Design and implement Hadoop platform to support enterprise-wide batch, real time, and ad-hoc data analytic and consumption
  • Responsible to manage data coming from different sources.
  • Developed the code for Importing from Netezza warehouse and exporting data into HDFS and Hive using Sqoop.
  • Used MapReduce for reading, cleaning the data and storing into HDFS.
  • Created Hive external tables by using partitioning applied on top of it.
  • Used Experienced in managing and reviewing Hadoop log files.
  • Involved in defining job flows using Oozie for scheduling jobs to manage apache Hadoop jobs by Directed Acyclic graph (DAG) of actions with control flows.
  • Setup real time data ingestion using Apache Kafka and Flume and storing the data into HBase.
  • Designed and Developed reports and dashboards for Finance domain using Tableau.

Java Developer

ISquare Infotech
05.2014 - 11.2015
  • Project: This was basic standalone school management project and the client was a local school in Rajkot, Gujarat. The project supports absence management, inventory tracking, marksheet, student and staff information storage and other required operations. In this project, I support the development of a backend side. We use spring boot framework for develop model and JDBCTemplate for database connection and transactions. We also used HTML5, CSS, jQuery, React.js for Front end side. We use git version control tool. We use the jenkins for the continues delivery (CI/CD) of the application.
  • Project: This was a fundraiser application for the local charity group who wants to broad their activity. The application contains regular admin side and user side login. It allows to create community to make money for non-profit organizations. It also includes other fund raiser ideas for charity. To develop this application, we use spring framework and tools like Spring Boot, Spring IOC. We use React.js, CSS, HTML 5 and jQuery for UI. We use IBM WebSphere for deployment of application. We use Jenkins for CI/CD and to build the application.
  • Responsibilities: Implemented the backend business layers using spring and web services
  • Implemented Java EE components using Spring MVC, Spring IOC, Spring Transactions, and Spring Security Modules. Developing microservices using Java Spring Boot, Hibernate, Restful services with good quality of code
  • Developed Web API controller with Spring Framework for handling restful requests and responses using
  • Configured the spring integration XML to route request to respective databases. Developing the graphical user interfaces by using, HTML, CSS, and Node.js. Responsible for validation of client interface using JavaScript and jQuery
  • Generate functions and procedures (PL/SQL) for day to day operations with development of applications
  • Generate GraphQL API to load the server data to the client
  • Involved in Single Page Application Framework design using AngularJS, Bootstrap, HTML5, CSS3, W3CSS, SASS, UI Routing
  • Designed and implementing the customized exception handling to handle the exceptions in the application
  • Design and develop the application using Agile Methodologies using JIRA and Confluence.

Education

Master of Science - IT & CA

H.N. Shukla College of Information Technology
India
04.2014

Bachelor of Computer Application -

Kamani Science and Prataprai Arts College
India
04.2012

Skills

  • Integrated development environments
  • Code editors: IntelliJ, Eclipse, Microsoft Visual Studio versions 2008-2019, PyCharm, VS Code
  • Relational databases: MySQL, Oracle, DB2
  • Big Data Technologies: Hadoop, Spark, Kafka, Hive, Sqoop, Hue, Impala
  • Development Languages: Java Core, Spring Boot, Scala, Python
  • Scripting languages: Shell, JavaScript
  • Operating systems: Unix/Linux and Windows family
  • Continuous integration tools: Maven, Jenkins, TeamCity
  • Version Control: GitHub, Bitbucket
  • Cloud: AWS, GCP

Timeline

BigData Developer

RBC
01.2021 - Current

BigData/Spark Developer

National Bank of Canada
01.2020 - 11.2020

Bigdata Developer

Zenith
05.2019 - 08.2019

Hadoop Application Developer-I

ISquare Infotech
12.2015 - 08.2018

Java Developer

ISquare Infotech
05.2014 - 11.2015

Master of Science - IT & CA

H.N. Shukla College of Information Technology

Bachelor of Computer Application -

Kamani Science and Prataprai Arts College
DHARA NAROLA