Over 12 years of experience in IT Industry with more than 10 years in Big Data Technologies.
Experience working as a Solutions Architect, Data Architect and Platform Architect, primarily in Big Data domain.
Having sound knowledge of Big Data Ecosystem, DWH, BI and ETL, ELT, IOT, DevOps, AWS and Azure.
Experience in Presales, deal solutioning, sales enablement and client presentations.
Migration planning of tools and technologies from legacy systems to microservices/cloud architecture.
Demonstrated ability to lead and manage teams effectively, motivating team members, and driving successful project delivery.
Proven track record of successfully managing and delivering complex projects within scope, timeline, and budget.
Proficient in setting up and working with large scale Infrastructure and Data Platforms.
Proficient in designing self-service architecture for internal teams, enabling them to use the Data Platforms as PaaS.
Good experience in Datalake/Data-Warehouse operations. Good experience working in Technology Roadmapping and Stakeholder management.
Good knowledge in optimizing performance through software-hardware co-design, including CPU, GPU, DPU and ASIC.
Proficient in evaluating emerging technologies and trends. Advising clients on technology selections, design improvements and implementations
Description: Starburst is the analytics engine for all types of data. It provides the fastest, most efficient analytics engine for a data warehouse, data lake, or data mesh. It unlocks the value of distributed data by making it fast and easy to access, no matter where it lives. Starburst queries data across any database, making it instantly actionable for data-driven organizations platform.
Role:
Tools Used: Trino, Multi Cloud, Kubernetes, Ansible, Kafka, Hive, Teradata, Postgres, Git, Docker, Ranger
Description: Working with multiple clients and teams in identifying the improvements areas in their projects. Designing and developing scalable and robust data engineering and data insight solutions using Nifi, Dremio, Python and Big Data platform.
Role:
Tools Used: Linux, Hadoop, Impala, Python, Nifi, Dremio, Infor, Docker, Grafana, Snowflake, AWS
Description: Delivering Big Data Platforms as a service to internal customers in a self-service environment. Apart from maintaining and automating the existing platform, the team is supposed to explore other useful technologies and integrate them in the existing platform to support analytics, data science and other projects.
Role:
Tools Used: Linux, Hadoop, HIVE, Python, Cloudera Manager, Docker, Nifi, Ansible, Graylog, Grafana, ELK, Impala, Spark, Presto
Description: EDW is the Enterprise level data Warehouse having a business use case for Analytics purposes. EDW is OLAP system used mainly for Analysis and Reporting for Regulatory and as a single ledger across the organization. As a part of Big Data Solution, the team is handling development of four different use cases and actively involved in POCs for upcoming use cases. Cluster size is 150 nodes in Prod.
Role:
Tools Used: Linux, Hadoop, HIVE, Python, Cloudera Manager, Sqoop, Kafka, Tableau, HBase, Ansible, Adobe Project: Big Data Platform
Description: Data is generated through multiple source systems which are fed into hdfs using different data importing tools such as Kafka, Sqoop, scripts etc. Data is processed using hive and pig and then later used by different BI teams for reporting purposes. Kyvos cubes are primarily used to reporting along with Tableau. Cluster size is 180 nodes in Prod.
Role:
Tools Used: Linux, Hadoop, HIVE, Python, Cloudera Manager, Sqoop, Kafka, Kyvos, Tableau, Hbase, Docker, Kubernetes
Description: Data is generated through calls and web usage. The data contains the call transactions, which are used as the source for the populating the data warehouse. The data is also generated through web usage by any user. Hadoop Team is getting daily feeds in form of csv files which are stored on HDFS, and then pig jobs are run on that data for ETL. After which Tableu is used to graphical analysis.
Role:
Tools Used: Linux, Hadoop, HUE, PIG, Ambari, Sqoop, Kerberos, Tableau
Description: Micron is a world leader in Semi-Conductors. Data is generated in the process of forming a semi-conductor chip from wafers, which is finally stored in csv files. Hadoop Team is getting daily feeds in form of csv files which are stored on HDFS, and then hive jobs are run on that data for manipulations. After which Neo4j is used to graphical analysis.
Role:
Tools Used: Linux, Hadoop, Hive, Ambari, Sqoop, Kerberos, Neo4j
Description: This was a data migration project for development as well as production support, based on onsite/offshore delivery model. The data is loaded into Siebel base tables after applying business logic to the raw data received from ODS using Informatica. This also includes monitoring and fixing production bugs, creating new Informatica mappings, scheduling of workflows for loading new data.
Role:
Tools Used: Informatica, PL/SQL, Unix, Cognos, Essbase
Distributed Computing, Big Data, Microservices, All major Clouds, Datalakes/Lakehouse, TOGAF, Snowflake
Python, SQL, Shell scripting, Starburst, CDP, Ansible, Docker, Kubernetes, Archi, Informatica, Trino, Druid, Dremio, Jenkins, NiFi, Kafka, ELK Stack, Grafana, Prometheus, Birst, Impala, Trino, Git, Jenkins, JMeter, LucidChart, Airflow, MySql, No-SQL Databases, Neo4j, Postgres
Solutions Archirect, Technical Architect, Technical Account Manager