Experienced Data and Cloud Professional with over 12 years of experience in Applied Machine learning, Data Engineering, Product Development, and Big Data Technologies. Excellent reputation for resolving problems and improving customer satisfaction.
Smart Meters Analytics (SMA) project (Live)
- Scope: E.on Smart meter analytics project focused on analyzing customer’s Smart meters performance metrics and compare them with real-world environment data.
E.ON TargetBI project (Live)
- Scope: E.ON have more than one sister companies. To create a unified MDM (Master Data management) for account, sales, BI data, EON is drawing data from Powercloud Innogy in order to make it available for analytical purposes i.e. Powercloud CDC Data migration from AWS cloud to Azure Cloud with AWS DMS (Database migration service).
LeanIX Integration Pipelines (Live)
- Design and Develop cost effective and secure solution to develop pipelines to integrate LeanIX source with multiple target locations in On-prem, Azure and multiple SaaS platforms like ServiceNow, Coplanner, Symbio, AxonIvy etc
CTL-Cloudera Big Data As Service Project, USA (Live)
Project Description:
Project link: https://www.cloudera.com/about/news-and-blogs/press-releases/2016-09-21-cloudera- centurylink-expand-strategic-alliance-deliver-big-data-service-for-customers.html
BDaaS Project includes complex architecture including product knowledge, Hadoop services knowledge, Cloudera API knowledge, and Cloud automation which initialize CenturyLink
managed a secured cluster for Data Warehousing, Big Data Analytics applications.
Played an integral role as Subject Matter Expert of CenturyLink Big Data as a
Service cloud product. Responsibilities are to develop an Automation API framework in Java/Python which will set up and manage clusters with all services up and running
automatically.
Panera Bread Data Science Platform , USA (Live)
Project Description: Panera, LLC is an American chain of bakery-cafe fast-casual restaurants in the United States and Canada. CenturyLink has SOW with Panera, LLC for Capacity Planning and Production setup. The client required Identification of methodology for tying the online business workload at an order level to the actual utilization of the IT infrastructure and building of sample/representative dashboard/s depicting measures defining the IT resource utilization per order.
CTL Data Lake Project: Data Ingestion API (PoC)
Project Description: CTL Data Lake is a CenturyLink internal project for creating an
application for comprehensive data access and management and then applies data analytics on scalable data.
My responsibilities were to develop and test the REST interface for a data pipeline which
takes data from the customer and dumps to the Kafka topic as well as parsing it with Streaming and stored to the HBase table and HDFS.
DLL Financial solution partner, Mumbai, India (Live)
Design and Develop Azure Data Pipeline with Azure Data Factory, HDInsight, Azure Databricks, Data Lake Analytics, Azure SQL (Stage), Azure DW, Azure ARM and Power BI
Michaels Store, USA (Live)
Design Distributed high-performance Data analysis and Business Intelligence platorm on top of Cloudera Impala, HAproxy, etc.
A&T Insight is a module of CMS (Amdocs Customer Management Suite) is a low-latency Big Data and predictive analytics Applications on top of customer billing data to provide insight-based reports to customer care executives during a call with a customer. Responsibilities/Deliverables
• Software Development with Hive and PIG scripts for analysis of Huge Telecom data.
• Develop dynamic map-reduce jobs and UDFs with Core-java.
• Automatic Data ingestion platform for migration of data from Oracle and H-Base.
• Worked in distributed Gigaspaces Grid clusters for Insight Application.
• Exposure to real-time Spark stream jobs and batch jobs.
• Developed modules for database connections and for structured programming.
• Experience with both Hortanworks and Cloudera distribution of Hadoop (CDH).
• Exposure to data manipulation with Hive queries and Pig scripts on HBase data (No SQL)
• Developed log analysis and real-time monitoring tools for the production application.
• Exposure to ETL job creation, flow diagrams, and job scheduling in DAG.
AT&T Telegence Mobility project create and manage a set of ETL workflows with Pre-Data validation and Post-Data Validations jobs to generate customers' bills from Billing data.
Automation of Campaign mailing system
The campaign mailing interface was designed for the client relation team to test emails and trigger jobs to send a huge number of targeted emails.
- Worked on distributed 7-node cluster queues, Solr, and Sphinx framework.
- Have a good experience with Lucene queries with Mysql on a distributed cluster.
- Experience in MySQL Server 5.2 including the views, Indexes, Procedures Optimization and performance.
- Software development with Java, Python, Perl, Javascript, and MemCache in Live Projects.
- Experience in Advanced Automation programming in Python and Perl for Bulk mailing systems and system tests.
- Impact: Manual processing of the mailer campaign had been stopped with fully automated data product where data can be selected with just a saved query id of target customers and 200k to 2 million emails execution at a time with mail tracking features for demographic analysis of email readers. Awards: Team got awarded by APAC leaders.
Monster online job posts aggregator
Description: A Framework built for the reusability of the jobs scraping process to run 200+ processes at the time
Impact: 121,000 Job posts with this automation in a single month. which saves the workforce hours of monster.com clients.
Admin Mannual of GulfTrip website: A Real Time Web application is B2B process specifically developed to the process of providing accessibility,Sales Reports, Profitability Reports, and Business Rules forGulftrip Website.In this Application, the Client will be able to modify GulfTrip Website which is basically a Tourism Module Website developed with Company Loop Methods.
Work is Development, Testing, and Support
Impact: long contract from client
Natural Language Processing
undefinedAWS Certified Solutions Architect - Associate
Data Hero - E.ON Deutschland, Jun 2022
Awarded as a Data hero award for the VTA OPT-In verification project at E.ON.
Centurylink Preservation Award - CenturyLink Business for Enterprise, Sep 2018
The highest award in Centurylink for consistent improvement in the betterment of the product in a continues feedback loop from customers.
AWS Security Best Practices for Developers
Architecting Microsoft Azure Solutions
Big Data Solutions with Azure Machine Learning
AWS Certified Solutions Architect - Associate
Cloudera Hadoop and Spark Developer - CCA175
Machine Learning With Big Data - UC San Diego