CakeResume Talent Search

上級
On
4〜6年
6〜10年
10〜15年
15年以上
Avatar of the user.
Avatar of the user.
Software Development Manager @Yo-Win Technology Co., Ltd.
2023 ~ 現在
Software Development
1ヶ月以内
PHP
就職中
就職を希望していません
フルタイム / リモートワークに興味あり
10〜15年
Ming Chuan University
Information Management
Avatar of Jason Chang.
Avatar of Jason Chang.
Senior Software Engineer @Innova Solutions Taiwan
2019 ~ 現在
Backend developer/Full-stack developer
1ヶ月以内
the development cycle. 2. Use Java and TestNG to design and develop a cross-team (four-team) CI framework to enhance the process of E2E testing. 3. Use Scala and Scalatest to design and develop an automation testing framework at the ingestion stage. 4. Clinical fake data generator, using Python and Faker. 5. Design test plans and test cases and coordinate the whole testing process in the team. 6. Build a centralized and standardized testing flow to give a clear guide to the QA folks. 7. Coordinate with
Java
Scala
JavaScript
就職中
就職を希望していません
フルタイム / リモートワークに興味あり
6〜10年
國立中興大學 (National Chung Hsin University, NCHU)
資訊管理
Avatar of Winnie Chen.
Avatar of Winnie Chen.
Past
Manufacturing Project Manager @YueHo Glass Co., LTD.
2017 ~ 2023
Project Manager
6ヶ月以内
. Collaborated with clients to define their requirements, achieving a 90% satisfaction rate by ensuring a clear understanding of their needs and expectations. Designed and developed tooling using Siemens NX (UG) and ATOS 3D measurement, leading to notable improvements in product precision Key driver in achieving QMS certification, reduced critical defects in audits by streamlining documentation and implementing internal training. Education BA in Interior Spatial Design • University of the Arts LondonCore Skills Project Management Budget Control Define Requirement Feasibility Study Communication Skills Problem Solving Product Development Tooling Development Computer Skills Siemens NX (UG) AutoCAD ATOS 3D Microsoft Office
Project Management
Budget Control
Feasibility Study
無職
フルタイム / リモートワークに興味あり
6〜10年
University of the Arts London
Interior Spatial Design
Avatar of David Bryant.
Avatar of David Bryant.
Past
Trade Services @John S.James Co
2017 ~ 2018
International Trade Specialist
1年以上
speaking. My ultimate goal is to gain full-time employment with a dynamic, forward-focused company and maintain a healthy work/life balance. Skills Import/Export Logistics • Cargo Consolidation • Cargo Planning • Ocean Transportation • Air cargo • Customer Service • Written and Oral Communication Skills • Public Speaking • Negotiating Contracts • Documentation • Creative Problem Solving • Detail Oriented • Critical Thinking • Spanish Communication Work Experience Account Executive • John S.James Co JanuarySeptember 2018 Create detailed business plans to facilitate paths to goals and quotas Manage the entire sales cycle from finding a client to securing a deal. Unearth new sales opportunities
Import/Export
logistics
Cargo Consolidation
無職
フルタイム / リモートワークに興味あり
15年以上
Strayer University
MBA-MARKETING
Avatar of 吳明赫.
Avatar of 吳明赫.
Test Equipment Technician @Micron Technology 台灣美光
2018 ~ 現在
製程工程師、設備工程師、半導體工程師
1ヶ月以内
installation, modification, project management, and maintenance support for pilot equipment sets *Solve straightforward problems with guidance to identify root cause of equipment shifts *Provide suggested solutions to real-time equipment problems *Maintain active projects to reduce equipment total cost of ownership *Demonstrate ownership of assigned projects including coordination, installation, maintenance, documentation, modification, and overall support of equipment projects *Recommend equipment changes to improve process metrics *Transfer owned equipment knowledge to manufacturing with good support and documentation *Research, gather, and analyze information related to specific assignments within your technical discipline *During the mass production stage, MFG production capacity requirements need to
Microsoft Office
critical thinking
Analytical Skills
就職中
面接の用意ができています
フルタイム / リモートワークに興味なし
10〜15年
南台科技大學
電機工程系
Avatar of Samantha Jae Lipit.
Avatar of Samantha Jae Lipit.
Consultant @Global Advanced IT Connections
2024 ~ 2024
Product Owner, Business Analyst, System Analyst
1ヶ月以内
Tech Leads to outline technical specifications on Confluence. • Owned the Release Management and streamlined Product Releases by: - Outlining deployment tasks for each iteration for efficient execution - Coordinating cross-functionally with Product and DevOps Teams - Crafting Hand-over documents - Creating internal Release Notes and Announcements - Maintaining accurate Product Version Trackers DOCUMENTATION CHAMPION • Acted as a documentation champion, responsible for developing and maintaining critical resources: - Runbooks: Created and maintained comprehensive runbooks with step-by-step instructions and illustrations to guide users on product usage and on critical process execution. - Release and Product Version Trackers: Established and maint...
YAML
AWS
Azure
就職中
面接の用意ができています
フルタイム / リモートワークに興味あり
6〜10年
Adamson University
Information Technology
Avatar of Devraj Kumar.
Avatar of Devraj Kumar.
Staff Engineer @NextGen Healthcare India
2019 ~ 現在
Senior Software Developer
3ヶ月以内
of tools and frameworks, showcasing expertise in software engineering and project execution. Software Engineer • KPIT Technologies Ltd MayApril 2016 • Spearheaded the enhancement of the CoDeg tool by leveraging technologies such as LaTeX, MKS, Total Commander, SQL, and C#. • Instrumental in devising robust failsafe specifications and effectively resolving critical bug issues. Software Engineer • GRASKO SOLUTIONS PRIVATE LIMITED JanuaryApril 2014 • Spearheaded full-stack development, expertly authoring SQL stored procedures for robust data handling and manipulation. • Leveraged C# ADO.Net for efficient UI binding, delivering seamless and responsive user experiences. • Designed intricate reports with Crystal
C#.NET development
PL/SQL
LINQ
就職中
面接の用意ができています
フルタイム / リモートワークに興味あり
10〜15年
North Maharastra University, Jalgaon, Maharastra
Computer Science & Engineering
Avatar of the user.
Avatar of the user.
Past
Project Leader | Senior Automation Test Engineer @Zealogics Inc
2023 ~ 2024
SR. Software QA Engineer
1ヶ月以内
python
Manual Testing
Automation Testing
無職
面接の用意ができています
フルタイム / リモートワークに興味あり
10〜15年
Feng Chia University
Avatar of Dhimas Prasetyo.
Avatar of Dhimas Prasetyo.
IT Data Center Infrasturcture Enginer @PT Smartfren Telecom Tbk
2020 ~ 現在
IT
1ヶ月以内
failures or emergencies. Conduct regular recovery tests to verify backup effectiveness. Troubleshooting and Issue Resolution: Identify and resolve issues related to backup operations, including software glitches, connectivity problems, and hardware failures. Performance Optimization: Continuously monitor and optimize backup processes to improve efficiency and minimize resource usage. Documentation and Reporting: Maintain comprehensive documentation of backup procedures, configurations, and troubleshooting steps. Generate regular reports on backup status, performance metrics, and compliance with backup policies. Hardware Management: Monitor and manage backup hardware components, including servers, storage systems, and networking equipment. Utilize tools like iDRAC for hardware
Microsoft Office
MySQL
Laravel Framework
就職中
就職希望
フルタイム / リモートワークに興味あり
6〜10年
Institut Teknologi Adhi Tama Surabaya
Computer Science
Avatar of the user.
Avatar of the user.
Marketing Graphic Design @Sinar Gading Group
2023 ~ 現在
Desain Grafis
1ヶ月以内
Presentation Skills
Critical Thinking
Collaboration & Teamwork
就職中
就職希望
フルタイム / リモートワークに興味あり
6〜10年
STMIK WIDURI
Information System

最も簡単で効果的な採用プラン

80万枚以上の履歴書を検索して、率先して求人応募者と連絡をとって採用効率を高めましょう。何百もの企業に選ばれています。

  • 検索結果をすべて閲覧
  • 新しい会話を無制限に始められます
  • 有料企業にのみ履歴書を公開
  • ユーザーのメールアドレスと電話番号を確認
検索のコツ
1
Search a precise keyword combination
senior backend php
If the number of the search result is not enough, you can remove the less important keywords
2
Use quotes to search for an exact phrase
"business development"
3
Use the minus sign to eliminate results containing certain words
UI designer -UX
無料プランでは公開済みの履歴書のみ利用できます。
上級プランにアップグレードして、CakeResume限定の何百万の履歴書など、すべての検索結果を閲覧しましょう。

Definition of Reputation Credits

Technical Skills
Specialized knowledge and expertise within the profession (e.g. familiar with SEO and use of related tools).
Problem-Solving
Ability to identify, analyze, and prepare solutions to problems.
Adaptability
Ability to navigate unexpected situations; and keep up with shifting priorities, projects, clients, and technology.
Communication
Ability to convey information effectively and is willing to give and receive feedback.
Time Management
Ability to prioritize tasks based on importance; and have them completed within the assigned timeline.
Teamwork
Ability to work cooperatively, communicate effectively, and anticipate each other's demands, resulting in coordinated collective action.
Leadership
Ability to coach, guide, and inspire a team to achieve a shared goal or outcome effectively.
1ヶ月以内
India
Professional Background
現在の状況
求人検索の進捗
Professions
Fields of Employment
職務経験
Management
スキル
言語
Job search preferences
希望のポジション
求人タイプ
希望の勤務地
リモートワーク
Freelance
学歴
学校
専攻
印刷

Subham Sahu

Seeking a challenging environment that encourages learning, creativity, provides exposures to new ideas and stimulates personal and professional growth along with organizational growth. 

[email protected]

+91- 9039347186

HSR Layout Sector 5, Bangalore, Karnataka, 560034

Technical Skills


Cloud Skills

MS Azure, Azure Data Factory, Data Lake, Azure Dev-Ops, Azure Synapse Analytics, Azure Data bricks, ETL/ELT, Storage Blob, Azure function, logic apps, Delta lake, Kafka, Grafana, Streaming Data

Programming 

Python, Pandas, PySpark, Beautiful Soup
Spark SQL, Scala

Database

MS SQL SERVER, Azure SQL Database, T-SQL, MySQL, Snowflake, Cosmos DB, Hive, Delta tables, IBM DB2 i series


Professional Summary


  • 6.3+ years of experience in Microsoft Azure cloud platform such as Azure Data Factory, Data Bricks, Data Lake, Azure Synapse analytics, functions, logic Apps, SQL DB, Cosmos DB, PySpark and Python along with Aviation, Oil & Gas, Energy Sector and Pharmaceutical domain knowledge.
  • Creates Several Relational and Non-relational Data Modelling of Data and Creates Prototype diagram using Draw.io.
  • Cleaning and transformation of complex data using data pipelines & notebook in the Azure cloud using Data Factory and Data Bricks with PySpark
  • Perform root cause analysis and resolve the production and data issues.
  • Responsible for design, development, modification, debug and maintenance of Data Pipelines.
  • Delivers technical accountability for team specific work products within an application and provide technical support during solution design for new requirements.
  • Maintains the Existing project along with new development using JIRA in Agile methodology to euiinhance the growth of productivity. 
  • Utilize sound engineering practices to deliver functional, stable, and scalable solutions to new or existing problems.
  • Involves in requirement analysis and business discussions with clients, delivery process, etc. 
  • Excellent interpersonal skills with strong analytical and problem-solving skills. 

Work Experience

Publicis Sapient, April 2023 - Present

Senior Associate Data Engineering L2 (MS Azure Cloud, PySpark)

  • Working on Orx RFP Healthcare & Insurance data of Client. 
  • Design the data model for data migration.
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines through Azure Data factory, Databricks and Kafka streaming/Batch data.
  • Maintains technical documentation using  Confluence platform
  • Manage the projects using Agile Methodology.

Ness Digital Engineering, May 2021 - April-2023

Senior Data Engineer (MS Azure Cloud, PySpark)

  • Works on Clinical and drug trial data of Giant organizations. 
  • Design the non-relational common Data model on Cosmos DB.
  • Analysis of existing relational data for data model creation. 
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines in Azure Data factory, Databricks and Azure Synapse analytics.
  • Maintains technical documentation using Docusaurus and Confluence platform
  • Manage the teams and projects using Agile Methodology along with Client Interaction.

IHS Markit ltd., Oct 2017 - Apr 2021

Data Engineer (MS Azure Cloud, PySparkPython)

  • Extract the complex and Bulk Data related to Aviation, Gas, Energy Sector
  • Investigate issues by reviewing/debugging pipelines, provide fixes, workarounds, and review changes for operability to maintain existing data solutions.
  • Experience of building data pipelines using Databricks (Azure Data Factory and Apache Spark).
  • We extract information by using Python and Kofax RPA tool for fast Crawling.
  • The general orientation of team to reduce manual efforts, growth of products and saving of time during ETL/ELT Process.
  • Have done the integration of third-party tools such as Shutil, Pandas, file transfer from local to azure storage container cloud services using Python.

Project

Healthcare and Insurance Analytics

ORX RFP DMA Explorer Analytics: -

We are building data Application and Framework that contains the data about the Healthcare & Insurance analytics and broadcast the visualizations through power BI reports for business users. Stores the huge data at Lakehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, IBM DB2 i series, Data Lake, Spark SQL, GCP Kafka, Scala, Delta lake & Tables, PySpark and Python.
  • We pull the data from source relational DB such as MySQL and IBM DB2 and move data to Data lake in form of Parquet.
  • Created the Common Data Model using draw.io.
  • Created Azure Data Factory pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Azure Databricks Notebook
  • Using Azure Monitor, Grafana to monitors the ADF Data pipelines and GCP Kafka jobs
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Confluence platform.
  • Generates the Claim, hospitalization and expenses reports using Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Clinical and Drug Trial Analytics

Pharma, Healthcare and Drugs trials: -

We are building Application that contains the data about the Pharma, Healthcare and Drugs trial and broadcast the visualizations through power BI reports. Stores the huge data at data warehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, MS SQL SERVER, Azure Synapse, Data Lake, PySpark and Cosmos DB.
  • We pull the data from relational DB such as MySQL and SQL Server and move data to cosmos DB after the several transformations.
  • Created the Common Data Model using SQL API of Cosmos DB using draw.io.
  • Created Azure Synapse data analytics pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Synapse Notebook
  • Using Azure Monitor, New Relic & analytics to monitors the Synapse Data pipelines
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Docusaurus and Confluence platform.
  • Generates the ingredient, excipients, numerator devices, dosage, artifact, sub artifact, products reports using Snowflake & Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Energy Analytics

Oil, Gas and Coal, OMDC:-

We are building products that contains the information about the oil and gas prices, tenders bidding, consumption and production data country-wise with other factors included.
  • Used various technologies like Python, Azure Database, Data Factory, Data Lake, Data bricks, PySpark, T-SQL, Pandas and Power BI
  • We crawl Complex data from Business, external resources and websites using Python and dump the Files into Azure blobs and data lake.
  • Created Azure Data Factory data pipelines, Activities, Linked services, IR, Triggers for process of ETL/ELT.
  • Write Azure Functions to Implement custom logics for transformation and automation in Python scripts.
  • Using Azure Monitor & analytics to monitors the ADF pipelines. Implemented CI/CD for moving pipelines/Scripts from one environment to another by ARM templates using Azure DevOps.
  • Generates the prices, production, consumption comparison reports using Power BI.

Aviation, IHS Markit Ltd.

Cargo & Flight BI: -

We are having the several pipeline which populates the data of Cargo, Shipment, Booking etc. for multiple Marts. On basis of this, Daily, weekly and monthly Report are generated.
  • Performed several transformations, structuring and cleansing on data including various transformations using Pyspark, Spark SQL and Delta tables.
  • Built multiple data pipelines and job clusters using Azure Data Factory and Databricks.
  • Handling of data on basis of refresh date and SQP date for Incremental load.
  • Worked in Agile methodology using JIRA
  • Highly proficient in using Spark-SQL for developing complex joins and aggregations.
  • Hands on experience on Synapse data warehousing of External tables through Parquet files in Data Lake.
  • Azure Data factory, Azure Data-lake, Azure Databricks, Delta Table, DevOps, Pyspark and Delta Tables. 
  • Cleaning of Cargo & flight data and movement from MS SQL, Hive, traditional Hadoop system and SFTP to Azure Data-lake and Delta tables in Databricks Using Azure Data factory pipelines and Databricks notebook.

Certifications

  • Certification in Azure Data Fundamentals from Microsoft Azure.    
  • Certification in Databricks Certified Data Engineer Associate from Databricks.
  • Certification in Databricks Certified Apache Spark Developer Associate 3.0 from Databricks.
  • Certification in Databricks Accredited Lakehouse fundamentals from Databricks. 
  • Certification In Master Data Analysis with Python - Intro To pandas from Udemy.

Rewards and Achievement

  • Achieved Team player award for Pharma & Clinical project in Q3, 2021.
  • Achieved Best performance Award for Energy analytics project in Q2, 2020.
  • Achieved Peer Award for Optimization of Parts intelligence pipelines Q3, 2019.
  • Achieved Rewards as Team player for Energy Analytics projects in Q4, 2018.   

Education

 B.E. in Electronics Engineering - 73.46% (2016) 
 Institute of Engineering, JIWAJI University, Gwalior 

Core Skills & Strengths 

 ● Team Management          ● Leadership Quality 

 ● Passionate and Creative  ● Quick Learner  

 ● Positive Thinking               ● Punctual 

 ● Motivated                           ● Flexible

Area of Interest 

 ● Interacting with people.  

 ● Willingness to learn new skills.

 ● Cooking                             

 ● Chess

Resume
プロフィール

Subham Sahu

Seeking a challenging environment that encourages learning, creativity, provides exposures to new ideas and stimulates personal and professional growth along with organizational growth. 

[email protected]

+91- 9039347186

HSR Layout Sector 5, Bangalore, Karnataka, 560034

Technical Skills


Cloud Skills

MS Azure, Azure Data Factory, Data Lake, Azure Dev-Ops, Azure Synapse Analytics, Azure Data bricks, ETL/ELT, Storage Blob, Azure function, logic apps, Delta lake, Kafka, Grafana, Streaming Data

Programming 

Python, Pandas, PySpark, Beautiful Soup
Spark SQL, Scala

Database

MS SQL SERVER, Azure SQL Database, T-SQL, MySQL, Snowflake, Cosmos DB, Hive, Delta tables, IBM DB2 i series


Professional Summary


  • 6.3+ years of experience in Microsoft Azure cloud platform such as Azure Data Factory, Data Bricks, Data Lake, Azure Synapse analytics, functions, logic Apps, SQL DB, Cosmos DB, PySpark and Python along with Aviation, Oil & Gas, Energy Sector and Pharmaceutical domain knowledge.
  • Creates Several Relational and Non-relational Data Modelling of Data and Creates Prototype diagram using Draw.io.
  • Cleaning and transformation of complex data using data pipelines & notebook in the Azure cloud using Data Factory and Data Bricks with PySpark
  • Perform root cause analysis and resolve the production and data issues.
  • Responsible for design, development, modification, debug and maintenance of Data Pipelines.
  • Delivers technical accountability for team specific work products within an application and provide technical support during solution design for new requirements.
  • Maintains the Existing project along with new development using JIRA in Agile methodology to euiinhance the growth of productivity. 
  • Utilize sound engineering practices to deliver functional, stable, and scalable solutions to new or existing problems.
  • Involves in requirement analysis and business discussions with clients, delivery process, etc. 
  • Excellent interpersonal skills with strong analytical and problem-solving skills. 

Work Experience

Publicis Sapient, April 2023 - Present

Senior Associate Data Engineering L2 (MS Azure Cloud, PySpark)

  • Working on Orx RFP Healthcare & Insurance data of Client. 
  • Design the data model for data migration.
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines through Azure Data factory, Databricks and Kafka streaming/Batch data.
  • Maintains technical documentation using  Confluence platform
  • Manage the projects using Agile Methodology.

Ness Digital Engineering, May 2021 - April-2023

Senior Data Engineer (MS Azure Cloud, PySpark)

  • Works on Clinical and drug trial data of Giant organizations. 
  • Design the non-relational common Data model on Cosmos DB.
  • Analysis of existing relational data for data model creation. 
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines in Azure Data factory, Databricks and Azure Synapse analytics.
  • Maintains technical documentation using Docusaurus and Confluence platform
  • Manage the teams and projects using Agile Methodology along with Client Interaction.

IHS Markit ltd., Oct 2017 - Apr 2021

Data Engineer (MS Azure Cloud, PySparkPython)

  • Extract the complex and Bulk Data related to Aviation, Gas, Energy Sector
  • Investigate issues by reviewing/debugging pipelines, provide fixes, workarounds, and review changes for operability to maintain existing data solutions.
  • Experience of building data pipelines using Databricks (Azure Data Factory and Apache Spark).
  • We extract information by using Python and Kofax RPA tool for fast Crawling.
  • The general orientation of team to reduce manual efforts, growth of products and saving of time during ETL/ELT Process.
  • Have done the integration of third-party tools such as Shutil, Pandas, file transfer from local to azure storage container cloud services using Python.

Project

Healthcare and Insurance Analytics

ORX RFP DMA Explorer Analytics: -

We are building data Application and Framework that contains the data about the Healthcare & Insurance analytics and broadcast the visualizations through power BI reports for business users. Stores the huge data at Lakehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, IBM DB2 i series, Data Lake, Spark SQL, GCP Kafka, Scala, Delta lake & Tables, PySpark and Python.
  • We pull the data from source relational DB such as MySQL and IBM DB2 and move data to Data lake in form of Parquet.
  • Created the Common Data Model using draw.io.
  • Created Azure Data Factory pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Azure Databricks Notebook
  • Using Azure Monitor, Grafana to monitors the ADF Data pipelines and GCP Kafka jobs
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Confluence platform.
  • Generates the Claim, hospitalization and expenses reports using Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Clinical and Drug Trial Analytics

Pharma, Healthcare and Drugs trials: -

We are building Application that contains the data about the Pharma, Healthcare and Drugs trial and broadcast the visualizations through power BI reports. Stores the huge data at data warehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, MS SQL SERVER, Azure Synapse, Data Lake, PySpark and Cosmos DB.
  • We pull the data from relational DB such as MySQL and SQL Server and move data to cosmos DB after the several transformations.
  • Created the Common Data Model using SQL API of Cosmos DB using draw.io.
  • Created Azure Synapse data analytics pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Synapse Notebook
  • Using Azure Monitor, New Relic & analytics to monitors the Synapse Data pipelines
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Docusaurus and Confluence platform.
  • Generates the ingredient, excipients, numerator devices, dosage, artifact, sub artifact, products reports using Snowflake & Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Energy Analytics

Oil, Gas and Coal, OMDC:-

We are building products that contains the information about the oil and gas prices, tenders bidding, consumption and production data country-wise with other factors included.
  • Used various technologies like Python, Azure Database, Data Factory, Data Lake, Data bricks, PySpark, T-SQL, Pandas and Power BI
  • We crawl Complex data from Business, external resources and websites using Python and dump the Files into Azure blobs and data lake.
  • Created Azure Data Factory data pipelines, Activities, Linked services, IR, Triggers for process of ETL/ELT.
  • Write Azure Functions to Implement custom logics for transformation and automation in Python scripts.
  • Using Azure Monitor & analytics to monitors the ADF pipelines. Implemented CI/CD for moving pipelines/Scripts from one environment to another by ARM templates using Azure DevOps.
  • Generates the prices, production, consumption comparison reports using Power BI.

Aviation, IHS Markit Ltd.

Cargo & Flight BI: -

We are having the several pipeline which populates the data of Cargo, Shipment, Booking etc. for multiple Marts. On basis of this, Daily, weekly and monthly Report are generated.
  • Performed several transformations, structuring and cleansing on data including various transformations using Pyspark, Spark SQL and Delta tables.
  • Built multiple data pipelines and job clusters using Azure Data Factory and Databricks.
  • Handling of data on basis of refresh date and SQP date for Incremental load.
  • Worked in Agile methodology using JIRA
  • Highly proficient in using Spark-SQL for developing complex joins and aggregations.
  • Hands on experience on Synapse data warehousing of External tables through Parquet files in Data Lake.
  • Azure Data factory, Azure Data-lake, Azure Databricks, Delta Table, DevOps, Pyspark and Delta Tables. 
  • Cleaning of Cargo & flight data and movement from MS SQL, Hive, traditional Hadoop system and SFTP to Azure Data-lake and Delta tables in Databricks Using Azure Data factory pipelines and Databricks notebook.

Certifications

  • Certification in Azure Data Fundamentals from Microsoft Azure.    
  • Certification in Databricks Certified Data Engineer Associate from Databricks.
  • Certification in Databricks Certified Apache Spark Developer Associate 3.0 from Databricks.
  • Certification in Databricks Accredited Lakehouse fundamentals from Databricks. 
  • Certification In Master Data Analysis with Python - Intro To pandas from Udemy.

Rewards and Achievement

  • Achieved Team player award for Pharma & Clinical project in Q3, 2021.
  • Achieved Best performance Award for Energy analytics project in Q2, 2020.
  • Achieved Peer Award for Optimization of Parts intelligence pipelines Q3, 2019.
  • Achieved Rewards as Team player for Energy Analytics projects in Q4, 2018.   

Education

 B.E. in Electronics Engineering - 73.46% (2016) 
 Institute of Engineering, JIWAJI University, Gwalior 

Core Skills & Strengths 

 ● Team Management          ● Leadership Quality 

 ● Passionate and Creative  ● Quick Learner  

 ● Positive Thinking               ● Punctual 

 ● Motivated                           ● Flexible

Area of Interest 

 ● Interacting with people.  

 ● Willingness to learn new skills.

 ● Cooking                             

 ● Chess