CakeResume 找人才

进阶搜寻
On
4 到 6 年
6 到 10 年
10 到 15 年
15 年以上
Avatar of the user.
Avatar of the user.
Software Development Manager @Yo-Win Technology Co., Ltd.
2023 ~ 现在
Software Development
一個月內
PHP
就职中
目前没有兴趣寻找新的机会
全职 / 对远端工作有兴趣
10 到 15 年
Ming Chuan University
Information Management
Avatar of Jason Chang.
Avatar of Jason Chang.
Senior Software Engineer @Innova Solutions Taiwan
2019 ~ 现在
Backend developer/Full-stack developer
一個月內
the development cycle. 2. Use Java and TestNG to design and develop a cross-team (four-team) CI framework to enhance the process of E2E testing. 3. Use Scala and Scalatest to design and develop an automation testing framework at the ingestion stage. 4. Clinical fake data generator, using Python and Faker. 5. Design test plans and test cases and coordinate the whole testing process in the team. 6. Build a centralized and standardized testing flow to give a clear guide to the QA folks. 7. Coordinate with
Java
Scala
JavaScript
就职中
目前没有兴趣寻找新的机会
全职 / 对远端工作有兴趣
6 到 10 年
國立中興大學 (National Chung Hsin University, NCHU)
資訊管理
Avatar of Winnie Chen.
Avatar of Winnie Chen.
曾任
Manufacturing Project Manager @YueHo Glass Co., LTD.
2017 ~ 2023
Project Manager
半年內
. Collaborated with clients to define their requirements, achieving a 90% satisfaction rate by ensuring a clear understanding of their needs and expectations. Designed and developed tooling using Siemens NX (UG) and ATOS 3D measurement, leading to notable improvements in product precision Key driver in achieving QMS certification, reduced critical defects in audits by streamlining documentation and implementing internal training. Education BA in Interior Spatial Design • University of the Arts LondonCore Skills Project Management Budget Control Define Requirement Feasibility Study Communication Skills Problem Solving Product Development Tooling Development Computer Skills Siemens NX (UG) AutoCAD ATOS 3D Microsoft Office
Project Management
Budget Control
Feasibility Study
待业中
全职 / 对远端工作有兴趣
6 到 10 年
University of the Arts London
Interior Spatial Design
Avatar of David Bryant.
Avatar of David Bryant.
曾任
Trade Services @John S.James Co
2017 ~ 2018
International Trade Specialist
超過一年
speaking. My ultimate goal is to gain full-time employment with a dynamic, forward-focused company and maintain a healthy work/life balance. Skills Import/Export Logistics • Cargo Consolidation • Cargo Planning • Ocean Transportation • Air cargo • Customer Service • Written and Oral Communication Skills • Public Speaking • Negotiating Contracts • Documentation • Creative Problem Solving • Detail Oriented • Critical Thinking • Spanish Communication Work Experience Account Executive • John S.James Co JanuarySeptember 2018 Create detailed business plans to facilitate paths to goals and quotas Manage the entire sales cycle from finding a client to securing a deal. Unearth new sales opportunities
Import/Export
logistics
Cargo Consolidation
待业中
全职 / 对远端工作有兴趣
15 年以上
Strayer University
MBA-MARKETING
Avatar of 吳明赫.
Avatar of 吳明赫.
Test Equipment Technician @Micron Technology 台灣美光
2018 ~ 现在
製程工程師、設備工程師、半導體工程師
一個月內
installation, modification, project management, and maintenance support for pilot equipment sets *Solve straightforward problems with guidance to identify root cause of equipment shifts *Provide suggested solutions to real-time equipment problems *Maintain active projects to reduce equipment total cost of ownership *Demonstrate ownership of assigned projects including coordination, installation, maintenance, documentation, modification, and overall support of equipment projects *Recommend equipment changes to improve process metrics *Transfer owned equipment knowledge to manufacturing with good support and documentation *Research, gather, and analyze information related to specific assignments within your technical discipline *During the mass production stage, MFG production capacity requirements need to
Microsoft Office
critical thinking
Analytical Skills
就职中
正在积极求职中
全职 / 暂不考虑远端工作
10 到 15 年
南台科技大學
電機工程系
Avatar of Samantha Jae Lipit.
Avatar of Samantha Jae Lipit.
Consultant @Global Advanced IT Connections
2024 ~ 2024
Product Owner, Business Analyst, System Analyst
一個月內
Tech Leads to outline technical specifications on Confluence. • Owned the Release Management and streamlined Product Releases by: - Outlining deployment tasks for each iteration for efficient execution - Coordinating cross-functionally with Product and DevOps Teams - Crafting Hand-over documents - Creating internal Release Notes and Announcements - Maintaining accurate Product Version Trackers DOCUMENTATION CHAMPION • Acted as a documentation champion, responsible for developing and maintaining critical resources: - Runbooks: Created and maintained comprehensive runbooks with step-by-step instructions and illustrations to guide users on product usage and on critical process execution. - Release and Product Version Trackers: Established and maint...
YAML
AWS
Azure
就职中
正在积极求职中
全职 / 对远端工作有兴趣
6 到 10 年
Adamson University
Information Technology
Avatar of Devraj Kumar.
Avatar of Devraj Kumar.
Staff Engineer @NextGen Healthcare India
2019 ~ 现在
Senior Software Developer
三個月內
of tools and frameworks, showcasing expertise in software engineering and project execution. Software Engineer • KPIT Technologies Ltd MayApril 2016 • Spearheaded the enhancement of the CoDeg tool by leveraging technologies such as LaTeX, MKS, Total Commander, SQL, and C#. • Instrumental in devising robust failsafe specifications and effectively resolving critical bug issues. Software Engineer • GRASKO SOLUTIONS PRIVATE LIMITED JanuaryApril 2014 • Spearheaded full-stack development, expertly authoring SQL stored procedures for robust data handling and manipulation. • Leveraged C# ADO.Net for efficient UI binding, delivering seamless and responsive user experiences. • Designed intricate reports with Crystal
C#.NET development
PL/SQL
LINQ
就职中
正在积极求职中
全职 / 对远端工作有兴趣
10 到 15 年
North Maharastra University, Jalgaon, Maharastra
Computer Science & Engineering
Avatar of the user.
Avatar of the user.
曾任
Project Leader | Senior Automation Test Engineer @Zealogics Inc
2023 ~ 2024
SR. Software QA Engineer
一個月內
python
Manual Testing
Automation Testing
待业中
正在积极求职中
全职 / 对远端工作有兴趣
10 到 15 年
Feng Chia University
Avatar of Dhimas Prasetyo.
Avatar of Dhimas Prasetyo.
IT Data Center Infrasturcture Enginer @PT Smartfren Telecom Tbk
2020 ~ 现在
IT
一個月內
failures or emergencies. Conduct regular recovery tests to verify backup effectiveness. Troubleshooting and Issue Resolution: Identify and resolve issues related to backup operations, including software glitches, connectivity problems, and hardware failures. Performance Optimization: Continuously monitor and optimize backup processes to improve efficiency and minimize resource usage. Documentation and Reporting: Maintain comprehensive documentation of backup procedures, configurations, and troubleshooting steps. Generate regular reports on backup status, performance metrics, and compliance with backup policies. Hardware Management: Monitor and manage backup hardware components, including servers, storage systems, and networking equipment. Utilize tools like iDRAC for hardware
Microsoft Office
MySQL
Laravel Framework
就职中
目前会考虑了解新的机会
全职 / 对远端工作有兴趣
6 到 10 年
Institut Teknologi Adhi Tama Surabaya
Computer Science
Avatar of the user.
Avatar of the user.
Marketing Graphic Design @Sinar Gading Group
2023 ~ 现在
Desain Grafis
一個月內
Presentation Skills
Critical Thinking
Collaboration & Teamwork
就职中
目前会考虑了解新的机会
全职 / 对远端工作有兴趣
6 到 10 年
STMIK WIDURI
Information System

最轻量、快速的招募方案,数百家企业的选择

搜寻简历,主动联系求职者,提升招募效率。

  • 浏览所有搜寻结果
  • 每日可无限次数开启陌生对话
  • 搜尋僅開放付費企業檢視的简历
  • 检视使用者信箱 & 电话
搜寻技巧
1
Search a precise keyword combination
senior backend php
If the number of the search result is not enough, you can remove the less important keywords
2
Use quotes to search for an exact phrase
"business development"
3
Use the minus sign to eliminate results containing certain words
UI designer -UX
免费方案仅能搜寻公开简历。
升级至进阶方案,即可浏览所有搜寻结果(包含数万笔览仅在 CakeResume 平台上公开的简历)。

职场能力评价定义

专业技能
该领域中具备哪些专业能力(例如熟悉 SEO 操作,且会使用相关工具)。
问题解决能力
能洞察、分析问题,并拟定方案有效解决问题。
变通能力
遇到突发事件能冷静应对,并随时调整专案、客户、技术的相对优先序。
沟通能力
有效传达个人想法,且愿意倾听他人意见并给予反馈。
时间管理能力
了解工作项目的优先顺序,有效运用时间,准时完成工作内容。
团队合作能力
具有向心力与团队责任感,愿意倾听他人意见并主动沟通协调。
领导力
专注于团队发展,有效引领团队采取行动,达成共同目标。
一個月內
India
专业背景
目前状态
求职阶段
专业
产业
工作年资
管理经历
技能
语言能力
求职偏好
希望获得的职位
预期工作模式
期望的工作地点
远端工作意愿
接案服务
学历
学校
主修科系
列印

Subham Sahu

Seeking a challenging environment that encourages learning, creativity, provides exposures to new ideas and stimulates personal and professional growth along with organizational growth. 

[email protected]

+91- 9039347186

HSR Layout Sector 5, Bangalore, Karnataka, 560034

Technical Skills


Cloud Skills

MS Azure, Azure Data Factory, Data Lake, Azure Dev-Ops, Azure Synapse Analytics, Azure Data bricks, ETL/ELT, Storage Blob, Azure function, logic apps, Delta lake, Kafka, Grafana, Streaming Data

Programming 

Python, Pandas, PySpark, Beautiful Soup
Spark SQL, Scala

Database

MS SQL SERVER, Azure SQL Database, T-SQL, MySQL, Snowflake, Cosmos DB, Hive, Delta tables, IBM DB2 i series


Professional Summary


  • 6.3+ years of experience in Microsoft Azure cloud platform such as Azure Data Factory, Data Bricks, Data Lake, Azure Synapse analytics, functions, logic Apps, SQL DB, Cosmos DB, PySpark and Python along with Aviation, Oil & Gas, Energy Sector and Pharmaceutical domain knowledge.
  • Creates Several Relational and Non-relational Data Modelling of Data and Creates Prototype diagram using Draw.io.
  • Cleaning and transformation of complex data using data pipelines & notebook in the Azure cloud using Data Factory and Data Bricks with PySpark
  • Perform root cause analysis and resolve the production and data issues.
  • Responsible for design, development, modification, debug and maintenance of Data Pipelines.
  • Delivers technical accountability for team specific work products within an application and provide technical support during solution design for new requirements.
  • Maintains the Existing project along with new development using JIRA in Agile methodology to euiinhance the growth of productivity. 
  • Utilize sound engineering practices to deliver functional, stable, and scalable solutions to new or existing problems.
  • Involves in requirement analysis and business discussions with clients, delivery process, etc. 
  • Excellent interpersonal skills with strong analytical and problem-solving skills. 

Work Experience

Publicis Sapient, April 2023 - Present

Senior Associate Data Engineering L2 (MS Azure Cloud, PySpark)

  • Working on Orx RFP Healthcare & Insurance data of Client. 
  • Design the data model for data migration.
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines through Azure Data factory, Databricks and Kafka streaming/Batch data.
  • Maintains technical documentation using  Confluence platform
  • Manage the projects using Agile Methodology.

Ness Digital Engineering, May 2021 - April-2023

Senior Data Engineer (MS Azure Cloud, PySpark)

  • Works on Clinical and drug trial data of Giant organizations. 
  • Design the non-relational common Data model on Cosmos DB.
  • Analysis of existing relational data for data model creation. 
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines in Azure Data factory, Databricks and Azure Synapse analytics.
  • Maintains technical documentation using Docusaurus and Confluence platform
  • Manage the teams and projects using Agile Methodology along with Client Interaction.

IHS Markit ltd., Oct 2017 - Apr 2021

Data Engineer (MS Azure Cloud, PySparkPython)

  • Extract the complex and Bulk Data related to Aviation, Gas, Energy Sector
  • Investigate issues by reviewing/debugging pipelines, provide fixes, workarounds, and review changes for operability to maintain existing data solutions.
  • Experience of building data pipelines using Databricks (Azure Data Factory and Apache Spark).
  • We extract information by using Python and Kofax RPA tool for fast Crawling.
  • The general orientation of team to reduce manual efforts, growth of products and saving of time during ETL/ELT Process.
  • Have done the integration of third-party tools such as Shutil, Pandas, file transfer from local to azure storage container cloud services using Python.

Project

Healthcare and Insurance Analytics

ORX RFP DMA Explorer Analytics: -

We are building data Application and Framework that contains the data about the Healthcare & Insurance analytics and broadcast the visualizations through power BI reports for business users. Stores the huge data at Lakehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, IBM DB2 i series, Data Lake, Spark SQL, GCP Kafka, Scala, Delta lake & Tables, PySpark and Python.
  • We pull the data from source relational DB such as MySQL and IBM DB2 and move data to Data lake in form of Parquet.
  • Created the Common Data Model using draw.io.
  • Created Azure Data Factory pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Azure Databricks Notebook
  • Using Azure Monitor, Grafana to monitors the ADF Data pipelines and GCP Kafka jobs
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Confluence platform.
  • Generates the Claim, hospitalization and expenses reports using Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Clinical and Drug Trial Analytics

Pharma, Healthcare and Drugs trials: -

We are building Application that contains the data about the Pharma, Healthcare and Drugs trial and broadcast the visualizations through power BI reports. Stores the huge data at data warehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, MS SQL SERVER, Azure Synapse, Data Lake, PySpark and Cosmos DB.
  • We pull the data from relational DB such as MySQL and SQL Server and move data to cosmos DB after the several transformations.
  • Created the Common Data Model using SQL API of Cosmos DB using draw.io.
  • Created Azure Synapse data analytics pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Synapse Notebook
  • Using Azure Monitor, New Relic & analytics to monitors the Synapse Data pipelines
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Docusaurus and Confluence platform.
  • Generates the ingredient, excipients, numerator devices, dosage, artifact, sub artifact, products reports using Snowflake & Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Energy Analytics

Oil, Gas and Coal, OMDC:-

We are building products that contains the information about the oil and gas prices, tenders bidding, consumption and production data country-wise with other factors included.
  • Used various technologies like Python, Azure Database, Data Factory, Data Lake, Data bricks, PySpark, T-SQL, Pandas and Power BI
  • We crawl Complex data from Business, external resources and websites using Python and dump the Files into Azure blobs and data lake.
  • Created Azure Data Factory data pipelines, Activities, Linked services, IR, Triggers for process of ETL/ELT.
  • Write Azure Functions to Implement custom logics for transformation and automation in Python scripts.
  • Using Azure Monitor & analytics to monitors the ADF pipelines. Implemented CI/CD for moving pipelines/Scripts from one environment to another by ARM templates using Azure DevOps.
  • Generates the prices, production, consumption comparison reports using Power BI.

Aviation, IHS Markit Ltd.

Cargo & Flight BI: -

We are having the several pipeline which populates the data of Cargo, Shipment, Booking etc. for multiple Marts. On basis of this, Daily, weekly and monthly Report are generated.
  • Performed several transformations, structuring and cleansing on data including various transformations using Pyspark, Spark SQL and Delta tables.
  • Built multiple data pipelines and job clusters using Azure Data Factory and Databricks.
  • Handling of data on basis of refresh date and SQP date for Incremental load.
  • Worked in Agile methodology using JIRA
  • Highly proficient in using Spark-SQL for developing complex joins and aggregations.
  • Hands on experience on Synapse data warehousing of External tables through Parquet files in Data Lake.
  • Azure Data factory, Azure Data-lake, Azure Databricks, Delta Table, DevOps, Pyspark and Delta Tables. 
  • Cleaning of Cargo & flight data and movement from MS SQL, Hive, traditional Hadoop system and SFTP to Azure Data-lake and Delta tables in Databricks Using Azure Data factory pipelines and Databricks notebook.

Certifications

  • Certification in Azure Data Fundamentals from Microsoft Azure.    
  • Certification in Databricks Certified Data Engineer Associate from Databricks.
  • Certification in Databricks Certified Apache Spark Developer Associate 3.0 from Databricks.
  • Certification in Databricks Accredited Lakehouse fundamentals from Databricks. 
  • Certification In Master Data Analysis with Python - Intro To pandas from Udemy.

Rewards and Achievement

  • Achieved Team player award for Pharma & Clinical project in Q3, 2021.
  • Achieved Best performance Award for Energy analytics project in Q2, 2020.
  • Achieved Peer Award for Optimization of Parts intelligence pipelines Q3, 2019.
  • Achieved Rewards as Team player for Energy Analytics projects in Q4, 2018.   

Education

 B.E. in Electronics Engineering - 73.46% (2016) 
 Institute of Engineering, JIWAJI University, Gwalior 

Core Skills & Strengths 

 ● Team Management          ● Leadership Quality 

 ● Passionate and Creative  ● Quick Learner  

 ● Positive Thinking               ● Punctual 

 ● Motivated                           ● Flexible

Area of Interest 

 ● Interacting with people.  

 ● Willingness to learn new skills.

 ● Cooking                             

 ● Chess

简历
个人档案

Subham Sahu

Seeking a challenging environment that encourages learning, creativity, provides exposures to new ideas and stimulates personal and professional growth along with organizational growth. 

[email protected]

+91- 9039347186

HSR Layout Sector 5, Bangalore, Karnataka, 560034

Technical Skills


Cloud Skills

MS Azure, Azure Data Factory, Data Lake, Azure Dev-Ops, Azure Synapse Analytics, Azure Data bricks, ETL/ELT, Storage Blob, Azure function, logic apps, Delta lake, Kafka, Grafana, Streaming Data

Programming 

Python, Pandas, PySpark, Beautiful Soup
Spark SQL, Scala

Database

MS SQL SERVER, Azure SQL Database, T-SQL, MySQL, Snowflake, Cosmos DB, Hive, Delta tables, IBM DB2 i series


Professional Summary


  • 6.3+ years of experience in Microsoft Azure cloud platform such as Azure Data Factory, Data Bricks, Data Lake, Azure Synapse analytics, functions, logic Apps, SQL DB, Cosmos DB, PySpark and Python along with Aviation, Oil & Gas, Energy Sector and Pharmaceutical domain knowledge.
  • Creates Several Relational and Non-relational Data Modelling of Data and Creates Prototype diagram using Draw.io.
  • Cleaning and transformation of complex data using data pipelines & notebook in the Azure cloud using Data Factory and Data Bricks with PySpark
  • Perform root cause analysis and resolve the production and data issues.
  • Responsible for design, development, modification, debug and maintenance of Data Pipelines.
  • Delivers technical accountability for team specific work products within an application and provide technical support during solution design for new requirements.
  • Maintains the Existing project along with new development using JIRA in Agile methodology to euiinhance the growth of productivity. 
  • Utilize sound engineering practices to deliver functional, stable, and scalable solutions to new or existing problems.
  • Involves in requirement analysis and business discussions with clients, delivery process, etc. 
  • Excellent interpersonal skills with strong analytical and problem-solving skills. 

Work Experience

Publicis Sapient, April 2023 - Present

Senior Associate Data Engineering L2 (MS Azure Cloud, PySpark)

  • Working on Orx RFP Healthcare & Insurance data of Client. 
  • Design the data model for data migration.
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines through Azure Data factory, Databricks and Kafka streaming/Batch data.
  • Maintains technical documentation using  Confluence platform
  • Manage the projects using Agile Methodology.

Ness Digital Engineering, May 2021 - April-2023

Senior Data Engineer (MS Azure Cloud, PySpark)

  • Works on Clinical and drug trial data of Giant organizations. 
  • Design the non-relational common Data model on Cosmos DB.
  • Analysis of existing relational data for data model creation. 
  • Design and Implementation of the Data ingestion using data pipelines and transformation pipelines in Azure Data factory, Databricks and Azure Synapse analytics.
  • Maintains technical documentation using Docusaurus and Confluence platform
  • Manage the teams and projects using Agile Methodology along with Client Interaction.

IHS Markit ltd., Oct 2017 - Apr 2021

Data Engineer (MS Azure Cloud, PySparkPython)

  • Extract the complex and Bulk Data related to Aviation, Gas, Energy Sector
  • Investigate issues by reviewing/debugging pipelines, provide fixes, workarounds, and review changes for operability to maintain existing data solutions.
  • Experience of building data pipelines using Databricks (Azure Data Factory and Apache Spark).
  • We extract information by using Python and Kofax RPA tool for fast Crawling.
  • The general orientation of team to reduce manual efforts, growth of products and saving of time during ETL/ELT Process.
  • Have done the integration of third-party tools such as Shutil, Pandas, file transfer from local to azure storage container cloud services using Python.

Project

Healthcare and Insurance Analytics

ORX RFP DMA Explorer Analytics: -

We are building data Application and Framework that contains the data about the Healthcare & Insurance analytics and broadcast the visualizations through power BI reports for business users. Stores the huge data at Lakehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, IBM DB2 i series, Data Lake, Spark SQL, GCP Kafka, Scala, Delta lake & Tables, PySpark and Python.
  • We pull the data from source relational DB such as MySQL and IBM DB2 and move data to Data lake in form of Parquet.
  • Created the Common Data Model using draw.io.
  • Created Azure Data Factory pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Azure Databricks Notebook
  • Using Azure Monitor, Grafana to monitors the ADF Data pipelines and GCP Kafka jobs
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Confluence platform.
  • Generates the Claim, hospitalization and expenses reports using Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Clinical and Drug Trial Analytics

Pharma, Healthcare and Drugs trials: -

We are building Application that contains the data about the Pharma, Healthcare and Drugs trial and broadcast the visualizations through power BI reports. Stores the huge data at data warehouse level and process the data for some ML application.  
  • Used various technologies like MySQL, MS SQL SERVER, Azure Synapse, Data Lake, PySpark and Cosmos DB.
  • We pull the data from relational DB such as MySQL and SQL Server and move data to cosmos DB after the several transformations.
  • Created the Common Data Model using SQL API of Cosmos DB using draw.io.
  • Created Azure Synapse data analytics pipelines for process of ETL\ELT.
  • Implement custom logics for transformation and automation in Synapse Notebook
  • Using Azure Monitor, New Relic & analytics to monitors the Synapse Data pipelines
  • Implemented CI/CD for moving pipelines/Scripts from one environment to another by Repos branch strategy using Azure DevOps
  • Manage the sprint planning, Backlog refinement and retrospectives Using JIRA in Agile methodology.
  • Maintained the documentation on Docusaurus and Confluence platform.
  • Generates the ingredient, excipients, numerator devices, dosage, artifact, sub artifact, products reports using Snowflake & Power BI.
  • Published the reports sharing the pbix file on product portal for Business purpose.

Energy Analytics

Oil, Gas and Coal, OMDC:-

We are building products that contains the information about the oil and gas prices, tenders bidding, consumption and production data country-wise with other factors included.
  • Used various technologies like Python, Azure Database, Data Factory, Data Lake, Data bricks, PySpark, T-SQL, Pandas and Power BI
  • We crawl Complex data from Business, external resources and websites using Python and dump the Files into Azure blobs and data lake.
  • Created Azure Data Factory data pipelines, Activities, Linked services, IR, Triggers for process of ETL/ELT.
  • Write Azure Functions to Implement custom logics for transformation and automation in Python scripts.
  • Using Azure Monitor & analytics to monitors the ADF pipelines. Implemented CI/CD for moving pipelines/Scripts from one environment to another by ARM templates using Azure DevOps.
  • Generates the prices, production, consumption comparison reports using Power BI.

Aviation, IHS Markit Ltd.

Cargo & Flight BI: -

We are having the several pipeline which populates the data of Cargo, Shipment, Booking etc. for multiple Marts. On basis of this, Daily, weekly and monthly Report are generated.
  • Performed several transformations, structuring and cleansing on data including various transformations using Pyspark, Spark SQL and Delta tables.
  • Built multiple data pipelines and job clusters using Azure Data Factory and Databricks.
  • Handling of data on basis of refresh date and SQP date for Incremental load.
  • Worked in Agile methodology using JIRA
  • Highly proficient in using Spark-SQL for developing complex joins and aggregations.
  • Hands on experience on Synapse data warehousing of External tables through Parquet files in Data Lake.
  • Azure Data factory, Azure Data-lake, Azure Databricks, Delta Table, DevOps, Pyspark and Delta Tables. 
  • Cleaning of Cargo & flight data and movement from MS SQL, Hive, traditional Hadoop system and SFTP to Azure Data-lake and Delta tables in Databricks Using Azure Data factory pipelines and Databricks notebook.

Certifications

  • Certification in Azure Data Fundamentals from Microsoft Azure.    
  • Certification in Databricks Certified Data Engineer Associate from Databricks.
  • Certification in Databricks Certified Apache Spark Developer Associate 3.0 from Databricks.
  • Certification in Databricks Accredited Lakehouse fundamentals from Databricks. 
  • Certification In Master Data Analysis with Python - Intro To pandas from Udemy.

Rewards and Achievement

  • Achieved Team player award for Pharma & Clinical project in Q3, 2021.
  • Achieved Best performance Award for Energy analytics project in Q2, 2020.
  • Achieved Peer Award for Optimization of Parts intelligence pipelines Q3, 2019.
  • Achieved Rewards as Team player for Energy Analytics projects in Q4, 2018.   

Education

 B.E. in Electronics Engineering - 73.46% (2016) 
 Institute of Engineering, JIWAJI University, Gwalior 

Core Skills & Strengths 

 ● Team Management          ● Leadership Quality 

 ● Passionate and Creative  ● Quick Learner  

 ● Positive Thinking               ● Punctual 

 ● Motivated                           ● Flexible

Area of Interest 

 ● Interacting with people.  

 ● Willingness to learn new skills.

 ● Cooking                             

 ● Chess