Nothing Special   »   [go: up one dir, main page]

0% found this document useful (0 votes)
16 views4 pages

Dimple Dhok Resume

Download as pdf or txt
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 4

Dimple Dhok

dimpledhok@gmail.com | +91 8983688745


linkedin.com/in/dimple-dhok | Nagpur, Maharashtra

WORK EXPERIENCE

Engineer in Delivery and Practices June 2020 - Present


Konverge.ai | Nagpur, India
● 3.8 years of experience in Data Science & Engineer, and a wealth of knowledge in Python, API
Development, and ETL Tools, as well as leveraging expertise to deliver data-driven solutions
tailored to specific needs.

PROJECTS

1).Project 1->Cloud based

Developed an automated, scalable, and resilient data pipeline for seamlessly migrating
on-premise data to the cloud. Transformed and stored the data in a cloud-based SQL
database, ensuring real-time accessibility for further analysis.
● Set up an Azure pipeline for on-premise and cloud integration using Integration
Runtime.
● Moved on-premise files to the cloud through Azure Data Factory, employing a
copy activity with a linked service to store raw data in a storage container.
● Processed unstructured data with Azure Databricks Notebook and saved the
output.Set up a logic app for pipeline failure notifications.

2).Project 2-(Health Care)

The project successfully consolidated cannabis-related government data from


various U.S. states into a single application, offering users comprehensive information
on one platform.

● Checked websites for scraping, addressing security and authentication concerns.


● Effectively gathered data from multiple websites.
● Automated web data extraction by navigating internal pages. Extracted
information from CSV sheets and saved it in MongoDB.
3).Project 3-(Database related)

The project was divided into two phases. In the first phase, public profile data from
LinkedIn was scraped and then recorded in a database. The second phase included
scraping public profile URLs from the Sales Navigator.

● Retrieved public profile data through web scraping.


● Automated the scraping process using Selenium.
● Incorporated the nohup process to concurrently execute two processes.
● Managed data in a MongoDB database and utilized Flask to create APIs.

4).Project 4-(Data science)

The project involved creating a Risk Quantification Engine that calculated the risk for all
nodes at the FAIR level, utilizing a Bayesian network.

● Researched efficient libraries for Bayesian networks, devised methods for


calculations at all levels, and managed data creation and storage in MongoDB.
● Built APIs using Flask, integrated SonarQube with the project, and authored test
cases to ensure comprehensive code coverage reflected on SonarQube.

5).Project 5-

Description:- (Data science)

Our project focused on reducing fuel and tire management costs in fleet operations. By
leveraging data-driven insights and targeted measures, we aimed to optimize these
expenses, enhancing overall cost efficiency and financial performance for the fleet.

● Collected relevant data from Palantir Foundry or other databases, then filtered to
remove unnecessary or redundant information.
● Conducted data sampling and preprocessing by selecting a representative sample
for prediction. Implemented necessary preprocessing steps to ready the data for
analysis and predictive modeling.
6).Project 6-

Description:- (Data Engineering)

The objective was to establish a complete data pipeline for a product, simplifying the
report creation process by monitoring a folder, detecting changes in files and data, and
generating Directed Acyclic Graphs (DAGs) using Apache Airflow.

● Designed a data pipeline utilizing Airflow to reduce manual intervention in the


report creation process. Introduced a file watcher functionality to alert users of
specified events in various scenarios.
● Explored Python libraries for monitoring file modifications.
● Initiated the file watcher functionality using the FastAPI framework.
● Recorded the modified event for the specified file, using the Dag ID associated
with that file to trigger the Airflow Dag.
● Generated a new report based on the triggered Dag and added it in a table
(Postgres).

7).Project 7-

Description:-(Data Migration)

The project's objective was to migrate from an on-premise infrastructure to a


cloud-based solution. The tasks undertaken and achievements include:

● In Databricks, I evaluated the functionality of shell code compared to Python.


Conducted end-to-end testing for all translated code using real data. Translated
shell code to Python and Perl scripts were converted to Python.
● Converted Scala code to a PySpark-compatible format and tested SQL scripts for
data extraction and loading during the project.
● Addressed production issues and applied necessary bug fixes.
EDUCATION

Bachelor of Engineering (B. E.) - Computer Science Engineering JUNE, 2015 - JUNE, 2019

SKILLS

Python | SQL server | Azure Blob storage| Azure Data Factory | Databricks | Selenium | MongoDB
| Agenarisk | Postgres | Linux | Airflow | FLASK | Microservice | Docker | GitHub Actions | Cloud |
Azure | FastAPI | Scripting | Linux | Logging | Monitoring | Project Management | Leadership |
Teamwork | Django | Development |

CERTIFICATION

-M220P: MongoDB for Python Developers


-Basic Python for beginners. Data structure and Algorithm for a Python developer.
-Selenium Webdriver with PYTHON from Scratch + Frameworks(udemy course)

You might also like