• Файл

Bohdan

Data engineer

Розглядає посади: Data engineer, Big data engineer
Вік: 24 роки
Місто проживання: Харків
Готовий працювати: Дистанційно, Дніпро, Київ, Львів, Одеса, Харків
Розглядає посади:
Data engineer, Big data engineer
Вік:
24 роки
Місто проживання:
Харків
Готовий працювати:
Дистанційно, Дніпро, Київ, Львів, Одеса, Харків

Контактна інформація

Шукач вказав телефон та ел. пошту.

Прізвище, контакти та світлина доступні тільки для зареєстрованих роботодавців. Щоб отримати доступ до особистих даних кандидатів, увійдіть як роботодавець або зареєструйтеся.

Завантажений файл

Файл містить ще 2 сторінки

Версія для швидкого перегляду

Це резюме розміщено у вигляді файлу. Ця версія для швидкого перегляду може бути гіршою за оригінал резюме.

Bohdan Kasumov
Date of birth: 20/04/2001 Nationality: Ukrainian Gender: Male Phone number:

[відкрити контакти](див. вище в блоці «контактна інформація») (Mobile) Email address: [відкрити контакти](див. вище в блоці «контактна інформація»)

ABOUT ME

Learner: striving for ongoing training, easy to digest new information, track, new technologies and apply
them in their work.
Teamwork: the development of modular software, ready to work in a team, polite, calm, tolerant.
Systems thinking: I analyze the tasks in order to find the optimal solutions to the problem. Determination of
cause-and-effect relationships, key indicators that determine the state of the problem or affect the result.

SKILLS

Database

MongoDB MySQL Relational databases Hive NoSQL DynamoDB Oracle PostgreSQL Azure
Data Lake Storage Gen2 Amazon RDS Redis Amazon Redshift MS SQL

Programming languages

Python Java Scala

Development environments

MySQL WorkBench SQL Server Management Studio Visual Studio Code PyCharm IntelliJ IDEA
JupyterLab

Technology and knowledge

Spring Spring Boot Hibernate Apache Spark JDBC RESTful Web Services Apache Kafka
SonarQube Agile (Scrum) Jenkins JSON Spark Streaming Git Zookeeper Docker
Kubernates Confluence Apache Avro Schema Registry Bamboo SOAP Code Review
Software Engineering Jira Continous Integration Kafka Stream XML Spring Streaming UML
HTML Unit Testing Web Development Github Bitbucket CSS HTML5 Hadoop Tomcat
TCP HTTP Mathematics Data Lake Data Warehouse Git Actions Regex Serialization
YAML Lombok Log4j/SLF4J Swagger Flask Machine learning Delta Lake

Design Principles and Patterns

SOLID MVC ETL RESTful api ACID Design Patterns REST Microservices Web API Object-
Oriented Programming API Integrations Clean Code MapReduce

Build and Dependency Management

Maven Gradle

Tools

Postman Insomnia Slack JMeter Confluent Control Center Power BI Tableau Grafana dbt

Cloud

Amazon s3 AWS Lambda AWS EC2 AWS Azure Databricks Google BigQuery

Data Orchestrator

AWS Glue Apache Airflow Azure Data Factory
LANGUAGE SKILLS

Mother tongue(s): UKRAINIAN SIGN LANGUAGE
Other language(s):

UNDERSTANDING SPEAKING WRITING
Spoken Spoken
Listening Reading
production interaction
ENGLISH B2 B2 B2 B2 B2

UKRAINIAN C2 C2 C2 C2 C2

RUSSIAN C1 C1 C1 C1 C1

Levels: A1 and A2: Basic user; B1 and B2: Independent user; C1 and C2: Proficient user

EDUCATION AND TRAINING

01/09/2016 – 02/07/2020 Kharkiv, Ukraine
JUNIOR SPECIALIST TECHNICIAN-PROGRAMMER Kharkiv patent-computer college

specialty "Software engineering"
Main areas:
• Develop algorithms for solving complex problems;
• Design databases of varying degrees of complexity;
• Develop web applications (sites);
• Create software products using client-server technologies;
• Develop graphical user interfaces;
• Modify and test software products and information resources;
• Design software products and use design patterns.

Website https://khpcc.com/

National classification Higher education institution of II level of accreditation of public ownership form

01/09/2020 – 30/06/2023 Kharkiv, Ukraine
BACHELOR'S DEGREE Kharkiv National University of Radio Electronics

Specialty "Computer science"
skills:
• Design, creation and re-engineering of folding software and information systems based on the
development of modern standards, technologies, flexible methodologies, programming languages (C,
C ++, C #, Java ...), methods and mechanisms of command work.
• Development of computer games, virtual and augmented reality (logic, plot, graphics).
• Design, development, implementation and administration of databases of various structures, both
relational and document-oriented (Oracle, MySQL, MongoDB) with the help of superiority, efficiency,
and scalability.
• Software testing.

Website https://nure.ua/

National classification Higher education institution of IV level of accreditation of public ownership form

Thesis Development of CRM system components for shoe sales

01/08/2023 – 01/06/2025 Kharkiv, Ukraine
MASTER'S DEGREE Kharkiv National University of Radio Electronics

Specialty "Computer science"
Educational program: "Systems design"
skills:
• Understanding the theoretical foundations of computer science.
• The ability to formalize the subject area of a certain project in the form of corresponding information
model..
• Ability to use mathematical methods for analysis formalized models of the subject area.
• The ability to collect and analyze data (including large ones), for ensuring the quality of project
decision-making.
• Ability to develop, describe, analyze and optimize architectural solutions of various information and
computer systems appointment.
• Ability to apply existing and develop new algorithms solving problems in the field of computer
science.

Website https://nure.ua/

National classification Higher education institution of IV level of accreditation of public ownership form

Thesis Research on Clustering Methods for Implementing a Recommendation Function Based on Collaborative
Filtering in Movie Viewing Systems

WORK EXPERIENCE

 E-COMMERCE
DATA ENGINEER – 04/2025 – CURRENT
Worked in the e-commerce system data team for E-commerce system.
Responsibilities as a developer:
• Designed the DWH architecture, including Raw, Stage, and Meta layers, to ensure efficient data
storage and processing;
• Developed and implemented ETL pipelines using Apache Airflow to extract data from 1C (MS SQL),
transform it, and load it into PostgreSQL;
• Created DAGs in Airflow to automate and schedule ETL operations, ensuring reliable and scalable
data workflows;
• Integrated Apache Kafka for real-time data streaming and processing between 1C and PostgreSQL;
• Stored and retrieved processed data in AWS S3 buckets for scalability and backup purposes;
• Performed data transformations and validations using Python and pandas to ensure data integrity
during migration;
• Configured ClickHouse for analytical queries and optimized performance for large-scale data
processing;
• Connected Power BI to the PostgreSQL DWH for data visualization and reporting;
• Conducted testing and validation of migrated data using SQL Server Management Studio and
PostgreSQL tools to ensure accuracy and consistency.
Stack: Python, Airflow, Fast API, PostgreSQL, ClickHouse, Power BI, 1C, MS SQL, Sql Server Management
Studio, pandas, AWS, Apache Kafka, AWS s3.

DATA ENGINEER – 04/2024 – 01/2025
Worked in the e-commerce system development team for data analysis movie viewing system.
Responsibilities as a developer:
• Investigated existing approaches to generating recommendations in movie viewing systems, which
allowed us to justify the choice of the Item-Based Collaborative Filtering method as the basic one;
• Proposed a modified version of the collaborative filtering method, which takes into account not only
user ratings, but also sets of movie genres, converted into vector form using a model based on
Word2Vec neural networks;
• Proposed the Kalinsky-Kharabash index, the Davis-Boulden index, the silhouette score, and the
execution time as indicators of clustering quality;
• Conducted a study of clustering of film data using the algorithms K-means, DBSCAN, OPTICS,
HDBSCAN, ST-DBSCAN. HDBSCAN was chosen to implement the recommendation function in the
system;
• Conducted an experimental study of the modified joint filtering method. The results confirmed that
taking into account genre sets increases the accuracy of recommendations by 10-15%;
• Developed a web application architecture that includes saving ratings in the database, clustering films
by genre, and generating personal recommendations for the user;
Stack: Python, pandas, sklearn (DBSCAN KMeans, davies_bouldin_score, OPTICS, ST_DBSCAN), seaborn,
gensim.

DATA ENGINEER – 11/2024 – 01/2025
Worked in the e-commerce system development team for data analysis.
Responsibilities as a developer:
• Checked the data for empty values, i.e. NaN. The empty data was filled using special tools;
• Proposed and implemented to assess the statistical significance of differences between periods of
work for each employee position, the Kruskal-Wallis Criteria were selected, due to the large difference
in variances of the selected groups. The result of the Kruskal-Wallis test showed that the level of
maturity significantly differs between teams;
• Constructed a boxplot diagram to visualize statistically significant differences.
Stack: Python, pandas, ANOVA, numpy, seaborn, matplotlib.scipy, glob.

DATA ENGINEER – 06/2024 – 11/2024
Worked in the e-commerce system development team for processing sales orders.
Responsibilities as a developer:
• Designed the DWH architecture with Bronze, Silver, and Gold layers to ensure structured data storage
and processing across PostgreSQL and Google BigQuery;
• Developed ETL pipelines using dbt and Azure Data Factory to extract data from Redis, transform it,
and load it into PostgreSQL and AWS Redshift;
• Implemented Apache Kafka to stream data from Redis to the DWH, enabling real-time data ingestion;
• Performed data transformations from the MySQL-based data lake using Python and Delta Lake within
Azure Databricks to ensure data consistency and quality;
• Created and managed DAGs in Azure Data Factory and dbt to automate and schedule ETL operations,
ensuring scalable and reliable data workflows;
• Built Flask-based APIs to facilitate data interactions and integration with downstream systems;
• Applied machine learning techniques within Azure Databricks to enhance data transformation
processes, improving data quality and predictive capabilities;
• Stored and managed processed data in Delta Lake for versioning and scalability;
• Integrated Power BI with Azure and Google BigQuery for data visualization and reporting, delivering
actionable insights from the Gold layer.
Stack: Python, Azure, Google BigQuery, Databricks, dbt, MySQL, PostgreSQL, Azure Data Factory, Flask,
Machine Learning, Power BI, Delta Lake

DATA ENGINEER – 01/2023 – 06/2024
Worked in the e-commerce system development team for processing sales orders.
Responsibilities as a developer:
• Designed the DWH architecture with Bronze, Silver, and Gold layers to ensure structured data
processing and storage in PostgreSQL and AWS Redshift;
• Developed ETL pipelines using AWS Glue to extract data from Redis, transform it, and load it into
PostgreSQL and AWS Redshift;
• Implemented Apache Kafka to read and stream data from Redis to the DWH, enabling real-time data
ingestion;
• Performed data transformations from the Redis-based data lake using Python and pandas to ensure
data consistency and quality;
• Created and managed DAGs in AWS Glue and dbt to automate and schedule ETL operations, ensuring
scalable and reliable data workflows;
• Utilized AWS Lambda for serverless processing to handle specific transformation tasks and
orchestrate ETL pipelines;
• Stored and managed processed data in AWS S3 buckets for scalability, archival, and integration with
AWS Redshift;
• Configured AWS RDS (PostgreSQL) for transactional data storage and AWS Redshift for analytical
queries and large-scale data processing;
• Integrated Tableau with AWS Redshift for data visualization and reporting, enabling business insights
from the Gold layer;
• Conducted testing and validation of migrated data using PostgreSQL tools to ensure accuracy and
integrity.
Stack: Python, AWS Glue, PostgreSQL, Tableau, dbt, pandas, AWS, Apache Kafka, AWS S3, AWS Lambda, AWS
RDS, AWS Redshift, Redis

DATA ENGINEER – 09/2022 – 01/2023
Worked in the AI system development team for recognizing people at pedestrian crossings.
Responsibilities as a developer:
• Implemented a system that allows using a trained neural network to recognize pedestrians crossing
the road.
Stack: Python, OpenCV, scikit-image, TensorFlow, PyTorch, Keras, scikit-learn, NumPy, Pandas, SOLID.

 EPAM SYSTEMS – KHARKIV, UKRAINE
BIG DATA SOFTWARE ENGINEER – 03/2022 – 09/2022
Worked in the Big Data development team for streaming solutions for processing large volumes of data
from 8 million records.
Responsibilities as a developer:
• Created and modified of Databricks notebook;
• Proposed and created several auxiliary pipelines for processing individual types of data, which
reduced the overall execution time by 8%;
• Created and modified of pipeline for Azure Data Factory;
• Created an uber JAR to run a job on Databricks;
• Resolved data parsing error;
• Used Apache Spark to implement distributed processing of unstructured and semi-structured data
(using Java and Scala);
• Improved test notebook for Product Owner;
• Created and conducted a demo for implemented tasks for the Client;
• Refactoring code.
Stack: Scala, Java, Azure, Python, Apache Spark, Databricks, Azure Data Factory, Bamboo, SonarQube, Data
lake gen 2 storage.

BIG DATA SOFTWARE ENGINEER – 09/2021 – 03/2022
Distributed Data Processing with Scala and Apache Spark.
Responsibilities as a developer:
• Developed a Scala-based application leveraging Scala Core to explore functional programming
paradigms, optimizing data transformation pipelines for improved performance and code
maintainability;
• Implemented an event-driven processing system using Scala Akka, enabling concurrent and fault-
tolerant handling of data streams with low-latency message passing;
• Designed and integrated Akka actors with Apache Spark to orchestrate distributed tasks, improving
system resilience and scalability for real-time data processing workflows;
• Created unit tests and integration tests using ScalaTest to validate Spark jobs and Akka actor systems,
ensuring robust and reliable data processing pipelines.
Stack: Scala, Scala Core, Akka, Apache Spark, ScalaTest

SOFTWARE ENGINEER – 03/2021 – 09/2021
Worked in the Big Data development team for streaming solutions for processing large volumes of data
from 6 million records.
Responsibilities as a developer:
• Implemented the eventing framework(Kafka clients) using Java technology stack;
• Integrated Confluent Schema Registry into Spring Boot Aplication to work with schemas;
• Implemented handlers for failed messages which accelerated the search for errors;
• Worked with Confluent Control Center;
• Created Get Started Guide for Teams;
• Set up Kafka locally for debug application;
• Deployed AWS DynamoDB in Docker instead file saved on datalake;
• Proposed and implemented the launch of test scenarios in a Docker container with result output at
Java report;
• Code review teammate.
Stack: Java, Apache Kafka, Spring, Maven, Streaming, Yaml, Docker, AWS DynamoDB, Confluent Control
Center, Schema Registry, Insomnia, Kafka client.

SOFTWARE ENGINEER – 10/2020 – 03/2021
Worked in the Big Data development team for solutions for processing large volumes of data from 10
million records.
Responsibilities as a developer:
• Ran Java Spark application on Hadoop Hortonworks sandbox;
• Saved and loaded data from HDFS as JSON, AVRO, CSV and Parquet file;
• Implemented Spark Streaming application with watermarking function;
• Implemented the eventing framework(Kafka clients) using Java technology stack and reduced the
sending time by 10%;
• Used Apache Spark to implemented distributed processing of unstructured and semi-structured data;
• Ran Java Spark application on Amazon EMR and save result at S3 bucket.
Stack: Java, Apache Spark, Hadoop Hortonworks, VirtualBox, Kafka, HDFS, Hive, AWS, EMR, S3, Maven,
MapReduce, Streaming

JUNIOR SOFTWARE ENGINEER – 08/2019 – 10/2020
Worked in the e-commerce system development team, an internal company project for ordering food.
Responsibilities as a developer:
• Created Java web-application with servlets;
• Created async query with using jQuery for calculation order price and reduced the calculation time
from 3 to 1 second;
• Implemented custom captcha for bot-checking;
• Saved data with MySQL Server;
• Implemented new transaction manager and and accelerated the execution of transactions by 30%;
• Implemented localization with properties files;
• Initiated the addition of a gift card to the system for users, which allows them to receive a flat or
percentage discount on goods;
• Updated framework versions, resolve conflict and refactoring code for new version;
• Incremented test cover from 60% to 70% with Grovy unit testig;
• Implemented user interface with jsp and TypeScript.
Stack: Java, MySQL, HTML, CSS JQuerry, Spring (Data, Security, JPA, WEB, Boot), Hibernate, TypeScipt, REST,
SOLID, GitLab, JMeter, Maven, Bamboo

Інші резюме цього кандидата

Схожі кандидати

Усі схожі кандидати


Порівняйте свої вимоги та зарплату з вакансіями інших підприємств: