Location
PolandRate
$65
/ per hour
Years of experience
10About
Big Data Engineer, specializing in developing robust data solutions leveraging Java, Python, and Scala. Proficient in Apache Spark, Kafka, and AWS, the focus is on underpinning innovative analytics platforms. Merging technical knowledge with business acumen, the goal is to deliver data architectures aligned with strategic objectives. Leading data engineers in dynamic environments, commitment lies in fostering collaborative success and staying updated with the latest trends in big data and cloud technologies.Tech Stack
Big Data, Airflow, AWS, Data modelling, ETL, GraphQL, Java, Kafka, MLOps, Python, SAS programming, Scala, Software design, Spark, SQL, TerraformEmployment history
Chapter Lead for Data Development
2022 - Present
- Building a community for data engineers
- Working on best practices, development and training plans
- Heavy weight on coaching and mentoring
- Conducting tech workshops
- Working on best practices, development and training plans
- Heavy weight on coaching and mentoring
- Conducting tech workshops
Staff Big Data Developer
2022 - Present
- Technology Stack: MLOps, Terraform, Airflow, Scala, Java, Python, Graph databases, SQL, ETL, AWS, Spark, Kafka, Data modelling
- Data pipelines
- Developing ETL/ELT pipelines in Spark
- Developing real time pipelines in Kafka
- Process scheduling in Airflow
- Developing custom operators/sensors in Airflow
Data modelling & analysis
- Designing data marts
- Designing global domains for data platform with DDD
- Preparing PowerBI dashboards
- Developing automatic data quality reports
- Creating REST services on AWS
- Creating POCs for emerging tech - e.g. Neo4J, GraphQL, MLFlow
- Close cooperation with product owner for new feature
- Close cooperation with data science teams
- Talent recruitment and team members coaching
- Implementing cross team best practices, tools & frameworks
Example projects :
- Customer service reporting data mart
- Graph based company normalisation pipeline
- GUI for ML model deployment
- Data quality rule based framework
- Framework for speedup of ETL pipeline creation (AWS Glue/Spark)
- Data mart & reporting for Redshift usage
- Data pipelines
- Developing ETL/ELT pipelines in Spark
- Developing real time pipelines in Kafka
- Process scheduling in Airflow
- Developing custom operators/sensors in Airflow
Data modelling & analysis
- Designing data marts
- Designing global domains for data platform with DDD
- Preparing PowerBI dashboards
- Developing automatic data quality reports
- Creating REST services on AWS
- Creating POCs for emerging tech - e.g. Neo4J, GraphQL, MLFlow
- Close cooperation with product owner for new feature
- Close cooperation with data science teams
- Talent recruitment and team members coaching
- Implementing cross team best practices, tools & frameworks
Example projects :
- Customer service reporting data mart
- Graph based company normalisation pipeline
- GUI for ML model deployment
- Data quality rule based framework
- Framework for speedup of ETL pipeline creation (AWS Glue/Spark)
- Data mart & reporting for Redshift usage
Senior Big Data Developer
2019 - 2022
Big Data Developer
2016 - 2019
Junior Big Data Developer
2015 - 2016
Associate Software Engineer
2014 - 2015
Education history
Warsaw University of Technology
2010 - 2014
Bachelor's degree - Electrical and Computer Engineering
We’ve helped 83 clients with IT recruitment and software development.
Read about a few of them below...