👋 Hi!

👋 Hi!

Originally from Kazakhstan 🇰🇿, I have been living and working in Germany 🇩🇪 for the past 10 years, 6 of which in Berlin working for small and medium-sized startups. I consider myself a Berliner 🐻.

In my last projects, I have built near real time data ingestion systems in AWS and GCP for mobile games and developed data pipelines in Airflow, Databricks, and Snowflake. I help companies in building scalable data ingestion solutions, designing ELT/ETL workloads, and analyzing data. I am a 3x AWS Certified:

    
    

Programming languages:

  • Python, SQL, Dart
  • Java, Scala

Data Warehousing:

  • Data warehousing concepts, ETL processes, data modeling, data normalization, data pipelines
  • Cloud-based data warehousing solutions (Amazon Redshift, Google BigQuery, Snowflake)

Big Data Technologies:

  • Apache Spark (Spark SQL, Spark Streaming)
  • NoSQL databases (MongoDB, Cassandra, DynamoDB)
  • Databricks
  • Streaming technologies such as AWS Kinesis, Firehose

Data Visualization:

  • Tableau
  • Looker

Data Governance:

  • Data governance and data security practices
  • Monitoring data quality, tracking data lineage, managing metadata

Cloud Computing:

  • Amazon Web Services (AWS)
  • Google Cloud
  • Microsoft Azure

Data Engineering Tools:

  • Apache Airflow for automating data pipelines
  • dbt (data build tool) for creating data pipelines
  • Docker for creating isolated and portable development environments

Collaboration Tools:

  • JIRA, Git, GitHub, GitLab, Stash
  • Agile/Scrum environments