Originally from Kazakhstan 🇰🇿, I have been living and working in Germany 🇩🇪 for the past 10 years, 6 of which in Berlin working for small and medium-sized startups. I consider myself a Berliner 🐻.
In my last projects, I have built near real time data ingestion systems in AWS and GCP for mobile games and developed data pipelines in Airflow, Databricks, and Snowflake. I help companies in building scalable data ingestion solutions, designing ELT/ETL workloads, and analyzing data. I am a 3x AWS Certified:
Programming languages:
- Python, SQL, Dart
- Java, Scala
Data Warehousing:
- Data warehousing concepts, ETL processes, data modeling, data normalization, data pipelines
- Cloud-based data warehousing solutions (Amazon Redshift, Google BigQuery, Snowflake)
Big Data Technologies:
- Apache Spark (Spark SQL, Spark Streaming)
- NoSQL databases (MongoDB, Cassandra, DynamoDB)
- Databricks
- Streaming technologies such as AWS Kinesis, Firehose
Data Visualization:
- Tableau
- Looker
Data Governance:
- Data governance and data security practices
- Monitoring data quality, tracking data lineage, managing metadata
Cloud Computing:
- Amazon Web Services (AWS)
- Google Cloud
- Microsoft Azure
Data Engineering Tools:
- Apache Airflow for automating data pipelines
- dbt (data build tool) for creating data pipelines
- Docker for creating isolated and portable development environments
Collaboration Tools:
- JIRA, Git, GitHub, GitLab, Stash
- Agile/Scrum environments