Results-driven Senior Data Engineer
with over 5 years of experience designing and implementing scalable, cloud-native data platforms. Specialized in Databricks, Delta Lake, Apache Kafka, AWS, and PySpark. Proven ability to deliver secure, high-performance data pipelines and frameworks using modern architectures like Lambda, Medallion, and stream-batch integration. Focused on automation, optimization, and data governance.
Technical Skills & Experience:
- Delivered complex, production-grade data migration and cloud modernization initiatives using Databricks, AWS Glue, Delta Lake, and Python-based frameworks.
- Built enterprise-scale ingestion systems with modular templates, YAML configurations, and dynamic orchestration using GitHub Actions and Databricks Asset Bundles.
- Strong hands-on expertise in Python, PySpark, Bash, T-SQL, with deep focus on scalable data transformation and automation-first design.
- Architected stream-batch hybrid pipelines integrating Kafka, Auto Loader, and CDF (Change Data Feed) for near real-time data processing and analytics.
- Developed robust data quality frameworks, SCD2 logic, CI/CD pipelines, and secure data governance layers (Unity Catalog, ACLs, masking).
- End-to-end ownership of the data lifecycle, from extraction and ingestion to historical modeling and business-ready output layers.
Selected Achievements:
- Reduced daily ETL processing time by 80% by implementing incremental data loading, dynamic partitioning, parallel loading, Delta Lake compaction, and other optimization strategies.
- Enabled data ingestion across 10+ business units by standardizing reusable frameworks and YAML-driven templates for source-agnostic ingestion.
- Deployed CI/CD automation using GitHub Actions and achieved 99.9% job reliability across environments.
Industry & Project Background:
- Successfully delivered data solutions across sectors including retail banking, heavy industry, and billing.
- Familiar with retail replenishment systems, cloud migration, and real-time analytics platforms.
Teamwork & Work Style:
- Comfortable working both independently and as part of distributed global teams.
- Experienced with Agile methodologies including Scrum and Kanban.
Education:
- Bachelor’s Degree, Information Technology, Turan University, Almaty, Kazakhstan;
- Bachelor’s Degree, Petroleum Engineering, Satbayev University, Almaty, Kazakhstan.
Certificates:
- Databricks Certified Data Engineer Professional, Databricks Certification;
- Databricks Certified Associate Developer for Apache Spark 3.0, Databricks Certification;
- Academy Accreditation - Generative AI Fundamentals, Databricks Accreditation;
- SQL (Advanced) Certificate, HackerRank Certification;
- ETL and Data Pipelines with Shell, Airflow, and Kafka, issued by IBM;
- Python and Pandas for Data Engineering, issued by Duke University.