Welcome
I build pragmatic data and software solutions across parsing standards, analytics enablement, and automation. I turn messy operational/log data into structured, queryable, value‑producing datasets—whether that’s air cargo message streams, platform usage telemetry, or API‑sourced operational data.
- Small composable tools, clear data contracts, automation first
- Predictable pipelines, repeatable builds, strong observability
Quick links
- Posts → /posts/
- Projects → /projects/
- About → /about/
- CV → /cv/
Featured projects
- CargoIMP Spark Parser — High‑performance Cargo‑IMP parsing into Spark DataFrames (Scala, Spark)
- Fantastic Spork — Catalyst‑native text & counting expressions for Spark (Scala, Spark SQL)
- Sparqlin — Structured Spark SQL job framework for Databricks (Spark, PySpark, YAML)
- RedactifyAI — AI PII detection + anonymization at scale (Python, Spark, Presidio)
- Tableau Usage → Databricks — Usage telemetry ingestion + modeling (.NET, Parquet, Databricks)
- SafetyCulture → Databricks — API → Delta ingestion automation (.NET, Delta Lake)
Explore all on the Projects page.
Latest posts
- Pre‑sales Patterns for Databricks Solutions Engineers
- RAG on Databricks: Embeddings, Vector Search, and Cost/Latency Tuning
- Databricks MLOps Playbook: From MLflow to Production
- Technology Timeline 2025
- SDE Interest Rates — Numerics
- Probabilistic Risk Modeling Platform
- Upstream Evaluation Framework
- Architecture — Petroleum Analytics
- Market Risk — Multicurrency Brent
Contact & Links
- GitLab: https://gitlab.com/rokorolev
- CV: /cv/