Essential Skills
- Establish import/export patterns, including handling data extracts, schema discovery, incremental loads, and multiple source instances
- Expertise in data‑transformation‑heavy pipelines: profiling, cleansing, standardisation, conformance, and publishing
- Advanced SQL, including profiling, joins/merges, deduplication, anomaly detection, and performance tuning
- Python scripting for automation, parsing, rules engines, anddata‑quality checks; ability to write maintainable code
- Experience with Python packages such as Pandas/Polars, scikit‑learn, matplotlib
- Experience with modern data tooling (e.g., Spark, cloud data‑integration tooling) or the ability to implement equivalent functionality in code
- Proven experience with geospatial data, including vector/raster formats, GeoJSON, shapefiles, CRS handling, and spatial‑analysis workflows
- Ability to interpret and apply geographical context in pipelines, including aggregating or translating local/regional geospatial insights into broader‑level datasets
- Experience using publicly available official datasets (e.g., census boundaries, geographic lookups, deprivation indices, population estimates)
- Ability to build rules for completeness/validity/consistency, implement exception handling, and support reconciliation flows
- Ability to build versioned, deterministic pipelines with logging, lineage, and traceability
- Comfortable working in a secure environment with least‑privilege access and handling sensitive datasets
Soft Skills
- Strong communication and collaboration skills
- Ability to work effectively with diverse team members and stakeholders
- Enthusiastic, engaged, and positive attitude
- Proactive, self‑driven, comfortable working without step‑by‑step direction
- Ability to handle ambiguity and changing requirements
Please send cv's to a.otoole@tenthrevolution.com