Position: Program Manager- L3 support engineer role (Dremio+Python)
Location: Halifax, NS (Hybrid- 2 days onsite)
Role & Responsibilities
- Engage and partner with data virtualization CTB, RTB, Dremio Vendor, and client applications.
- Analyze and resolve production and non-production issues, analyze product logs and find RCA.
- Identify repetitive issues, raise problem tickets and work upon tactical/strategy fixes.
- Dremio version upgrade in all non-prod instances (standalone as well as containerized)
- Enhancements in data virtualization services (written in python)
- Identify and resolve system vulnerabilities.
- Identify opportunities to automate manual work.
- Identify long running queries and tune the same.
- Partner with the platform engineering teams to ensure technical capabilities and data requirements are well understood, identify dependencies and integration points.
- Work with upstream to onboard new datasets, perform validations to confirm if client's ask is complaint with DV's guidelines.
- Able to derive insights from Grafana telemetries, monitor cluster's health, automate SOW/SOD/EOD checks.
- Participate in vendor calls, sprint ceremonies, reviews, or inspections, in collaboration with the project team, to ensure the quality of the deliverables.
- Participate in the user acceptance testing, implementation planning, post implementation validations and L3 production incident analysis as needed.
Minimum Requirements
- Strong knowledge of Data Fabric, Data Virtualization concepts & containers
- 8+ years of experience in Dremio, Data virtualization product, running Dremio in K8 containers
- Ability to work independently, apply critical and strategic thinking to issues, and manage deadlines while multi-tasking amongst different priorities.
- 8+ years of experience working as a Developer / L3 in Data engineering field, handling large data volume.
- Proven experience in Lake, databases (big data, nosql, rdbms), SQL, query tuning, able to read & analyze query execution plans, debug application/server logs, assist team in analyzing the issues.
- Understanding of different infrastructures (physical servers, containers, virtual machines, cloud)
- Strong tech expertise – Dremio, Python, SQL, RestAPI, Hive/HDFS, Kubernetes, Helm chart, Linux and data formats e.g. Iceberg, Delta format, Parquet, Avro, JSON
- Knowledge of containers, kubernetese, pods, nodes, PV/PVC, Load balancer, helm deployment
- Knowledge of application security, vault, SSL, TSL, secrets, connectivity channels - JDBC, ODBC, Arrowflight
- Knowledge of different Storage classes – S3, NAS, Trident, SSDs
- Understanding of Azure cloud, ADLS, Shipment pipeline would be plus