Inizio Partners
Databricks Architect
Inizio Partners, Dublin, Ohio, United States, 43016
About the job Databricks Architect
OUR CLIENT
Our client provides data-driven, action-oriented solutions to business problems through statistical data mining, cutting-edge analytics techniques, and a consultative approach. Leveraging proprietary methodology and best-of-breed technology, our client's analytics team takes an industry-specific approach to transform decision-making and embed analytics more deeply into their business processes. They have a global footprint of 2,000+ data scientists and analysts who assist client organizations with complex risk minimization methods, advanced marketing, pricing and CRM strategies, internal cost analysis, and cost and resource optimization within the organization. They serve the insurance, healthcare, banking, capital markets, utilities, retail and e-commerce, travel, transportation and logistics industries.
Role & Responsibilities Overview:
Develop and optimize ETL pipelines from various data sources using Databricks on cloud (AWS, Azure, etc.)Experienced in implementing standardized pipelines with automated testing, Airflow scheduling, Azure DevOps for CI/CD, Terraform for infrastructure as code, and Splunk for monitoringContinuously improve systems through performance enhancements and cost reductions in computing and storageData Processing and API Integration: Utilize Spark Structured Streaming for real-time data processing and integrate data outputs with REST APIsLead Data Engineering Projects to manage and implement data-driven communication systemsExperienced with Scrum and Agile Methodologies to coordinate global delivery teams, run scrum ceremonies, manage backlog items, and handle escalationsIntegrate data across different systems and platformsStrong verbal and written communication skills to manage client discussionsCandidate Profile:
8+ years' experience in developing and implementing ETL pipelines from various data sources using Databricks on cloudSome experience in the insurance domain/ data is a mustProgramming Languages SQL, PythonTechnologies - IaaS (AWS or Azure or GCP), Databricks platform, Delta Lake storage, Spark (PySpark, Spark SQL).
Good to have - Airflow, Splunk, Kubernetes, Power BI, Git, Azure DevOps
Project Management using Agile, ScrumB.S. Degree in a data-centric field (Mathematics, Economics, Computer Science, Engineering or other science field), Information Systems, Information Processing or engineering.Excellent communication & leadership skills, with the ability to lead and motivate team membersAbility to work independently with some level of ambiguity and juggle multiple demands
This is a Hybrid role - 2 days per week in Dublin, OH. Client is currently considering candidates who are local to Dublin, OH or those who are open to relocating themselves to the Dublin, OH area.
OUR CLIENT
Our client provides data-driven, action-oriented solutions to business problems through statistical data mining, cutting-edge analytics techniques, and a consultative approach. Leveraging proprietary methodology and best-of-breed technology, our client's analytics team takes an industry-specific approach to transform decision-making and embed analytics more deeply into their business processes. They have a global footprint of 2,000+ data scientists and analysts who assist client organizations with complex risk minimization methods, advanced marketing, pricing and CRM strategies, internal cost analysis, and cost and resource optimization within the organization. They serve the insurance, healthcare, banking, capital markets, utilities, retail and e-commerce, travel, transportation and logistics industries.
Role & Responsibilities Overview:
Develop and optimize ETL pipelines from various data sources using Databricks on cloud (AWS, Azure, etc.)Experienced in implementing standardized pipelines with automated testing, Airflow scheduling, Azure DevOps for CI/CD, Terraform for infrastructure as code, and Splunk for monitoringContinuously improve systems through performance enhancements and cost reductions in computing and storageData Processing and API Integration: Utilize Spark Structured Streaming for real-time data processing and integrate data outputs with REST APIsLead Data Engineering Projects to manage and implement data-driven communication systemsExperienced with Scrum and Agile Methodologies to coordinate global delivery teams, run scrum ceremonies, manage backlog items, and handle escalationsIntegrate data across different systems and platformsStrong verbal and written communication skills to manage client discussionsCandidate Profile:
8+ years' experience in developing and implementing ETL pipelines from various data sources using Databricks on cloudSome experience in the insurance domain/ data is a mustProgramming Languages SQL, PythonTechnologies - IaaS (AWS or Azure or GCP), Databricks platform, Delta Lake storage, Spark (PySpark, Spark SQL).
Good to have - Airflow, Splunk, Kubernetes, Power BI, Git, Azure DevOps
Project Management using Agile, ScrumB.S. Degree in a data-centric field (Mathematics, Economics, Computer Science, Engineering or other science field), Information Systems, Information Processing or engineering.Excellent communication & leadership skills, with the ability to lead and motivate team membersAbility to work independently with some level of ambiguity and juggle multiple demands
This is a Hybrid role - 2 days per week in Dublin, OH. Client is currently considering candidates who are local to Dublin, OH or those who are open to relocating themselves to the Dublin, OH area.