Willis Towers Watson
Data Engineer
Willis Towers Watson, Chicago, IL
Job Description
Job Summary:
As a Senior Data Engineer, you will play a pivotal role in designing, building, and optimizing complex data pipelines and ETL processes while leveraging your expertise in Azure Synapse and PySpark to build advanced data analytics systems and data outputs for downstream consumption. You will be responsible for architecting scalable and efficient data models to support the business processes for reporting and data integrations. You will utilize cutting-edge Generative AI (GenAI) technologies to drive innovative data extraction and analysis solutions for our downstream consumers. The ideal candidate will bring over 10 years of experience in data engineering with a strong background in ETL, Data Modeling through Azure cloud solutions, and the integration of AI into data workflows.
The Role:
Key Responsibilities:
ETL Development & Maintenance: Lead the design, development, and optimization of complex ETL processes and pipelines that enable reliable data ingestion, transformation, and loading across a variety of sources.
Azure Synapse Analytics: Architect and develop scalable data solutions utilizing Azure Synapse with a deep focus on performance optimization. Create optimized PySpark notebooks for advanced data transformations and analytical queries.
Data Modeling: Design and maintain logical and physical data models, ensuring data structures align with business needs, scalability, and optimization for data warehousing and analytics.
GenAI Integration for Data: Lead the application of Generative AI technologies for data analysis and data extraction, including leveraging GenAI for predictive analytics, automated data transformation, and natural language query processing.
Data Pipeline Automation: Develop, implement, and manage automated data pipelines for continuous integration and deployment of data solutions. Incorporate best practices for monitoring and error-handling in production environments.
Collaboration: Work closely with other Data Engineers, analysts, and business stakeholders to understand their data requirements and provide innovative, scalable data solutions.
Performance Tuning & Optimization: Continuously monitor, evaluate, and optimize data pipelines and queries to enhance the performance of data systems, minimize latency, and ensure real-time data availability.
Cloud Engineering: Drive cloud-native engineering best practices on the Azure platform including security, scalability, high availability, disaster recovery, and cost-efficiency in data storage and processing.
Documentation & Best Practices: Create and maintain clear, concise documentation for data pipelines, models, and processes. Promote best practices for data governance, quality, and security.
Qualifications
The Requirements:
Required Skills & Qualifications:
10+ years of experience in Data Engineering and ETL processes with proven expertise in building and optimizing data pipelines.
5+ years of experience in Azure Synapse Analytics, including hands-on work with PySpark notebooks.
5+ years of experience in Data Modeling, with strong expertise in relational and dimensional modeling that have been proven as successful deliverables of desired end usage as it relates to consumer report outcomes. Designing and maintaining Data Model documentation for use across the Data Ecosystem
3+ years of experience in integrating and leveraging Generative AI technologies for data-related functions such as data analysis, extraction, and transformation.
Expertise in Azure Cloud Ecosystem, including Azure Data Lake, Azure Data Factory, and Synapse Studio.
Strong proficiency in PySpark and distributed computing frameworks as it relates to the ingestion, transformation and distribution of data into established data frameworks or processes.
Solid understanding of Data Governance principles, including data security, data quality, and master data management (MDM), and understanding or usage of Purview for cataloging the lineage of the data assets for consumption
Strong SQL skills and proficiency in other languages like Python, Scala, or R.
Experience with DevOps practices such as CI/CD pipelines for data workflows, including tools like Azure Git
Familiarity with AI/ML workflows and the intersection of AI with data engineering processes.
Preferred Qualifications:
Azure Certifications: Microsoft Certified: Azure Data Engineer Associate or Azure Solutions Architect Expert.
Experience with real-time data streaming solutions such as Azure Event Hubs.
Hands-on experience with BI Tools such as Power BI or similar.
Knowledge of NoSQL databases such as Cosmos DB or MongoDB.
Experience with containerization (Docker, Kubernetes) for data engineering workloads.
Experience in the Property and Casualty Insurance industry to allow accelerated learning of business structure and data needs.
Soft Skills:
Problem-Solving Orientation: Ability to think critically and solve complex problems with innovative and practical solutions.
Strong Communication Skills: Capable of communicating effectively with both technical and non-technical stakeholders.
Leadership & Mentorship: Experience in mentoring junior engineers and leading technical initiatives across teams.
Adaptability & Continuous Learning: Openness to continuously learning new tools and technologies in the evolving landscape of cloud data engineering and AI.
Location: Near major WTW offices in the United States
Expected Project deliverables within the first 18 months (in order of importance)
Build Data Models from existing Applications as well as integrations with the Data Lake technology to enable Business user to build analytics reports based on the inputs from the applications.
Utilizing GenAI technologies to Extract, Transform, and fill in data gaps based on existing data and analytics approaches.
Building out data extraction processes and utilizing existing data to enrich the data where gaps exist.
Building out data update processes where data from other systems needs to be integrated to form core data record.
This position will remain posted for a minimum of three business days from the date posted or until a sufficient/appropriate candidate slate has been identified.
Compensation and Benefits
Base salary range and benefits information for this position are being included in accordance with requirements of various state/local pay transparency legislation. Please note that salaries may vary for different individuals in the same role based on several factors, including but not limited to location of the role, individual competencies, education/professional certifications, qualifications/experience, performance in the role and potential for revenue generation (Producer roles only).
Compensation
The base salary compensation range being offered for this role is $110,000 - $118,000 USD per year. This role is also eligible for an annual short-term incentive bonus.
Company Benefits
WTW provides a competitive benefit package which includes the following (eligibility requirements apply):
At WTW, we trust you to know your work and the people, tools and environment you need to be successful. The majority of our colleagues work in a "hybrid" style, with a mix of remote, in-person and in-office interactions dependent on the needs of the team, role and clients. Our flexibility is rooted in trust and "hybrid" is not a one-size-fits-all solution.
We understand flexibility is key to supporting an inclusive and diverse workforce and so we encourage requests for all types of flexible working as well as location-based arrangements. Please speak to your recruiter to discuss more.
EOE, including disability/vets
Job Summary:
As a Senior Data Engineer, you will play a pivotal role in designing, building, and optimizing complex data pipelines and ETL processes while leveraging your expertise in Azure Synapse and PySpark to build advanced data analytics systems and data outputs for downstream consumption. You will be responsible for architecting scalable and efficient data models to support the business processes for reporting and data integrations. You will utilize cutting-edge Generative AI (GenAI) technologies to drive innovative data extraction and analysis solutions for our downstream consumers. The ideal candidate will bring over 10 years of experience in data engineering with a strong background in ETL, Data Modeling through Azure cloud solutions, and the integration of AI into data workflows.
The Role:
Key Responsibilities:
ETL Development & Maintenance: Lead the design, development, and optimization of complex ETL processes and pipelines that enable reliable data ingestion, transformation, and loading across a variety of sources.
Azure Synapse Analytics: Architect and develop scalable data solutions utilizing Azure Synapse with a deep focus on performance optimization. Create optimized PySpark notebooks for advanced data transformations and analytical queries.
Data Modeling: Design and maintain logical and physical data models, ensuring data structures align with business needs, scalability, and optimization for data warehousing and analytics.
GenAI Integration for Data: Lead the application of Generative AI technologies for data analysis and data extraction, including leveraging GenAI for predictive analytics, automated data transformation, and natural language query processing.
Data Pipeline Automation: Develop, implement, and manage automated data pipelines for continuous integration and deployment of data solutions. Incorporate best practices for monitoring and error-handling in production environments.
Collaboration: Work closely with other Data Engineers, analysts, and business stakeholders to understand their data requirements and provide innovative, scalable data solutions.
Performance Tuning & Optimization: Continuously monitor, evaluate, and optimize data pipelines and queries to enhance the performance of data systems, minimize latency, and ensure real-time data availability.
Cloud Engineering: Drive cloud-native engineering best practices on the Azure platform including security, scalability, high availability, disaster recovery, and cost-efficiency in data storage and processing.
Documentation & Best Practices: Create and maintain clear, concise documentation for data pipelines, models, and processes. Promote best practices for data governance, quality, and security.
Qualifications
The Requirements:
Required Skills & Qualifications:
10+ years of experience in Data Engineering and ETL processes with proven expertise in building and optimizing data pipelines.
5+ years of experience in Azure Synapse Analytics, including hands-on work with PySpark notebooks.
5+ years of experience in Data Modeling, with strong expertise in relational and dimensional modeling that have been proven as successful deliverables of desired end usage as it relates to consumer report outcomes. Designing and maintaining Data Model documentation for use across the Data Ecosystem
3+ years of experience in integrating and leveraging Generative AI technologies for data-related functions such as data analysis, extraction, and transformation.
Expertise in Azure Cloud Ecosystem, including Azure Data Lake, Azure Data Factory, and Synapse Studio.
Strong proficiency in PySpark and distributed computing frameworks as it relates to the ingestion, transformation and distribution of data into established data frameworks or processes.
Solid understanding of Data Governance principles, including data security, data quality, and master data management (MDM), and understanding or usage of Purview for cataloging the lineage of the data assets for consumption
Strong SQL skills and proficiency in other languages like Python, Scala, or R.
Experience with DevOps practices such as CI/CD pipelines for data workflows, including tools like Azure Git
Familiarity with AI/ML workflows and the intersection of AI with data engineering processes.
Preferred Qualifications:
Azure Certifications: Microsoft Certified: Azure Data Engineer Associate or Azure Solutions Architect Expert.
Experience with real-time data streaming solutions such as Azure Event Hubs.
Hands-on experience with BI Tools such as Power BI or similar.
Knowledge of NoSQL databases such as Cosmos DB or MongoDB.
Experience with containerization (Docker, Kubernetes) for data engineering workloads.
Experience in the Property and Casualty Insurance industry to allow accelerated learning of business structure and data needs.
Soft Skills:
Problem-Solving Orientation: Ability to think critically and solve complex problems with innovative and practical solutions.
Strong Communication Skills: Capable of communicating effectively with both technical and non-technical stakeholders.
Leadership & Mentorship: Experience in mentoring junior engineers and leading technical initiatives across teams.
Adaptability & Continuous Learning: Openness to continuously learning new tools and technologies in the evolving landscape of cloud data engineering and AI.
Location: Near major WTW offices in the United States
Expected Project deliverables within the first 18 months (in order of importance)
Build Data Models from existing Applications as well as integrations with the Data Lake technology to enable Business user to build analytics reports based on the inputs from the applications.
Utilizing GenAI technologies to Extract, Transform, and fill in data gaps based on existing data and analytics approaches.
Building out data extraction processes and utilizing existing data to enrich the data where gaps exist.
Building out data update processes where data from other systems needs to be integrated to form core data record.
This position will remain posted for a minimum of three business days from the date posted or until a sufficient/appropriate candidate slate has been identified.
Compensation and Benefits
Base salary range and benefits information for this position are being included in accordance with requirements of various state/local pay transparency legislation. Please note that salaries may vary for different individuals in the same role based on several factors, including but not limited to location of the role, individual competencies, education/professional certifications, qualifications/experience, performance in the role and potential for revenue generation (Producer roles only).
Compensation
The base salary compensation range being offered for this role is $110,000 - $118,000 USD per year. This role is also eligible for an annual short-term incentive bonus.
Company Benefits
WTW provides a competitive benefit package which includes the following (eligibility requirements apply):
- Health and Welfare Benefits: Medical (including prescription coverage), Dental, Vision, Health Savings Account, Commuter Account, Health Care and Dependent Care Flexible Spending Accounts, Group Accident, Group Critical Illness, Life Insurance, AD&D, Group Legal, Identify Theft Protection, Wellbeing Program and Work/Life Resources (including Employee Assistance Program)
- Leave Benefits: Paid Holidays, Annual Paid Time Off (includes paid state/local paid leave where required), Short-Term Disability, Long-Term Disability, Other Leaves (e.g., Bereavement, FMLA, ADA, Jury Duty, Military Leave, and Parental and Adoption Leave), Paid Time Off
- Retirement Benefits: Contributory Pension Plan and Savings Plan (401k). All Level 38 and more senior roles may also be eligible for non-qualified Deferred Compensation and Deferred Savings Plans.
At WTW, we trust you to know your work and the people, tools and environment you need to be successful. The majority of our colleagues work in a "hybrid" style, with a mix of remote, in-person and in-office interactions dependent on the needs of the team, role and clients. Our flexibility is rooted in trust and "hybrid" is not a one-size-fits-all solution.
We understand flexibility is key to supporting an inclusive and diverse workforce and so we encourage requests for all types of flexible working as well as location-based arrangements. Please speak to your recruiter to discuss more.
EOE, including disability/vets