Cynet Systems
Azure Data Architect with Pyspark - Remote / Telecommute
Cynet Systems, New York, New York, United States,
Job Description:
Pay Range: $60hr - $65hr
Responsibilities:
Design and develop scalable, high-performance data architecture solutions using PySpark, ADF, and Power BI to support business intelligence, analytics, and reporting needs.Build and manage robust data pipelines using PySpark and Azure Data Factory, ensuring efficient data extraction, transformation, and loading (ETL) processes across various data sources.Develop and maintain data models that optimize query performance and support the needs of analytics and reporting teams.Design and implement integration strategies to automate data flows between systems and ensure data consistency and accuracy.Collaboration Work closely with data engineers, data analysts, business intelligence teams, and other stakeholders to understand data requirements and deliver effective solutions.Data Governance and Security Ensure data solutions adhere to best practices in data governance, security, and compliance, including data privacy regulations and policies.Continuously monitor and optimize data processes and architectures for performance, scalability, and cost-efficiency.Utilize Power BI to design and develop interactive dashboards and reports that provide actionable insights for business stakeholders.Documentation:
Create comprehensive documentation for data architecture, data flows, ETL processes, and reporting solutions.Troubleshooting & Support: Provide technical support and troubleshooting for data-related issues, ensuring timely resolution and minimal impact on business operations.Qualifications:
Education: Bachelor's degree in Computer Science, Information Technology, Data Science, or a related field.Experience:
Minimum of 15+ years of experience in data architecture and engineering, with a focus on PySpark, ADF, and Power BI.Proven experience in designing and implementing data pipelines, ETL processes, and data integration solutions.Strong experience in data modeling and data warehouse design.Technical Skills:
Proficiency in PySpark for big data processing and transformation.Extensive experience with Azure Data Factory (ADF) for data orchestration and ETL workflows.Strong expertise in Power BI for data visualization, dashboard creation, and reporting.Knowledge of Azure services (e.g., Azure Data Lake, Azure Synapse) and other relevant cloud-based data technologies.Strong SQL skills and experience with relational databases.Soft Skills:
Excellent problem-solving and analytical skills.Strong communication and collaboration skills, with the ability to work effectively with cross-functional teams.Ability to manage multiple priorities in a fast-paced environment.
Pay Range: $60hr - $65hr
Responsibilities:
Design and develop scalable, high-performance data architecture solutions using PySpark, ADF, and Power BI to support business intelligence, analytics, and reporting needs.Build and manage robust data pipelines using PySpark and Azure Data Factory, ensuring efficient data extraction, transformation, and loading (ETL) processes across various data sources.Develop and maintain data models that optimize query performance and support the needs of analytics and reporting teams.Design and implement integration strategies to automate data flows between systems and ensure data consistency and accuracy.Collaboration Work closely with data engineers, data analysts, business intelligence teams, and other stakeholders to understand data requirements and deliver effective solutions.Data Governance and Security Ensure data solutions adhere to best practices in data governance, security, and compliance, including data privacy regulations and policies.Continuously monitor and optimize data processes and architectures for performance, scalability, and cost-efficiency.Utilize Power BI to design and develop interactive dashboards and reports that provide actionable insights for business stakeholders.Documentation:
Create comprehensive documentation for data architecture, data flows, ETL processes, and reporting solutions.Troubleshooting & Support: Provide technical support and troubleshooting for data-related issues, ensuring timely resolution and minimal impact on business operations.Qualifications:
Education: Bachelor's degree in Computer Science, Information Technology, Data Science, or a related field.Experience:
Minimum of 15+ years of experience in data architecture and engineering, with a focus on PySpark, ADF, and Power BI.Proven experience in designing and implementing data pipelines, ETL processes, and data integration solutions.Strong experience in data modeling and data warehouse design.Technical Skills:
Proficiency in PySpark for big data processing and transformation.Extensive experience with Azure Data Factory (ADF) for data orchestration and ETL workflows.Strong expertise in Power BI for data visualization, dashboard creation, and reporting.Knowledge of Azure services (e.g., Azure Data Lake, Azure Synapse) and other relevant cloud-based data technologies.Strong SQL skills and experience with relational databases.Soft Skills:
Excellent problem-solving and analytical skills.Strong communication and collaboration skills, with the ability to work effectively with cross-functional teams.Ability to manage multiple priorities in a fast-paced environment.