Logo
PETADATA

ETL Developer ( Matillion Tool )

PETADATA, Seattle, WA, United States


POSITION

ETL Developer with Matillion Tool

LOCATION

Seattle, WA. - ONSITE

Experience: 10+ years

Work type: Fulltime ( W2 / C2C )

PETADATA is currently looking to hire for the ETL Developer role experience with Matillion tool for one of their clients.

Roles & Responsibilities:

  • The ideal candidate should have experience with both streaming and batch workflows will be essential in ensuring the efficient flow and processing of data to support our clients.
  • Collaborate with cross-functional teams to understand data requirements and design robust data architecture solutions.
  • Designing, develop, and implement scalable data processing solutions using Apache Spark
  • The ability to organize and to keep the projects well-arranged and structured.
  • Good knowledge on data analysis, design, and programming skills such as JavaScript, SQL and XML, and DOM.
  • Ensure data quality, integrity, and consistency throughout the ETL pipeline.
  • Integrate data from different systems and sources to provide a unified view for analytical purposes.
  • Collaborate with data analysts to implement solutions that meet their data integration needs.
  • Design and implement streaming workflows using PySpark Streaming or other relevant technologies.
  • Build strings of configured components inside a job and then run that job to accomplish a desired task such as loading or transforming data. using Matillion ETL tool
  • Develop batch processing workflows for large-scale data processing and analysis.
  • Has to analyze the business requirement to determine the volume of data extracted from different sources, data models, to ensure the quality of the data involved.
  • Should be able to figure out best storage medium required for the data warehouse needed.
  • Has to identify the data storage needs to determine the amount of data to deal with the company’s requirement.
  • Must ensure the data quality that everything is in place at the transformation stage to eliminate errors and fix unstructured and unorganized data extracted.
  • Must be responsible to ensure that the data is loaded into the warehouse system and meets the business needs and standards.
  • The Candidate must responsible for data flow validation, creating and building a secured database warehouse that meets a given company’s needs and standards.
  • Must be responsible for determining the storage needs of a business and the volume of data involved.

Required skills:

  • Implement ETL processes to extract, transform, and load data from various sources to ensure data quality, integrity, and consistency throughout the ETL pipeline.
  • The candidate should be expertise in Matillion ETL, Python, PySpark, ETL processes, CI/CD (Jenkins or GitHub).
  • Expertise in Python and PySpark to develop efficient data processing and analysis scripts.
  • Optimize code for performance and scalability, keeping up-to-date with the latest industry best practices.
  • Must load data and be proficient in valuable technical skills such as SQL, ? JAVA, XML, and DOM, among others.
  • Extensive knowledge and Hands on Experience on Spark and its technologies.
  • Hands on Experience with Apache Spark framework, including the Spark SQL module for querying databases.
  • Familiar with various coding languages used in web development, including HTML, CSS and JavaScript, Python, Java, Scala, or R proficiency.
  • Applicants must be able to write clean code that’s free of bugs and reproducible by other developers.
  • Candidates should have experience managing SQL databases and organizing big data.
  • Hands on experience with ETL such as Matillion ETL, MS SQL, SSIS (Server Integration Services), Python / Perl, Oracle, SQL Server/ MySQL.
  • Solid understanding of Data warehousing Schemas, Dimensional modeling, implementing data storage solutions to supports efficient data retrieval and analysis.
  • The skill set necessary debugging ETL processes, optimizing data flows, and ensuring that the data pipeline is robust and error-free.

Educational Qualification:

Bachelor's/ Master’s degree in Computer Science, Engineering, or a related field.

We offer a professional work environment and are given every opportunity to grow in the Information technology world.

Note:

Candidates required to attend Phone/Video Call / In person interviews and after Selection of candidate (He/She) should go through all background checks on Education and Experience.

Please email your resume to: careers@petadata.co

After carefully reviewing your experience and skills one of our HR team members will contact you on the next steps.