Logo
United Software Group

Data Engineer - w2 contract Texas local only apply

United Software Group, Dallas, Texas, United States, 75215


Data Engineer - w2 contract Texas resident apply

Kindly Note:1st interview to be conducted on Zoom.Client interviews will be conducted in-person.

Data Engineer

ETL

Must be Proficient

At least 7 years

PL/SQL

Must be Proficient

Data Lake Exp

Preferred

Snowflake

Preferred

Goldman Sachs Experience with Alloy Services

Plus

Responsibilities:•Develop, maintain, and optimize data pipelines to extract, transform, and load large datasets from diverse sources into our data ecosystem.•Design and implement efficient and scalable data models that align with business requirements, ensuring data integrity and performance.•Collaborate with cross-functional teams to understand data needs and deliver solutions that meet those requirements.•Work closely with data scientists, analysts, and software engineers to ensure seamless integration of data solutions into larger systems.•Identify and resolve data quality issues, ensuring accuracy, reliability, and consistency of the data infrastructure.•Continuously monitor and improve data pipelines and processes, identifying opportunities for automation and optimization.•Stay updated with emerging trends, technologies, and best practices in data engineering, data modeling, and backend Java engineering.•Provide technical guidance and mentorship to junior team members, fostering their growth and development.

Requirements:•Bachelor's or Master's degree in Computer Science, Engineering, or a related field.•5+years of hands-on experience as a Data Engineer, working on complex data projects and implementing data modeling solutions.

Data Engineering

Must have:•Solid understanding of SQL and expertise in working with relational databases (e.g., PostgreSQL, MySQL).•In-depth knowledge of data modeling techniques and experience with data modeling tools•Proficiency in designing and optimizing data pipelines using ETL/ELT frameworks and tools (e.g., Informatica, Apache Spark, Airflow, AWS Glue).•Working knowledge on Data warehousing•Familiarity with cloud-based data platforms and services (e.g., Snowflake, AWS, Google Cloud, Azure).•Experience with version control systems (e.g., Git) and agile software development methodologies.•Strong communication skills to effectively convey technical concepts to both technical and non-technical stakeholders.•Excellent problem-solving skills and the ability to work independently and collaboratively in a fast-paced environment.

Good to Have:•JAVA 8, REST APIs, and microservices, Spring Boot framework•UNIX scripting

Primary Skillset: Data Engineering

Solid understanding of SQL and expertise in working with relational databases (e.g., DB2, MySQL).

Data Modelling knowledge

cloud-based data platforms like Snowflake

Working knowledge on Data warehousing

Cloud-based data platforms and services (e.g., AWS, Google Cloud, Azure)

Alteryx (good to have)

ETL/ELT tools like Informatica, Apache Spark

UNIX scripting

Good to have:

GS specific tools like Alloy Registry, LEGEND tool for Alloy data modelling, PURE, Data Browser, Data Lake etc