Logo
Systems Technology Group Inc.

Data Engineer with Google Cloud Platform (GCP) Experience 9.12.24

Systems Technology Group Inc., Dearborn, Michigan, United States, 48120


Title: Data Engineer with Google Cloud Platform (GCP) Experience 9.12.24Description:

STG is a SEI CMMi Level 5 company with several Fortune 500 and State Government clients. STG has an opening for

Data Engineer with Google Cloud Platform (GCP) Experience.Please note that this project assignment is with our own direct clients. We do not go through any vendors. STG only does business with direct end clients. This is expected to be a long-term position. STG will provide immigration and permanent residency sponsorship assistance to those candidates who need it.Job Description:The right person for this position will have high energy, a positive attitude, a high attention to detail, and an unwavering commitment to teamwork.

Data Engineer with Google Cloud Platform (GCP) Experience in Dearborn, MI.Position Description:Materials Management Platform (MMP) is a multi-year transformation initiative aimed at transforming Materials Requirement Planning & Inventory Management capabilities. This is part of a larger Industrial Systems IT Transformation effort. This position responsibility is to design & deploy Data Centric Architecture in GCP for Materials Management platform which would get / give data from multiple applications modern & Legacy in Product Development, Manufacturing, Finance, Purchasing, N-Tier Supply Chain, Supplier Collaboration.Skills Required:Design and implement data-centric solutions on Google Cloud Platform (GCP) using various GCP tools like Big Query, Google Cloud Storage, Cloud SQL, Memory Store, Dataflow, Dataproc, Artifact Registry, Cloud Build, Cloud Run, Vertex AI, Pub/Sub, GCP APIs.Build ETL pipelines to ingest the data from heterogeneous sources into our systemDevelop data processing pipelines using programming languages like Java and Python to extract, transform, and load (ETL) dataCreate and maintain data models, ensuring efficient storage, retrieval, and analysis of large datasetsDeploy and manage databases, both SQL and NoSQL, such as Bigtable, Firestore, or Cloud SQL, based on project requirementsOptimize data workflows for performance, reliability, and cost-effectiveness on the GCP infrastructure. Implement version control and CI/CD practices for data engineering workflows to ensure reliable and efficient deployments.Utilize GCP monitoring and logging tools to proactively identify and address performance bottlenecks and system failuresTroubleshoot and resolve issues related to data processing, storage, and retrieval.Promptly address code quality issues using SonarQube, Checkmarx, Fossa, and Cycode throughout the development lifecycleImplement security measures and data governance policies to ensure the integrity and confidentiality of dataCollaborate with stakeholders to gather and define data requirements, ensuring alignment with business objectives.Develop and maintain documentation for data engineering processes, ensuring knowledge transfer and ease of system maintenance.Participate in on-call rotations to address critical issues and ensure the reliability of data engineering systems. Provide mentorship and guidance to junior team members, fostering a collaborative and knowledge-sharing environment.Experience Required:8 years of professional experience in: o Data engineering, data product development and software product launchesAt least three of the following languages: Java, Python, Spark, Scala, SQL and experience performance tuning.4 years of cloud data/software engineering experience building scalable, reliable, and cost-effective production batch and streaming data pipelines using: o Data warehouses like Google BigQuery. o Workflow orchestration tools like Airflow. o Relational Database Management System like MySQL, PostgreSQL, and SQL Server. o Real-Time data streaming platform like Apache Kafka, GCP Pub/Sub o Microservices architecture to deliver large-scale real-time data processing application. o REST APIs for compute, storage, operations, and security. o DevOps tools such as Tekton, GitHub Actions, Git, GitHub, Terraform, Docker. o Project management tools like Atlassian JIRAAutomotive experience is preferredSupport in an onshore/offshore model is preferredExcellent at problem solving and prevention.Knowledge and practical experience of agile deliveryExperience in IDOC processing, APIs and SAP data migration projects.Experience working in SAP S4 Hana environmentEducation Required:Bachelor's degree or equivalent qualification in computer science, engineering or related disciplines.Data Engineer with Google Cloud Platform (GCP) Experience is based in

Dearborn, MI.

A great opportunity to experience the corporate environment leading personal career growth.Resume Submittal Instructions:

Interested/qualified candidates should email their word formatted resumes to

Vasavi Konda - vasavi.konda(.@)

and/or contact @ (Two-Four-Eight) Seven- One-Two - Six-Seven-Two-Five (@248.712.6725) . In the subject line of the email please include:

First and Last Name

-

Data Engineer with Google Cloud Platform (GCP) Experience.For more information about STG, please visit us at .Sincerely,Vasavi Konda| Recruiting SpecialistOpportunities don't happen, you create them.Systems Technology Group (STG)3001 W. Big Beaver Road, Suite 500Troy, Michigan 48084Phone: @ (Two-Four-Eight) Seven- One-Two - Six-Seven-Two-Five: @ 248.712.6725(O)Email: vasavi.konda(.@)