TechnoGen
"Data Engineer"
TechnoGen, Alpharetta, Georgia, United States, 30239
Data Engineer
Location:
Alpharetta GA
Title : Data Engineer, Modelling, design, integration, Python, Spark, DB2, NoSQL, Jenkins - 12 mths Alpharetta GA- C2H
Description:
Hybrid 3 days a week onsite
Potential to convert
The
Data System Engineer
will be responsible for tasks such as data engineering, data modeling, ETL processes, data warehousing, and data analytics & science. Our platform run both on premise and on the cloud (AWS/Azure).
Knowledge/Skills:
Able to establish, modify or maintain data structures and associated components according to design Understands and documents business data requirements Able to come up with
Conceptual and Logical Data Models
at Enterprise, Business Unit/Domain Level Understands
XML/JSON
and
schema development/reuse, database concepts, database designs,
Open Source and
NoSQL concepts Partners with
Sr. Data Engineers
and
Sr. Data architects
to
create platform level data models and database designs Takes part in reviews of own work and reviews of colleagues' work Has working knowledge of the core tools used in the
planning, analyzing, designing, building, testing, configuring and maintaining
of assigned application(s) Able to participate in assigned team's software delivery methodology (Agile, Scrum, Test-Driven Development, Waterfall, etc.) in support of data engineering pipeline development Understands infrastructure technologies and components like
servers, databases, and networking concepts Write code
to develop, maintain and optimized batch and event driven for storing, managing, and analyzing
large volumes of structured and unstructured data both Metadata integration
in data pipelines Automate
build and deployment processes
using
Jenkins
across all environments to enable faster, high-quality releases Qualification:
Up to 4 years
of software development experience in a professional environment and/or comparable experience such as:
Understanding of Agile or other rapid application development methods Exposure to
design and development
across one or more
database management
systems
DB2, SybaseIQ, Snowflake
as appropriate Exposure to methods relating to
application and database design, development, and automated testing Understanding of
big data technology and NOSQL
design and development with variety of data stores (document, column family, graph, etc.) General knowledge of
distributed (multi-tiered) systems, algorithms, and relational & non-relational databases Experience with
Linux and Python scripting
as well as large scale data processing technology such as
Spark Exposure to
Big data technology
and
NOSQL design
and
coding
with variety of data stores (document, column family, graph, etc.) Experience with
cloud technologies
such as
AWS and Azure , including deployment, management, and optimization of data analytics & science pipelines Nice to have:
Collibra, Terraform, Java, Golang, Ruby, Machine Learning Operation deployment Bachelor's degree in computer science, computer science engineering, or related field required
Location:
Alpharetta GA
Title : Data Engineer, Modelling, design, integration, Python, Spark, DB2, NoSQL, Jenkins - 12 mths Alpharetta GA- C2H
Description:
Hybrid 3 days a week onsite
Potential to convert
The
Data System Engineer
will be responsible for tasks such as data engineering, data modeling, ETL processes, data warehousing, and data analytics & science. Our platform run both on premise and on the cloud (AWS/Azure).
Knowledge/Skills:
Able to establish, modify or maintain data structures and associated components according to design Understands and documents business data requirements Able to come up with
Conceptual and Logical Data Models
at Enterprise, Business Unit/Domain Level Understands
XML/JSON
and
schema development/reuse, database concepts, database designs,
Open Source and
NoSQL concepts Partners with
Sr. Data Engineers
and
Sr. Data architects
to
create platform level data models and database designs Takes part in reviews of own work and reviews of colleagues' work Has working knowledge of the core tools used in the
planning, analyzing, designing, building, testing, configuring and maintaining
of assigned application(s) Able to participate in assigned team's software delivery methodology (Agile, Scrum, Test-Driven Development, Waterfall, etc.) in support of data engineering pipeline development Understands infrastructure technologies and components like
servers, databases, and networking concepts Write code
to develop, maintain and optimized batch and event driven for storing, managing, and analyzing
large volumes of structured and unstructured data both Metadata integration
in data pipelines Automate
build and deployment processes
using
Jenkins
across all environments to enable faster, high-quality releases Qualification:
Up to 4 years
of software development experience in a professional environment and/or comparable experience such as:
Understanding of Agile or other rapid application development methods Exposure to
design and development
across one or more
database management
systems
DB2, SybaseIQ, Snowflake
as appropriate Exposure to methods relating to
application and database design, development, and automated testing Understanding of
big data technology and NOSQL
design and development with variety of data stores (document, column family, graph, etc.) General knowledge of
distributed (multi-tiered) systems, algorithms, and relational & non-relational databases Experience with
Linux and Python scripting
as well as large scale data processing technology such as
Spark Exposure to
Big data technology
and
NOSQL design
and
coding
with variety of data stores (document, column family, graph, etc.) Experience with
cloud technologies
such as
AWS and Azure , including deployment, management, and optimization of data analytics & science pipelines Nice to have:
Collibra, Terraform, Java, Golang, Ruby, Machine Learning Operation deployment Bachelor's degree in computer science, computer science engineering, or related field required