Input Technology Solutions
Data Engineer
Input Technology Solutions, Honolulu, Hawaii, United States, 96814
Input Technology Solutions is seeking a
Data Engineer
to support a program in Honolulu, HI. As a Data Engineer, you will play a crucial role in designing, developing, and maintaining our Advana data infrastructure, pipelines and systems. Proficiency in ETL, Databricks, Python, Spark, Scala, JavaScript/JSON, SQL, and Jupyter Notebooks is crucial for ensuring efficient data ingest, processing and analysis. You will develop and manage ETL processes and data pipelines across various government platforms, working with a wide range of data types.This is an ideal opportunity to be part of one of the fastest growing AI/ML companies in the industry. At BigBear.ai, we're in this business together. We own it, we make it thrive, and we enjoy the challenges of our work. We know that our employees play the largest role in our continual success. That is why we foster an environment of growth and development, with an emphasis on opportunity, recognition, and work-life balance. We give the same high level of commitment to our employees that we give to our clients. If BigBear.ai sounds like the place where you want to be, we'd enjoy speaking with you.What you will do:Design, develop, and implement end-to-end data pipelines, utilizing ETL processes and technologies such as Databricks, Python, Spark, Scala, JavaScript/JSON, SQL, and Jupyter Notebooks.Create and optimize data pipelines from scratch, ensuring scalability, reliability, and high-performance processing.Perform data cleansing, data integration, and data quality assurance activities to maintain the accuracy and integrity of large datasets.Leverage big data technologies to efficiently process and analyze large datasets, particularly those encountered in a federal agency.Troubleshoot data-related problems and provide innovative solutions to address complex data challenges.Implement and enforce data governance policies and procedures, ensuring compliance with regulatory requirements and industry best practices.Work closely with cross-functional teams to understand data requirements and design optimal data models and architectures.Collaborate with data scientists, analysts, and stakeholders to provide timely and accurate data insights and support decision-making processes.Maintain documentation for software applications, workflows, and processes.Stay updated with emerging trends and advancements in data engineering and recommend suitable tools and technologies for continuous improvement.What you need to have:
Bachelor's Degree and 0 to 2 years of experience; 6 to 8 years with no degreeClearance:TS/SCI clearance1+ years of Python experience including developing, running, packaging, and testing Python scriptsExperience with distributed version control systems (VCS)Experience with the entire ETL/ELT pipeline, including data ingestion, data normalization, data preparation, and database designExperience with conducting exploratory data analysis to communicate qualitative and quantitative findings to analystsExperience processing and fusing structured and unstructured dataExperience with data engineering projects supporting data science and AI/ML workloadsExperience creating solutions within a collaborative, cross-functional team environment in team sprint cyclesBachelor’s degreeWhat we'd like you to have:Experience with using Palantir products for data manipulation, correlation, and visualizationExperience with AWS or other cloud computing servicesExperience with Kafka and NiFi developmentExperience with containerization tools, including Docker and KubernetesTS/SCI with Counterintelligence Polygraph
Data Engineer
to support a program in Honolulu, HI. As a Data Engineer, you will play a crucial role in designing, developing, and maintaining our Advana data infrastructure, pipelines and systems. Proficiency in ETL, Databricks, Python, Spark, Scala, JavaScript/JSON, SQL, and Jupyter Notebooks is crucial for ensuring efficient data ingest, processing and analysis. You will develop and manage ETL processes and data pipelines across various government platforms, working with a wide range of data types.This is an ideal opportunity to be part of one of the fastest growing AI/ML companies in the industry. At BigBear.ai, we're in this business together. We own it, we make it thrive, and we enjoy the challenges of our work. We know that our employees play the largest role in our continual success. That is why we foster an environment of growth and development, with an emphasis on opportunity, recognition, and work-life balance. We give the same high level of commitment to our employees that we give to our clients. If BigBear.ai sounds like the place where you want to be, we'd enjoy speaking with you.What you will do:Design, develop, and implement end-to-end data pipelines, utilizing ETL processes and technologies such as Databricks, Python, Spark, Scala, JavaScript/JSON, SQL, and Jupyter Notebooks.Create and optimize data pipelines from scratch, ensuring scalability, reliability, and high-performance processing.Perform data cleansing, data integration, and data quality assurance activities to maintain the accuracy and integrity of large datasets.Leverage big data technologies to efficiently process and analyze large datasets, particularly those encountered in a federal agency.Troubleshoot data-related problems and provide innovative solutions to address complex data challenges.Implement and enforce data governance policies and procedures, ensuring compliance with regulatory requirements and industry best practices.Work closely with cross-functional teams to understand data requirements and design optimal data models and architectures.Collaborate with data scientists, analysts, and stakeholders to provide timely and accurate data insights and support decision-making processes.Maintain documentation for software applications, workflows, and processes.Stay updated with emerging trends and advancements in data engineering and recommend suitable tools and technologies for continuous improvement.What you need to have:
Bachelor's Degree and 0 to 2 years of experience; 6 to 8 years with no degreeClearance:TS/SCI clearance1+ years of Python experience including developing, running, packaging, and testing Python scriptsExperience with distributed version control systems (VCS)Experience with the entire ETL/ELT pipeline, including data ingestion, data normalization, data preparation, and database designExperience with conducting exploratory data analysis to communicate qualitative and quantitative findings to analystsExperience processing and fusing structured and unstructured dataExperience with data engineering projects supporting data science and AI/ML workloadsExperience creating solutions within a collaborative, cross-functional team environment in team sprint cyclesBachelor’s degreeWhat we'd like you to have:Experience with using Palantir products for data manipulation, correlation, and visualizationExperience with AWS or other cloud computing servicesExperience with Kafka and NiFi developmentExperience with containerization tools, including Docker and KubernetesTS/SCI with Counterintelligence Polygraph