Logo
Infoblox

Staff Data Engineer

Infoblox, Santa Clara, California, us, 95053


DescriptionHybrid role: Requires to be in the office in Tacoma (WA) or Santa Clara (CA), at least 2 days a week*It’s an exciting time to be at Infoblox. Named a Top 25 Cyber Security Company by The Software Report and one of Inc. magazine’s Best Workplaces for 2020, Infoblox is the leader in cloud-first networking and security services. Our solutions empower organizations to take full advantage of the cloud to deliver network experiences that are inherently simple, scalable, and reliable for everyone. Infoblox customers are among the largest enterprises in the world and include 70% of the Fortune 500, and our success depends on bright, energetic, talented people who share a passion for building the next generation of networking technologies—and having fun along the way.We are looking for a Staff Data Engineer to join our Cloud Engineering team in Tacoma, WA or Santa Clara, CA, reporting to the senior manager of Software Engineering. In this role, you will develop platforms and products for Infoblox’s SaaS product line delivering next level networking for our customers. This is an opportunity to work closely with data scientists and product teams to curate and refine data powering our latest cloud products. Come join our growing Cloud Engineering team and help us build world class solutions.You are the ideal candidate if you are passionate about the nexus between data and computer science and driven to figure out how best to represent and summarize data in a way that informs good decisions and drives new products.What you’ll do:Curate large-scale data from a multitude of sources into appropriate sets for research and development for the data scientists, threat analysts, and developers across the companyDesign, test, and implement storage solutions for various consumers of the data. Especially data warehouses like ClickHouse and OpenSearch.Design and implement mechanisms to monitor data sources over time for changes using summarization, monitoring, and statistical methodsDesign, develop, and maintain APIs that enable seamless data integration and retrieval processes for internal and external applications, and ensure these APIs are scalable, secure, and efficient to support high-volume data interactionsLeverage computer science algorithms and constructs, including probabilistic data structures, to distill large data into sources of insight and enable future analyticsConvert prototypes into production data engineering solutions through disciplined software engineering practices, Spark optimizations, and modern deployment pipelinesCollaborate on design, implementation, and deployment of applications with the rest of software engineeringSupport data scientists and Product teams in building, debugging, and deploying Spark applications that best leverage dataBuild and maintain tools for automation, deployment, monitoring, and operationsCreate test plans, test cases, and run tests with automated toolsWhat you’ll bring:12+ years of experience with Python3, and 2+ years of experience with Spark; Scala experience is helpful5+ years of experience in data engineering, data science, and related data-centric fields using large-scale data environments3+ years of experience in using SQL and working with modern relational databases, including MySQL or PostgreSQL3+ years of experience with developing ETL pipelines and data manipulation scriptsExperience working with ClickHouse data warehouseExpertise in designing and developing APIs, preferably RESTful, for data access and manipulation in a distributed environmentProficient in Object Oriented Design and S.O.L.I.D principlesEmphasis on unit testing and code qualityProficient with AWS products, including EMR, S3, VPC, EC2, Athena, etc.Bachelor’s or master’s degree in computer science or a related field, or equivalent work experience requiredWhat success looks like:After six months, you will…Complete onboarding by demonstrating knowledge of the Data Lake and associated technologies and CI/CD processes by deploying ETL pipelines for curation and warehousing to productionComplete rotations on support/duty where you will gain experience with the different systems, tools, and processes in the Data Lake by resolving reported issuesContribute to the team's velocity by participating in Scrum and driving stories to completionAfter About a Year, You Will…Be an expert on the Data Lake target state architecture and drive engineering design and grooming sessions for new feature developmentApply coding best practices and provide in-depth reviews on the team’s pull requestsBe a thought leader in one or more domains of the Data Lake, driving development and mentoring teammates in this domain

#J-18808-Ljbffr