Selby Jennings
Senior Software Engineer, Data Platform
Selby Jennings, Chicago, Illinois, United States, 60290
CAN'T SIT REMOTELY OUT OF CALIFORNIARequirements for This Role:Experience:
7+ years working with modern data technologies and/or building data-intensive distributed systems.Programming Skills:
Expert-level proficiency in Java/Scala or Python, with a proven ability to write high-quality, maintainable code.Database and Scripting:
Strong knowledge of SQL and Bash.Cloud Technologies:
Experience in leveraging and building cloud-native technologies for scalable data processing.Data Systems:
Previous experience with both batch and streaming systems, understanding their limitations and challenges.Data Processing Technologies:
Familiarity with a range of technologies such as Flink, Spark, Polars, Dask, etc.Data Storage Solutions:
Knowledge of various storage technologies, including S3, RDBMS, NoSQL, Delta/Iceberg, Cassandra, Clickhouse, Kafka, etc.Data Formats and Serialization:
Experience with multiple data formats and serialization systems like Arrow, Parquet, Protobuf/gRPC, Avro, Thrift, JSON, etc.ETL Pipelines:
Proven track record of managing complex data ETL pipelines using tools like Kubernetes, Argo Workflows, Airflow, Prefect, Dagster, etc.Schema Governance:
Prior experience dealing with schema governance and schema evolution.Data Quality Control:
Experience in developing data quality control processes to detect and address data gaps or inaccuracies.Mentorship:
A desire to mentor less experienced team members and promote best practices and high standards in code quality.Problem-Solving:
Strong technical problem-solving abilities.Agile Environment:
Proven capability to work in an agile, fast-paced environment, prioritize multiple tasks and projects, and handle the demands of a trading environment efficiently.
#J-18808-Ljbffr
7+ years working with modern data technologies and/or building data-intensive distributed systems.Programming Skills:
Expert-level proficiency in Java/Scala or Python, with a proven ability to write high-quality, maintainable code.Database and Scripting:
Strong knowledge of SQL and Bash.Cloud Technologies:
Experience in leveraging and building cloud-native technologies for scalable data processing.Data Systems:
Previous experience with both batch and streaming systems, understanding their limitations and challenges.Data Processing Technologies:
Familiarity with a range of technologies such as Flink, Spark, Polars, Dask, etc.Data Storage Solutions:
Knowledge of various storage technologies, including S3, RDBMS, NoSQL, Delta/Iceberg, Cassandra, Clickhouse, Kafka, etc.Data Formats and Serialization:
Experience with multiple data formats and serialization systems like Arrow, Parquet, Protobuf/gRPC, Avro, Thrift, JSON, etc.ETL Pipelines:
Proven track record of managing complex data ETL pipelines using tools like Kubernetes, Argo Workflows, Airflow, Prefect, Dagster, etc.Schema Governance:
Prior experience dealing with schema governance and schema evolution.Data Quality Control:
Experience in developing data quality control processes to detect and address data gaps or inaccuracies.Mentorship:
A desire to mentor less experienced team members and promote best practices and high standards in code quality.Problem-Solving:
Strong technical problem-solving abilities.Agile Environment:
Proven capability to work in an agile, fast-paced environment, prioritize multiple tasks and projects, and handle the demands of a trading environment efficiently.
#J-18808-Ljbffr