ACL Digital
Programmer Analyst|1911 Programmer Analyst|1911
ACL Digital, San Diego, California, 92189
Job Description: Top 5 Required Skills 1. AWS 2. Data Engineering 3. SQL 4. Python 5. Informatica Technologies (specific tools and/or programs you would want to see on their resumes) • AWS, Redshift, Apache Airflow, Glue, Python, Informatica, Snowflake, IICS, ERP, Salesforce Keywords (specific words, projects, programs you would want to see on their resumes) • AWS, Redshift, Apache Airflow, Glue, Python, Informatica, Snowflake, IICS, ERP, Salesforce • Data discovery, modeling, engineering, pipeline, data warehouse Education Requirement • Bachelor of Science in Computer Science, Information Technology, Data Science, or a related field. Required Years of Experience • 3 years min Physical Requirements • Push Max Weight Limit 0 • Pull Max Weight Limit 0 • Lift Max Weight Limit 0 • Forklift Required (Y/N): N Driving Requirements • Are there driving responsibilities no matter how minimal with this role? No • (If Yes)How many hours per week? n/a • Work Location Requirement: 100% Onsite • Work Address: 5775 Morehouse Drive, San Diego, CA • Qualcomm Home Building: O • Work Days: Mon-Fri • Exact Shift Time: 8:30-5:00pm PST • Weekly / Daily Expected Hours: 40.0 / 8.0 As a Data Engineer at Qualcomm, you will play a pivotal role in shaping and implementing robust data architecture while constructing efficient data pipelines aligned with our organizational goals. This position supports various business functions, including Supply Chain, Finance, Sales & Marketing, and other corporate areas. The ideal candidate will possess a deep understanding of data architecture principles, data discovery/modeling, and the seamless integration of emerging technologies into a cohesive data ecosystem that fosters innovation, operational excellence, and strategic insights. Principal Duties and Responsibilities: • Collaborate with cross-functional and global teams to comprehend business requirements and translate them into effective solutions. • Design and manage resilient, scalable data models and architectures that cater to the evolving needs of the business, emphasizing efficiency, quality, and security. • Implement data pipelines using tools such as Apache Airflow, AWS Glue, Redshift, S3, Python, and Informatica Cloud. • Facilitate communication within and outside the project team to resolve conflicts related to implementation schedules, design complexities, and other challenges. Good to have qualifications • Experience in Data Engineering leveraging AWS Redshift, S3, Glue, Airflow, Python, SQL and Kubernetes • Familiarity with Informatica Cloud and Informatica Power Center is essential. • Strong expertise in data modeling tools and methodologies, encompassing both structured and unstructured data environments. • Demonstrated experience in data governance, data quality management, and data security practices. • Exceptional analytical and problem-solving skills, enabling the translation of intricate technical challenges into actionable solutions. • Ability to quickly learn and adapt to new technologies. • Strong sense of ownership and growth mindset. • Curiosity about the problem domain and an analytical approach. • Strong influence skills to drive business adoption and change. • Experience in data discovery of source systems like Oracle E-Business Suite and Salesforce is preferable. Comments for Suppliers: