Programmer Analyst 5022006

Website ManpowerWest Manpower San Diego

Jobs in San Diego and surroundings areas

Job Id: 1033
Top 5 Required Skills
1. AWS
2. Data Engineering
3. SQL
4. Python
5. Informatica

Technologies (specific tools and/or programs you would want to see on their resumes)
• AWS, Redshift, Apache Airflow, Glue, Python, Informatica, Snowflake, IICS, ERP, Salesforce

Keywords (specific words, projects, programs you would want to see on their resumes)
• AWS, Redshift, Apache Airflow, Glue, Python, Informatica, Snowflake, IICS, ERP, Salesforce
• Data discovery, modeling, engineering, pipeline, data warehouse

Education Requirement
• Bachelor of Science in Computer Science, Information Technology, Data Science, or a related field.

Required Years of Experience
• 3 years min

Physical Requirements
• Push Max Weight Limit = 0
• Pull Max Weight Limit = 0
• Lift Max Weight Limit = 0
• Forklift Required (Y/N): N

Driving Requirements
• Are there driving responsibilities no matter how minimal with this role? No
• (If Yes)How many hours per week? n/a

• Work Location Requirement: 100% Onsite
• Work Address: 5775 Morehouse Drive, San Diego, CA
• Qualcomm Home Building: O
• Work Days: Mon-Fri
• Exact Shift Time: 8:30-5:00pm PST
• Weekly / Daily Expected Hours: 40.0 / 8.0

As a Data Engineer at Qualcomm, you will play a pivotal role in shaping and implementing robust data architecture while constructing efficient data pipelines aligned with our organizational goals. This position supports various business functions, including Supply Chain, Finance, Sales & Marketing, and other corporate areas. The ideal candidate will possess a deep understanding of data architecture principles, data discovery/modeling, and the seamless integration of emerging technologies into a cohesive data ecosystem that fosters innovation, operational excellence, and strategic insights.

Principal Duties and Responsibilities:
• Collaborate with cross-functional and global teams to comprehend business requirements and translate them into effective solutions.
• Design and manage resilient, scalable data models and architectures that cater to the evolving needs of the business, emphasizing efficiency, quality, and security.
• Implement data pipelines using tools such as Apache Airflow, AWS Glue, Redshift, S3, Python, and Informatica Cloud.
• Facilitate communication within and outside the project team to resolve conflicts related to implementation schedules, design complexities, and other challenges.

Good to have qualifications
• Experience in Data Engineering leveraging AWS Redshift, S3, Glue, Airflow, Python, SQL and Kubernetes
• Familiarity with Informatica Cloud and Informatica Power Center is essential.
• Strong expertise in data modeling tools and methodologies, encompassing both structured and unstructured data environments.
• Demonstrated experience in data governance, data quality management, and data security practices.
• Exceptional analytical and problem-solving skills, enabling the translation of intricate technical challenges into actionable solutions.
• Ability to quickly learn and adapt to new technologies.
• Strong sense of ownership and growth mindset.
• Curiosity about the problem domain and an analytical approach.
• Strong influence skills to drive business adoption and change.
• Experience in data discovery of source systems like Oracle E-Business Suite and Salesforce is preferable.

Leave a Reply