As a Data Engineer, you will contribute to the development and maintenance of our data infrastructure using technologies like Snowflake, Amazon S3, and Airflow. You will work within a team of data engineers to assist in designing and implementing data solutions that facilitate scalable data ingestion, storage, and processing. Your efforts will help maintain the quality, efficiency, and reliability of our data operations.
Key Responsibilities:
1. Development (70%)
Assist in the development and maintenance of scalable and efficient data pipelines using Python, SQL, and Snowflake, Snowpark. Support complex ETL processes and frameworks to manage data transformation, data structures, metadata, dependency, and workload management.
Contribute to enhancing our data ingestion processes using AWS Lambda functions and Snowpipe for real-time data processing and analytics.
Collaborate with data scientists and analysts to help build datasets that support business needs.
Create proof of concepts to support new innovation for the CareJourney product and understand new features of the underlying (Snowflake, AWS, Airflow, etc).
Create views and other layers of abstraction for data consumption both internal and external to CareJourney. This includes logical views, materialized views and secure views to support reports and data sharing.
2. Code Review (15%)
Participate in code review sessions to ensure code quality, adherence to best practices, and maintainability.
Learn from technical guidance and mentorship provided by senior team members and strive for continuous improvement.
3. Agile Participation (10%)
Support the data engineering team in agile practices such as sprint planning, daily stand-ups, retrospectives, and sprint reviews.
Help maintain clear communication between project stakeholders and the team, ensuring alignment on goals and deliverables.
4. Release Support (5%)
Assist in coordinating with project managers and IT stakeholders to plan release schedules.
Support the staging and deployment of data pipeline releases to ensure smooth rollouts with minimal downtime.
Required Skills & Qualifications:
1. At least 5 years of experience in a Data Engineer role or similar.
2. Proficient with Snowflake (or other cloud databases), Apache Airflow, and event driven orchestration patterns.
3. Strong programming skills in Python and Advanced SQL used within transformation data pipelines.
4. Understanding of Stored Procedures, Functions and when they are appropriate to use.
5. Experience with AWS services, especially Lambda functions and data streaming technologies.
6. Understanding of ELT tools and data orchestration processes.
7. Familiarity with data modeling, data access, and data storage techniques.
8. Knowledge of agile methodologies.
9. Excellent problem-solving abilities and willingness to learn.
10. Strong communication and organizational skills.
Desired Skills & Qualifications:
1. Working knowledge of American health care data.
2. Experience with data privacy rules and regulations such as HIPAA and GDPR.
3. Understanding of Snowflakes Snowpark development framework.
4. Familiarity with Snowflakes data sharing capabilities.
Education:
Bachelors degree in Computer Science, Engineering, or related field.
¡Postulate ahora y súmate a la querida Darwoft!
Envía tu CV a :
Consultas? Sigue al Recruiter
#J-18808-Ljbffr