The Data Engineer helps build and maintain enriched data infrastructure to improve and strengthen Northwest Permanente’s capabilities in using data to drive analytical thinking, storytelling, and decision-making. The Engineer collects, cleanses, and transforms raw data from disparate sources and platforms into stored, integrated, and enriched data that can be used and interpreted by data scientists and business intelligence analysts. The Data Engineer responsibilities and required experience include database architecture, engineering, design, optimization, security, and administration, as well as data modeling, Extract, Transform, and Load (ETL) development, storage engineering, data warehousing, and data provisioning.
Major Responsibilities:
- Develops data warehouse solutions through ingesting, integrating, and curating data, building database solutions, and administering systems to deliver enriched data to business intelligence analysts and data scientists. Utilizes Spark SQL and Python in Databricks to build and optimize scalable data pipelines.
- Assists in efforts to facilitate, analyze, design, and execute architecture solutions and ensure solutions are leveraged; assists in creating documents and communicates the integration approach for components of the solution.
- Applies knowledge of ETL design, development, and processes, including building and orchestrating Databricks workflows and using Spark SQL and Python for data transformation.
- Takes accountability for ensuring specific interfaces, methods, parameters, procedures, and functions support technical solutions and are aligned with architectural designs.
- Applies knowledge of relational database skills including advanced SQL knowledge and the ability to create complex queries and stored procedures.
- Assists in efforts to identify and design solutions that allow performing root cause analyses to enable proactive issue resolution and data quality maintenance; assists in efforts to build processes and diagnostics tools/measures into the development process so that data pipelines can be monitored, and issues detected proactively.
- Gathers scope and business requirements for data infrastructure projects.
- Assists in efforts to translate business requirements and functional specifications into physical program designs, code modules, stable and reliable data solutions by partnering with analysts and other team members to understand business needs and functional specifications.
- Assists in efforts to create data dictionaries to document data lineages, data definitions, transformations, and metadata for data infrastructure projects; identifies and reconciles inconsistencies in data definitions; takes steps to assure metadata accuracy and validity.
- Documents programming changes and design, system modifications and their associated maintenance.
- Assists in executing strategies and plans for data security, backup, recovery, business continuity, and archiving.
- Identifies and develop opportunities for data reuse, migration, or retirement.
- Reviews and verifies resource estimates for technical design, coding, and testing efforts.
- Practices self-leadership and promotes learning in others by soliciting and acting on performance feedback; builds collaborative, cross-functional relationships.
Minimum Education and Work Experience:
- Bachelor’s degree in Computer Information Systems, Computer Science, or related field.
- Three (3) years of hands-on Databricks experience (Notebooks, Jobs, Workflows) in ETL design, implementation, and maintenance using technologies using Python, Spark SQL, Delta Lake, GitHub, SFTP, and TCP/IP.
- Four (4) years of experience in IT with a broad range of exposure to all aspects of business planning, systems analysis, application development, and/or data warehouse development using SDLC methodologies.
- Four (4) years data analytical experience.
- Two (2) years of hands-on experience in architecture, data modeling, and implementation of the enterprise data solutions.
- Strong passion for data analytics and fostering a data driven culture.
- Demonstrates curiosity and a growth mindset; fosters a culture that supports learning, innovation, and intelligent risk-taking; and exhibits resilience in the face of challenge.
- Familiarity with legal, security and regulatory issues associated with managing and exposing sensitive institutional data.
- Experience working with relational databases, data extraction and manipulation language, and in architecture principles and techniques across master data, transaction data and derived/analytic data.
- Experience in the use of data warehouse ETL methodologies and tools.
About Northwest Permanente:
We are the Permanente in Kaiser Permanente. Northwest Permanente is a self-governed, multi-specialty group of 1,500 physicians, clinicians, and administrative professionals caring for 630,000 members in Oregon and Southwest Washington. Together with Kaiser Foundation Health Plans and Kaiser Foundation Hospitals, we form Kaiser Permanente of the Northwest, an integrated health care program. Kaiser Permanente is one of the nation's preeminent health care systems, a benchmark for comprehensive, integrated, value-based, and high-quality care.
Our Northwest Permanente administrative professionals enjoy a wide range of company sponsored benefits:
- 15% employer contribution to retirement programs, including pension
- 90% employer-paid health plan
- Tuition Reimbursement
- Child Care Benefits
- Flexible Work Schedules
- Paid Parental Leave
- Self-Care Days + Paid Time Off
Equal Opportunity Employer
Northwest Permanente is an equal opportunity employer committed to fair, respectful, and inclusive workplaces. Applicants will be considered for employment without regard to race, religion, sex, age, national origin, disability, veteran status, or any other protected characteristic or status.