Posted in

Sr. Data Engineer

Sr. Data Engineer

CompanyPatientPoint
LocationCincinnati, OH, USA
Salary$Not Provided – $Not Provided
TypeFull-Time
DegreesBachelor’s
Experience LevelSenior

Requirements

  • 5+ years of experience working on cloud data warehouses and data pipelines with a focus on data engineering, building scalable, sustainable and secure data platforms powering intelligent applications.
  • Bachelor’s degree in Informatics, Business Technology, Analytics, Computer Science or a related field.
  • 2+ years of hands-on experience in working with dbt
  • 3+ years of hands-on experience with Snowflake.
  • Advanced experience in a Data Engineering and ELT Engineering role.
  • Expert-level proficiency in SQL query and stored procedure development.
  • Competent with Python, Airflow, GitHub and DAG construction.
  • Experience with unstructured datasets and ability to handle Parquet, JSON, AVRO and XML file formats.
  • Strong understanding of CI/CD principles, DevOps practices, software testing and quality
  • Hands-on Experience with cloud-based orchestration tools such as Apache Airflow Astronomer.
  • Advanced experience implementing data quality initiatives, monitoring, and auditing.

Responsibilities

  • Responsible for architecting end-to-end data solutions that meet our business partners’ expectations and integrate into our Data Platform.
  • Engage with the Data team using scrum framework to comprehensively grasp requirements for each deliverable.
  • Accountable to follow Data team’s best practices in solution design, build, orchestration and automation, data ingest, monitoring performance optimizations, data quality, solution accuracy and compliance throughout the lifecycle.
  • Leverage a modern tool stack including Snowflake, Atlan, Fivetran, Docker, AWS, and Astronomer (Airflow) to cultivate an environment where analysts and data engineers can autonomously enact changes in an automated, thoroughly tested and high-quality manner.
  • Lead design, development, prototyping, operations and implementation of data solutions and pipelines.
  • Analyze the impact of changes to downstream systems/products and recommend alternatives to minimize the impact.
  • Build a deep knowledge in each PatientPoint data information domain.
  • Responsible for testing and release process for data pipelines using best practices for frequent releases.
  • Participate in Code Reviews
  • Mentor and support Data team members new to integrating the modern stack.
  • Partner to deliver a data model that aligns with DevOps principles, ensuring and standards for continuous integration/ continuous delivery (CI/CD) processes.
  • Drive Results by motivating self and others to exceed goals and achieve breakthrough results while exhibiting persistence to remove barriers to achieving results.
  • Develop and maintain data documentation, including data dictionaries, data lineage, and data flow diagrams, best practices and data recovery processes to provide clear visibility into the data ecosystem.

Preferred Qualifications

  • Advanced experience working with large data sets and streaming data.
  • Basic experience with various patterns of data ingestion, processing, and curation along with various streaming data concepts, such as Kafka.
  • Intermediate experience protecting PPI or PHI data during the ELT process, data security and data access controls and design.
  • Prior experience as a tech-lead or team lead.
  • Exposure to industry standard BI tools like Power BI and Looker.
  • Healthcare / Medical Devices domain experience will be an added advantage.