Data Engineer

Honeysuckle Health
Newcastle, Maitland & Hunter
Information & Communication Technology
Engineering - Software
Full time
Posted 30d ago

About Us

Honeysuckle Health exists to help people lead healthier lives. As a healthcare services company we develop and deliver digital and telephonic health programs, health services contracting and healthcare analytics.  We have an exciting opportunity for a Mid Data Engineer to join Honeysuckle Health.

A career at Honeysuckle Health presents a unique and exciting opportunity to provide support to health consumers, health funds and providers in the best possible application of health care services.


The Role

As a Data Engineer you must possess passion for learning and growing in the Data Science and Artificial Intelligence space, thus using technology to drive better health care solutions - When it comes to buying a car it's easier to know what one is paying for, however in healthcare there is a big gap in value vs outcomes. You will be working in a highly collaborative environment, where people learn from each other, understand different perspectives and deliver outcomes fast, to create better pathways to better health. Your responsibilities may include but not limited to below:
 

Core Responsibilities:

  • Taking ownership of business-critical data pipelines, refactor them and improve the design if needed
  • Making sure pipelines adhere to good ops practices and are well logged and reliable
  • Define and build end to end monitoring into our pipelines, from time of data extraction all the way through consumption
  • Define and co build data quality framework to deliver acceptable quality data for data science use cases. e.g. Data modelling, and data quality test cases to make sure data is not stale, accurate and complete
  • Review and improve designs, practices and tooling on an ongoing basis

Secondary Responsibilities:

  • Selecting sound technology for a given data science use case, e.g. choosing between Sagemaker vs Databricks for a use case
  • Writing modular code, leveraging software engineering practices and bringing to data science domain, e.g. Writing configs, reproducible pipeline code 
  • Co-designing MLOPs framework with data scientists and help the team to deliver ROI with machine learning and AI. e.g. Monitoring ML pipeline, and automating training due to data degradation
  • Tuning data science workloads for faster feedback. e.g. Choosing high compute distributed cluster to parallelize model training
  • Design practices around model packaging and serving via Rest API on cloud. e.g. Choosing Django or Fast Api for model serving


About your skills:

Essential (Commercial not necessary, we appreciate personal experience):

  • Experience building out and maintaining data pipelines (ETL/ELT) within AWS using Python/Spark/Scala. Familiarity with data partitioning, clustering in distributed environment.
  • Experience with orchestrating pipelines with open-sourced tools such as - Airflow/Luigi/Gitlab/Oozie
  • Experience with any top data warehousing tools – (Bigquery, Snowflake, Redshift)
  • Ability to communicate technical ideas clearly to other team members, business stake holders.
  • Experience with writing ymls and defining complex workflows for CICD.
  • A bachelor/masters diploma/degree in technology, data science or analytics
  • Experience with any or combination of three machine learning framework – Pytorch/Scikit learn/TensorFlow
  • Building containerised applications on cloud (Docker, Docker Compose, ECS, EKS)
  • Knowledge of best practices for large sets of data, writing advanced queries.
  • Experience with infrastructure as code
  • Previous commercial experience on any three major clouds – (GCP, AWS, AZURE)

Desirable:

  • A bachelor/masters diploma/degree in technology, data science or analytics
  • Experience with any or combination of three machine learning framework – Pytorch/Scikit learn/TensorFlow
  • Building containerised applications on cloud (Docker, Docker Compose, ECS, EKS)
  • Knowledge of best practices for large sets of data, writing advanced queries.
  • Experience with infrastructure as code 
  • Monitoring and Deploying Machine learning models in a commercial setting
  • Building and consuming REST or Grah APIs
  • Statistical knowledge and tuning machine learning algorithm
  • Streaming data concepts and technologies – in any (Kafka, Kinesis, Beam, Pubsub) would do


Our Stack (100% Cloud):

  • Snowflake for Datawarehousing and data share
  • DBT to interact and deploy our code on snowflake
  • AWS Codebuild and Codepipeline with GitHub for CICD
  • AWS serverless offerings for almost all our applications
  • AWS Sagemaker and Databricks for machine learning
  • Terraform for infra as code
  • Golang, Typescript, Python, Nodejs, Javascript for our application code
  • Quicksight, Tableau for data visualisations

 

What We Offer:

  • Training budget and time during the week for learning & development
  • Work with truly amazing people who have a passion for health care and improving the lives of our community members
  • Opportunity to participate in our short term incentive program
  • Good work health life balance - Inclusion in almost daily fitness activities with the team around Newcastle office site, ranging from swims, exercise sessions, runs, bike rides, etc
  • Mix of remote and on-site work. We are currently resourced to work from home, however once we are able to come together in our Newcastle CBD space, we may encourage you to travel occasionally
  • Other perks: Study assistance, Paid parental leave, Volunteer leave, Relocation budget for interstate or overseas individuals, Employee health insurance discounts

 

How to Apply

We strongly encourage applications from female candidates, Aboriginal and Torres Strait Islanders, people from culturally and linguistically diverse backgrounds and people with a disability, as we recognise that these groups are underrepresented throughout the technology industry.

If you're seeking professional growth and enjoy working on large, distributed, cloud-based applications, love coffee runs, and enjoy techy conversations, then apply now.

You must possess the right to work in Australia.

We’re not big on resumes: to apply, complete a short technical test, provide us with you contact details and LinkedIn profile, and we’ll get back to you if your profile matches our requirements.

Apply here: https://apply.hire.toggl.com/288VZ81JDKQ1D23HKK

Employer questions

Your application will include the following questions:
  • Which of the following statements best describes your right to work in Australia?
  • Which of the following statements best describes your Covid-19 vaccination status?
  • How many years' experience do you have as a Data Engineer?

Report this job advert

Be careful
Don’t provide your bank or credit card details when applying for jobs.
Learn how to protect yourself here.