We are partnering with a global tech consultancy to find a Data Engineer for an initial 3 month contract. If you're someone who enjoys designing scalable data pipelines and building modern, high performance data platforms, this is a great opportunity for you! - Must be Brisbane based as this is a h ybrid work environment. - 12 week contract with the possibility to extend. - You will demonstrate your expertise and gain exposure to innovative tech within a short term, high impact project. What You'll Be Doing : Build scalable batch and streaming pipelines with Python and PySpark Design Delta Lake architectures on Databricks Orchestrate workflows and jobs in Databricks Tune performance and manage code with Databricks libraries Manage AWS S3 data lakes for secure data access Deploy infrastructure using Terraform or CloudFormation Automate AWS services using Boto3 Collaborate across teams to ensure data reliability Maintain data quality and observability standards What We're Looking For : Proven experience in data engineering with Databricks and AWS. Strong programming skills in Python and PySpark. Hands-on experience with Delta Lake and structured streaming. Deep understanding of data lake architecture and ETL pipeline design. Experience with Terraform, CloudFormation, or similar IaC tools. Strong problem-solving skills and ability to work autonomously in a fast-paced environment. Excellent communication and collaboration abilities. Experience working in a multi cloud or multi account AWS environment will be highly regarded. Familiarity with data governance, data cataloguing, and security best practices is a bonus. Does this sound like you, please apply here!