Analyse and organise raw data Design develop and maintain data pipelines using AWS technologies such as AWS Glue, Airflow, Redshift and databricks Implement data modelling techniques to optimize data and storage and retrieval process. Must have knowledge of Python , Pyspark and SQL Familiarity with code versioning using GIT and Jenkins Evaluate business needs and objectives Conduct complex data analysis and report on results Build algorithms and prototypes Explore ways to enhance data quality and reliability Develop analytical tools and programs Collaborate with data scientists and architects on several projects Hands-on experience with SQL database design Good to have AWS or Databricks associated certification