






YOUR JOB APPLICATION






Keen on solving an interesting data problem to fast-track your application?
Pssst. Want to skip the queue and 2X your chances of being hired?
Set yourself apart by solving a real world problem that simulates the work we do at Teraflow and wow us with your technical skills!
The Challenge
Consulting on a Banking group's Data Engineering project, your mission is to create a monthly extract of data from their Aurora cluster into a CSV file on S3 for consumption by an external system. #challengeaccepted

Your Solution:
- AWS Glue (Jobs, Crawlers, Data Catalogues, ETL Scripts)
- Terraform
- PySpark/SQL
- Solution Design
*Please note that the Aurora instance details don't point to a real instance and that this is mainly a theoretical exercise. We don't expect you to submit proof that you ran your code in a real AWS environment. We mainly want to assess the way you think and problem solve.



Fork this repo on Github
Or it might just GIT forked by someone else.
Submit a pull request into the parent repo once you’ve
completed the assessment
(naming convention for your development branch should be: NameSurname, or FirstnameSecondnameSurname so we can easily link your submission back to your application form)
