Keen on solving an interesting data
problem to fast-track your application?
Pssst. Want to skip the queue and 2X your chances of being hired?
Set yourself apart by solving a real world problem that simulates the work we do at Teraflow and wow us with your technical skills!
The Challenge
Consulting on a Banking group's Data Engineering project,
your mission is to create a monthly extract
of data from their Aurora cluster into a CSV file on S3 for
consumption by an external system.#challengeaccepted
Your Solution:
AWS Glue (Jobs, Crawlers, Data Catalogues, ETL Scripts)
Terraform
PySpark/SQL
Solution Design
*Please note that the Aurora instance details don't point to a
real instance and that this is mainly a theoretical exercise.
We don't expect you to submit proof that you ran your code
in a real AWS environment. We mainly want to assess the
way you think and problem solve.
Submit a pull request into the parent repo once you’ve
completed the assessment
(naming convention for your development branch should be: NameSurname, or FirstnameSecondnameSurname so we can easily link your submission back to your application form)