I'm learning Glue with Pyspark by following this page: https://aws-dojo.com/ws8/labs/configure-crawler/.
My question is: is crawler & creating a database in Lake Formation required for creating a glue job?
I have some issue with my aws role and I'm not authorised to create resourse in LakeFormation, so I'm thinking if I can skip them to only create a glue job and test my script?
For example, I only want to test my pyspark script for one single input .txt file, I store it in S3, do I still need crawler? Can I just use boto3
to create a glue job to test the script and do some preprocessing and write data back to s3?