top of page

Automate Data Type Standardization for patient_data_bronze via S3 Mapping File

Requirement

Requirement:

 

Develop a Databricks PySpark script to perform dynamic data type conversion on the patient_data_bronze table using a JSON file stored in S3. Read the JSON file follows in S3 which has the format → {data_type: column_name} (eg: STRING: <column_name>). Display the output after applying the conversions.

 

 

 

S3 path: s3://agilisium-playground-dev/filestore/mapping-files/patient_datatype_mapping.json.

 

Databricks Secret Information: The access_key and secret_key have been stored in Databricks secrets under the scope named aws_keys.

Purgo AI Agentic Code

bottom of page