Sagemaker processing job with PySpark and Step Functions
问题 this is my problem: I have to run a Sagemaker processing job using custom code written in PySpark. I've used the Sagemaker SDK by running these commands: spark_processor = sagemaker.spark.processing.PySparkProcessor( base_job_name="spark-preprocessor", framework_version="2.4", role=role_arn, instance_count=2, instance_type="ml.m5.xlarge", max_runtime_in_seconds=1800, ) spark_processor.run( submit_app="processing.py", arguments=['s3_input_bucket', bucket_name, 's3_input_file_path', file_path ]