I am trying to perform an ETL job on AWS using Glue and pySpark, but unfortunately, I'm really new to this.
For the most part I dont have any problem working with glue dynamic dataframe to perform applymapping and some of the other transformations that I must perform. But I am facing a problem with a particular column that I must convert from string to integer array. In this column, value, we have the datatype set as string that is infact an array of integers converted to string and separated by space, for example a data entry in the value column looks like '111 222 333 444 555 666'. I must convert this column to be an integer array so that my data is transformed into '[111, 222, 333, 444, 555, 666]'.
How can I achieve this in AWS Glue and using pySpark? Any help is really appreciated.