Upgrading from PySpark 2.3 to 2.4ΒΆ
In PySpark, when Arrow optimization is enabled, previously
toPandasjust failed when Arrow optimization is unable to be used whereascreateDataFramefrom Pandas DataFrame allowed the fallback to non-optimization. Now, bothtoPandasandcreateDataFramefrom Pandas DataFrame allow the fallback by default, which can be switched off byspark.sql.execution.arrow.fallback.enabled.