問題
Databricks在擬合SparkML模型或Pipeline時拋出錯誤:
org.apache.spark.SparkException: Job abort to stage failure: Task 0 in stage 162.0 failed 4 times,最近的failure: Lost Task 0.3 in stage 162.0 (TID 168, 10.205.250.130, executor 1): org.apache.spark.SparkException: failed to execute user defined function($anonfun$9: (string) =>雙)
導致
通常,在擬合SparkML模型或Pipeline時的錯誤是訓練數據問題的結果。
解決方案
檢查以下問題: