We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Current Behavior When running v2 excel pySpark code below in Databricks 13.3 LTS Runtime:
df = spark.read.format("excel") .option("header", True) .option("inferSchema", True) .load(fr"{folderpath}//.xlsx") display(df)
I receive the following error upon attempting to display or use the resulting dataframe:
AbstractMethodError: org.apache.spark.sql.execution.datasources.v2.FilePartitionReaderFactory.options()Lorg/apache/spark/sql/catalyst/FileSourceOptions;
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3) (10.178.42.202 executor 0): java.lang.AbstractMethodError: org.apache.spark.sql.execution.datasources.v2.FilePartitionReaderFactory.options()Lorg/apache/spark/sql/catalyst/FileSourceOptions;
This issue is same as #682 that was addressed for older versions.
The resulting dataframe should display the data.
set the folderpath variable to a location containing excel files, and run the below python code in latest runtime of Databricks:
- Spark version: 3.4.1 - Spark-Excel version: 0.18.7 - OS: N/A - Cluster environment
No response
The text was updated successfully, but these errors were encountered:
Please check these potential duplicates:
Sorry, something went wrong.
Please always try the newest version before creating issues. Closing this until the issue is reproduced with the newest version.
No branches or pull requests
Is there an existing issue for this?
Current Behavior
Current Behavior
When running v2 excel pySpark code below in Databricks 13.3 LTS Runtime:
df = spark.read.format("excel")
.option("header", True)
.option("inferSchema", True)
.load(fr"{folderpath}//.xlsx")
display(df)
I receive the following error upon attempting to display or use the resulting dataframe:
AbstractMethodError: org.apache.spark.sql.execution.datasources.v2.FilePartitionReaderFactory.options()Lorg/apache/spark/sql/catalyst/FileSourceOptions;
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3) (10.178.42.202 executor 0): java.lang.AbstractMethodError: org.apache.spark.sql.execution.datasources.v2.FilePartitionReaderFactory.options()Lorg/apache/spark/sql/catalyst/FileSourceOptions;
This issue is same as #682 that was addressed for older versions.
Expected Behavior
The resulting dataframe should display the data.
Steps To Reproduce
set the folderpath variable to a location containing excel files, and run the below python code in latest runtime of Databricks:
df = spark.read.format("excel")
.option("header", True)
.option("inferSchema", True)
.load(fr"{folderpath}//.xlsx")
display(df)
Environment
Anything else?
No response
The text was updated successfully, but these errors were encountered: