You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
The RAPIDS plugin added support to insertIntoHiveTable NVIDIA/spark-rapids#10912 . However, the dataSource.csv did not add a new entry to indicate that.
InsertHiveTable in the eventlog shows the serDe library but I don't see the outputFormatClass.
How to find whether a table is bucketed or not ?
the qual prediction might be off if the user disable spark.sql.hive.convertMetastoreParquet on the GPU. Even if we recommend turning that on, there is no guarantee users can do it due to compatibility issues.
Other Formats:
I also see that HiveORC is converted into GPU. We might need to change the format to be "ORC" when the Spark configurations use native ORC.
InsertIntoHadoopFsRelationCommand HiveORC
# this will be transformed into
GpuInsertIntoHadoopFsRelationCommand HiveORC
I kept eventlog representing that case in my scratch-folder under the same issue number (app-20250305030545-0000.zstd).
Is your feature request related to a problem? Please describe.
The RAPIDS plugin added support to insertIntoHiveTable NVIDIA/spark-rapids#10912 . However, the dataSource.csv did not add a new entry to indicate that.
According to NVIDIA/spark-rapids#10912, the HiveParquet is supported when:
The problems:
spark.sql.hive.convertMetastoreParquet
on the GPU. Even if we recommend turning that on, there is no guarantee users can do it due to compatibility issues.Other Formats:
I also see that HiveORC is converted into GPU. We might need to change the format to be "
ORC
" when the Spark configurations use native ORC.I kept eventlog representing that case in my scratch-folder under the same issue number (
app-20250305030545-0000.zstd
).Related issues
#1583
The text was updated successfully, but these errors were encountered: