1
1
package ai .chronon .integrations .cloud_gcp
2
2
3
- import com .google .cloud .spark .bigquery .BigQueryCatalog
4
3
import com .google .cloud .bigquery .{
5
4
BigQuery ,
6
5
BigQueryOptions ,
@@ -9,10 +8,10 @@ import com.google.cloud.bigquery.{
9
8
TableDefinition ,
10
9
TableId
11
10
}
11
+ import com .google .cloud .spark .bigquery .BigQueryCatalog
12
12
import org .apache .iceberg .gcp .bigquery .BigQueryMetastoreCatalog
13
13
import org .apache .iceberg .spark .SparkCatalog
14
14
import org .apache .spark .sql .SparkSession
15
- import org .apache .spark .sql .catalyst .analysis .NoSuchTableException
16
15
import org .apache .spark .sql .connector .catalog ._
17
16
import org .apache .spark .sql .connector .catalog .functions .UnboundFunction
18
17
import org .apache .spark .sql .connector .expressions .Transform
@@ -23,6 +22,7 @@ import org.apache.spark.sql.util.CaseInsensitiveStringMap
23
22
24
23
import java .util
25
24
import scala .jdk .CollectionConverters ._
25
+ import org .apache .spark .sql .catalyst .analysis .NoSuchTableException
26
26
import scala .util .{Failure , Success , Try }
27
27
28
28
/** Galactus catalog that allows us to interact with BigQuery metastore as a spark catalog. This allows for
@@ -126,7 +126,6 @@ class DelegatingBigQueryMetastoreCatalog extends TableCatalog with SupportsNames
126
126
// ideally it should be the below:
127
127
// val connectorTable = connectorCatalog.loadTable(ident)
128
128
connectorCatalog.loadTable(Identifier .of(Array (tId.getDataset), tId.getTable))
129
-
130
129
}
131
130
case _ => throw new IllegalStateException (s " Cannot support table of type: ${table.getDefinition}" )
132
131
}
0 commit comments