File tree 6 files changed +17
-14
lines changed
main/scala/ai/chronon/integrations/cloud_gcp
test/scala/ai/chronon/integrations/cloud_gcp
main/scala/ai/chronon/spark
test/scala/ai/chronon/spark/test 6 files changed +17
-14
lines changed Original file line number Diff line number Diff line change @@ -224,7 +224,6 @@ lazy val cloud_gcp = project
224
224
libraryDependencies += " org.json4s" %% " json4s-native" % " 3.7.0-M11" ,
225
225
libraryDependencies += " org.json4s" %% " json4s-core" % " 3.7.0-M11" ,
226
226
libraryDependencies += " org.yaml" % " snakeyaml" % " 2.3" ,
227
- // libraryDependencies += "com.google.cloud.spark.bigtable" %% "spark-bigtable" % "0.2.1",
228
227
libraryDependencies += " com.google.cloud.bigtable" % " bigtable-hbase-2.x" % " 2.14.2" ,
229
228
libraryDependencies ++= avro,
230
229
libraryDependencies ++= spark_all_provided,
Original file line number Diff line number Diff line change @@ -102,7 +102,7 @@ object DataprocSubmitter {
102
102
.map((jVal) => render(jVal))
103
103
.map(compact)
104
104
.map(parse(_).extract[SubmitterConf ])
105
- .getOrElse(throw new IllegalArgumentException (s " Yaml conf not found or invalid yaml " ))
105
+ .getOrElse(throw new IllegalArgumentException (" Yaml conf not found or invalid yaml" ))
106
106
107
107
}
108
108
}
Original file line number Diff line number Diff line change @@ -5,13 +5,13 @@ import ai.chronon.spark.TableUtils
5
5
import com .google .cloud .hadoop .fs .gcs .GoogleHadoopFS
6
6
import com .google .cloud .hadoop .fs .gcs .GoogleHadoopFileSystem
7
7
import com .google .cloud .hadoop .fs .gcs .GoogleHadoopFileSystemConfiguration
8
+ import com .google .cloud .hadoop .fs .gcs .HadoopConfigurationProperty
9
+ import com .google .cloud .hadoop .gcsio .GoogleCloudStorageFileSystem
8
10
import org .apache .spark .sql .SparkSession
9
11
import org .junit .Assert .assertEquals
10
12
import org .junit .Assert .assertTrue
11
13
import org .scalatest .funsuite .AnyFunSuite
12
14
import org .scalatestplus .mockito .MockitoSugar
13
- import com .google .cloud .hadoop .gcsio .GoogleCloudStorageFileSystem
14
- import com .google .cloud .hadoop .fs .gcs .HadoopConfigurationProperty
15
15
16
16
class BigQueryCatalogTest extends AnyFunSuite with MockitoSugar {
17
17
Original file line number Diff line number Diff line change 1
1
package ai .chronon .integrations .cloud_gcp
2
2
3
3
import ai .chronon .spark .SparkSessionBuilder
4
- import org .apache .spark .sql .{Row , SaveMode , SparkSession }
4
+ import org .apache .spark .sql .Row
5
+ import org .apache .spark .sql .SaveMode
6
+ import org .apache .spark .sql .SparkSession
5
7
import org .apache .spark .sql .functions ._
6
- import org .apache .spark .sql .types .{StringType , StructField , StructType }
7
-
8
+ import org .apache .spark .sql .types .StringType
9
+ import org .apache .spark .sql .types .StructField
10
+ import org .apache .spark .sql .types .StructType
8
11
import org .junit .Assert .assertEquals
9
12
import org .scalatest .funsuite .AnyFunSuite
10
13
Original file line number Diff line number Diff line change @@ -49,11 +49,14 @@ import org.apache.spark.sql.streaming.StreamingQueryListener
49
49
import org .apache .spark .sql .streaming .StreamingQueryListener .QueryProgressEvent
50
50
import org .apache .spark .sql .streaming .StreamingQueryListener .QueryStartedEvent
51
51
import org .apache .spark .sql .streaming .StreamingQueryListener .QueryTerminatedEvent
52
+ import org .json4s ._
53
+ import org .json4s .jackson .JsonMethods ._
52
54
import org .rogach .scallop .ScallopConf
53
55
import org .rogach .scallop .ScallopOption
54
56
import org .rogach .scallop .Subcommand
55
57
import org .slf4j .Logger
56
58
import org .slf4j .LoggerFactory
59
+ import org .yaml .snakeyaml .Yaml
57
60
58
61
import java .io .File
59
62
import java .nio .file .Files
@@ -69,9 +72,6 @@ import scala.reflect.internal.util.ScalaClassLoader
69
72
import scala .util .Failure
70
73
import scala .util .Success
71
74
import scala .util .Try
72
- import org .json4s ._
73
- import org .json4s .jackson .JsonMethods ._
74
- import org .yaml .snakeyaml .Yaml
75
75
76
76
// useful to override spark.sql.extensions args - there is no good way to unset that conf apparently
77
77
// so we give it dummy extensions
Original file line number Diff line number Diff line change @@ -18,17 +18,18 @@ package ai.chronon.spark.test
18
18
19
19
import ai .chronon .spark .Driver .OfflineSubcommand
20
20
import org .apache .spark .sql .SparkSession
21
+ import org .json4s ._
22
+ import org .json4s .jackson .JsonMethods ._
21
23
import org .junit .Assert .assertEquals
22
24
import org .junit .Assert .assertTrue
23
25
import org .junit .Test
24
26
import org .rogach .scallop .ScallopConf
25
- import org .json4s ._
26
- import org .json4s .jackson .JsonMethods ._
27
27
import org .yaml .snakeyaml .Yaml
28
- import collection .JavaConverters ._
29
28
30
29
import scala .io .Source
31
30
31
+ import collection .JavaConverters ._
32
+
32
33
class OfflineSubcommandTest {
33
34
34
35
class TestArgs (args : Array [String ]) extends ScallopConf (args) with OfflineSubcommand {
@@ -83,7 +84,7 @@ class OfflineSubcommandTest {
83
84
.map((jVal) => render(jVal))
84
85
.map(compact)
85
86
.map(parse(_).extract[Map [String , String ]])
86
- .getOrElse(throw new IllegalArgumentException (s " Yaml conf not found or invalid yaml " ))
87
+ .getOrElse(throw new IllegalArgumentException (" Yaml conf not found or invalid yaml" ))
87
88
88
89
val confKey = " spark.chronon.table.format_provider.class"
89
90
assertEquals(confs.get(confKey), sparkSession.conf.getOption(confKey))
You can’t perform that action at this time.
0 commit comments