@@ -48,11 +48,7 @@ object TypedBigQueryIT {
48
48
timestamp : Instant ,
49
49
date : LocalDate ,
50
50
time : LocalTime ,
51
- // BQ DATETIME is problematic with avro as BQ api uses different representations:
52
- // - BQ export uses 'string(datetime)'
53
- // - BQ load uses 'long(local-timestamp-micros)'
54
- // BigQueryType avroSchema favors read with string type
55
- // datetime: LocalDateTime,
51
+ datetime : LocalDateTime ,
56
52
geography : Geography ,
57
53
json : Json ,
58
54
bigNumeric : BigNumeric
@@ -81,23 +77,12 @@ object TypedBigQueryIT {
81
77
y <- Gen .numChar
82
78
} yield Geography (s " POINT( $x $y) " )
83
79
)
84
- implicit val arbJson : Arbitrary [Json ] = Arbitrary {
85
- import Arbitrary ._
86
- import Gen ._
87
- Gen
88
- .oneOf(
89
- // json object
90
- alphaLowerStr.flatMap(str => arbInt.arbitrary.map(num => s """ {" $str": $num} """ )),
91
- // json array
92
- alphaLowerStr.flatMap(str => arbInt.arbitrary.map(num => s """ [" $str", $num] """ )),
93
- // json literals
94
- alphaLowerStr.map(str => s """ " $str" """ ),
95
- arbInt.arbitrary.map(_.toString),
96
- arbBool.arbitrary.map(_.toString),
97
- Gen .const(" null" )
98
- )
99
- .map(wkt => Json (wkt))
100
- }
80
+ implicit val arbJson : Arbitrary [Json ] = Arbitrary (
81
+ for {
82
+ key <- Gen .alphaStr
83
+ value <- Gen .alphaStr
84
+ } yield Json (s """ {" $key":" $value"} """ )
85
+ )
101
86
102
87
implicit val arbBigNumeric : Arbitrary [BigNumeric ] = Arbitrary {
103
88
// Precision: 76.76 (the 77th digit is partial)
@@ -119,7 +104,7 @@ object TypedBigQueryIT {
119
104
private val tableRowTable = table(" records_tablerow" )
120
105
private val avroTable = table(" records_avro" )
121
106
122
- private val records = Gen .listOfN(5 , recordGen).sample.get
107
+ private val records = Gen .listOfN(100 , recordGen).sample.get
123
108
private val options = PipelineOptionsFactory
124
109
.fromArgs(
125
110
" --project=data-integration-test" ,
@@ -131,9 +116,8 @@ object TypedBigQueryIT {
131
116
class TypedBigQueryIT extends PipelineSpec with BeforeAndAfterAll {
132
117
import TypedBigQueryIT ._
133
118
134
- private val bq = BigQuery .defaultInstance()
135
-
136
119
override protected def afterAll (): Unit = {
120
+ val bq = BigQuery .defaultInstance()
137
121
// best effort cleanup
138
122
Try (bq.tables.delete(typedTable.ref))
139
123
Try (bq.tables.delete(tableRowTable.ref))
@@ -156,14 +140,6 @@ class TypedBigQueryIT extends PipelineSpec with BeforeAndAfterAll {
156
140
runWithRealContext(options) { sc =>
157
141
sc.parallelize(records)
158
142
.map(Record .toTableRow)
159
- .map { row =>
160
- // TableRow BQ save API uses json
161
- // TO disambiguate from literal json string,
162
- // field MUST be converted to parsed JSON
163
- val jsonLoadRow = new TableRow ()
164
- jsonLoadRow.putAll(row.asInstanceOf [java.util.Map [String , _]]) // cast for 2.12
165
- jsonLoadRow.set(" json" , Json .parse(row.getJson(" json" )))
166
- }
167
143
.saveAsBigQueryTable(
168
144
tableRowTable,
169
145
schema = Record .schema,
@@ -177,9 +153,9 @@ class TypedBigQueryIT extends PipelineSpec with BeforeAndAfterAll {
177
153
}
178
154
}
179
155
180
- it should " handle records as avro format" in {
156
+ // TODO fix if in beam 2.61
157
+ ignore should " handle records as avro format" in {
181
158
implicit val coder : Coder [GenericRecord ] = avroGenericRecordCoder(Record .avroSchema)
182
-
183
159
runWithRealContext(options) { sc =>
184
160
sc.parallelize(records)
185
161
.map(Record .toAvro)
@@ -191,8 +167,7 @@ class TypedBigQueryIT extends PipelineSpec with BeforeAndAfterAll {
191
167
}.waitUntilFinish()
192
168
193
169
runWithRealContext(options) { sc =>
194
- val data =
195
- sc.bigQueryTable(avroTable, Format .GenericRecordWithLogicalTypes ).map(Record .fromAvro)
170
+ val data = sc.bigQueryTable(avroTable, Format .GenericRecord ).map(Record .fromAvro)
196
171
data should containInAnyOrder(records)
197
172
}
198
173
}
0 commit comments