|
17 | 17 | */
|
18 | 18 | package org.apache.hadoop.hive.serde2.avro;
|
19 | 19 |
|
| 20 | +import static org.apache.avro.Schema.Field.NULL_DEFAULT_VALUE; |
| 21 | + |
20 | 22 | import org.apache.avro.Schema;
|
21 | 23 | import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector;
|
22 | 24 | import org.apache.hadoop.hive.serde2.typeinfo.CharTypeInfo;
|
|
28 | 30 | import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
|
29 | 31 | import org.apache.hadoop.hive.serde2.typeinfo.UnionTypeInfo;
|
30 | 32 | import org.apache.hadoop.hive.serde2.typeinfo.VarcharTypeInfo;
|
31 |
| -import org.codehaus.jackson.JsonNode; |
32 |
| -import org.codehaus.jackson.node.JsonNodeFactory; |
33 | 33 |
|
34 | 34 | import java.util.ArrayList;
|
35 | 35 | import java.util.Arrays;
|
@@ -235,14 +235,13 @@ private Schema createAvroArray(TypeInfo typeInfo) {
|
235 | 235 | private List<Schema.Field> getFields(Schema.Field schemaField) {
|
236 | 236 | List<Schema.Field> fields = new ArrayList<Schema.Field>();
|
237 | 237 |
|
238 |
| - JsonNode nullDefault = JsonNodeFactory.instance.nullNode(); |
239 | 238 | if (schemaField.schema().getType() == Schema.Type.RECORD) {
|
240 | 239 | for (Schema.Field field : schemaField.schema().getFields()) {
|
241 |
| - fields.add(new Schema.Field(field.name(), field.schema(), field.doc(), nullDefault)); |
| 240 | + fields.add(new Schema.Field(field.name(), field.schema(), field.doc(), NULL_DEFAULT_VALUE)); |
242 | 241 | }
|
243 | 242 | } else {
|
244 | 243 | fields.add(new Schema.Field(schemaField.name(), schemaField.schema(), schemaField.doc(),
|
245 |
| - nullDefault)); |
| 244 | + NULL_DEFAULT_VALUE)); |
246 | 245 | }
|
247 | 246 |
|
248 | 247 | return fields;
|
|
0 commit comments