In addition to the answers above, we can also observe when an error arrives, for example
In my case, I had two types of data columns that caused the error: json and binary
for the json column, an error occurred when the Java class was executed at the very beginning of the import process:
/04/19 09:37:58 ERROR orm.ClassWriter: Cannot resolve SQL type
for a binary column , an error was selected when importing into hive tables (after importing data and adding it to HDFS files)
16/04/19 09:51:22 ERROR tool.ImportTool: Encountered IOException running import job: java.io.IOException: Hive does not support the SQL type for column featured_binary
To get rid of these two errors, I had to provide the following parameters
--map-column-java column1_json=String,column2_json=String,featured_binary=String --map-column-hive column1_json=STRING,column2_json=STRING,featured_binary=STRING
Therefore, we may need to provide
--map-column-java
or
--map-column-hive
depending on the failure.
source share