public final enum class org.apache.hadoop.hive.ql.ErrorMsg extends java.lang.Enum
{
public static final enum org.apache.hadoop.hive.ql.ErrorMsg GENERIC_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_OR_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_TABLE_OR_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_PARTITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_TABLE_ALIAS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_ALIAS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_TABLE_ALIAS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_FUNCTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_FUNCTION_SIGNATURE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_OPERATOR_SIGNATURE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT_LENGTH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_1;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_2;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_3;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TRANSFORM;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_MULTIPLE_DISTINCTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_SUBQUERY_ALIAS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_INSERT_INSUBQUERY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NON_KEY_EXPR_IN_GROUPBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_XPATH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_PATH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ILLEGAL_PATH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_NUMERICAL_CONSTANT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARRAYINDEX_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPINDEX_CONSTANT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPINDEX_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NON_COLLECTION_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SELECT_DISTINCT_WITH_GROUPBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMN_REPEATED_IN_PARTITIONING_COLS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DUPLICATE_COLUMN_NAMES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_BUCKET_NUMBER;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMN_REPEATED_IN_CLUSTER_SORT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SAMPLE_RESTRICTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SAMPLE_COLUMN_NOT_FOUND;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_PARTITION_PREDICATE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_DOT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TBL_DDL_SERDE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TARGET_TABLE_COLUMN_MISMATCH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TABLE_ALIAS_NOT_ALLOWED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_DISTRIBUTEBY_CONFLICT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ORDERBY_DISTRIBUTEBY_CONFLICT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_SORTBY_CONFLICT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ORDERBY_SORTBY_CONFLICT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_ORDERBY_CONFLICT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_LIMIT_WITH_ORDERBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNION_NOTIN_SUBQ;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_INPUT_FORMAT_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_OUTPUT_FORMAT_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_VALID_PARTN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_OUTER_MAPJOIN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPJOIN_HINT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPJOIN_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NON_BUCKETED_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg BUCKETED_NUMERATOR_BIGGER_DENOMINATOR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NEED_PARTITION_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_CTLT_COEXISTENCE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LINES_TERMINATED_BY_NON_NEWLINE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_COLLST_COEXISTENCE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTLT_COLLST_COEXISTENCE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_SELECT_SCHEMA;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_PARCOL_COEXISTENCE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_MULTI_LOADFILE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_EXTTBL_COEXISTENCE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INSERT_EXTERNAL_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DATABASE_NOT_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TABLE_ALREADY_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMN_ALIAS_ALREADY_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_MULTIPLE_EXPR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_REQUIRE_AS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_GROUP_BY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_SORT_BY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_CLUSTER_BY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_DISTRIBUTE_BY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_INVALID_LOCATION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_LATERAL_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDTF_ALIAS_MISMATCH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDF_STATEFUL_INVALID_LOCATION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LATERAL_VIEW_WITH_JOIN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LATERAL_VIEW_INVALID_CHILD;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OUTPUT_SPECIFIED_MULTIPLE_TIMES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_AS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg VIEW_COL_MISMATCH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DML_AGAINST_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ANALYZE_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg VIEW_PARTITION_TOTAL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg VIEW_PARTITION_MISMATCH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTITION_DYN_STA_ORDER;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITION_DISABLED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITION_STRICT_MODE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NONEXISTPARTCOL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CREATE_NON_NATIVE_AS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOAD_INTO_NON_NATIVE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCKMGR_NOT_SPECIFIED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCKMGR_NOT_INITIALIZED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCK_CANNOT_BE_ACQUIRED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ZOOKEEPER_CLIENT_COULD_NOT_BE_INITIALIZED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OVERWRITE_ARCHIVED_PART;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_METHODS_DISABLED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_ON_MULI_PARTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNARCHIVE_ON_MULI_PARTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_ON_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg RESERVED_PART_VAL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OFFLINE_TABLE_OR_PARTITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NEED_PARTITION_SPECIFICATION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_METADATA;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NEED_TABLE_SPECIFICATION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTITION_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TABLE_DATA_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INCOMPATIBLE_SCHEMA;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg EXIM_FOR_NON_NATIVE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_BUCKETIZED_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTSPEC_DIFFER_FROM_SCHEMA;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTITION_COLUMN_NON_PRIMITIVE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_DYNAMICPARTITION_IFNOTEXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UDAF_INVALID_LOCATION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DROP_PARTITION_NON_STRING_PARTCOLS_NONEQUALITY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_COMMAND_FOR_VIEWS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_COMMAND_FOR_TABLES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_VIEW_DISALLOWED_OP;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NON_NATIVE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SORTMERGE_MAPJOIN_FAILED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg BUCKET_MAPJOIN_NOT_POSSIBLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg BUCKETED_TABLE_METADATA_INCORRECT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg JOINNODE_OUTERJOIN_MORETHAN_16;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_JDO_FILTER_EXPRESSION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_BUCKETNUM_NONBUCKETIZED_TBL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_FOR_NON_MANAGED_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_FOR_NON_NATIVE_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTSPEC_FOR_NON_PARTITIONED_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_IN_ON_CLAUSE_OF_MERGE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOAD_INTO_STORED_AS_DIR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_STOREDASDIR_NOT_SKEWED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_SKEWED_LOC_NO_LOC;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_SKEWED_LOC_NO_MAP;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_NO_COLUMN_NAME;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_NO_COLUMN_VALUE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_DUPLICATE_COLUMN_NAMES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_INVALID_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_1;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_2;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_3;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NOT_ALLOWED_RENAME_SKEWED_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_NOMAPAGGR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_EXPRESSION_INVALID;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_EMPTY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_UNION_REMOVE_OPTIMIZATION_NEEDS_SUBDIRECTORIES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_EXPR_NOT_IN_GROUPBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_PARTITION_SPEC;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_UNSET_NON_EXIST_PROPERTY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_VIEW_AS_SELECT_NOT_EXIST;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPLACE_VIEW_WITH_PARTITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg EXISTING_TABLE_IS_NOT_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_SUPPORTED_ORDERBY_ALLCOLREF_POS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_POSITION_ALIAS_IN_GROUPBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_POSITION_ALIAS_IN_ORDERBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_SKEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_DISTINCTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OPERATOR_NOT_ALLOWED_WITH_MAPJOIN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ANALYZE_TABLE_NOSCAN_NON_NATIVE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTITION_VALUE_NOT_CONTINUOUS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TABLES_INCOMPATIBLE_SCHEMAS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg EXCHANGE_PARTITION_NOT_ALLOWED_WITH_TRANSACTIONAL_TABLES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_COLUMN_NOT_RC;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_COLUMN_ARCHIVED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_BUCKETED_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_LIST_BUCKETED_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TABLE_NOT_PARTITIONED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DATABASE_ALREADY_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CANNOT_REPLACE_COLUMNS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg BAD_LOCATION_VALUE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_ALTER_TBL_OP;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_BIGTABLE_MAPJOIN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MISSING_OVER_CLAUSE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARTITION_SPEC_TYPE_MISMATCH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_SUBQUERY_EXPRESSION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_SUBQUERY_EXPRESSION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_HDFS_URI;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_DIR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_VALID_LOCATIONS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_PRINCIPAL_TYPE_GROUP;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_NAME;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_IMMUTABLE_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_GLOBAL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_SPECIFIED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_INSTANTIATED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TXN_NO_SUCH_TRANSACTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TXN_ABORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DBTXNMGR_REQUIRES_CONCURRENCY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_ACID;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCK_NO_SUCH_LOCK;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCK_REQUEST_UNSUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg METASTORE_COMMUNICATION_FAILED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg METASTORE_COULD_NOT_INITIATE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_COMPACTION_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NO_COMPACTION_PARTITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TOO_MANY_COMPACTION_PARTITIONS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DISTINCT_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NONACID_COMPACTION_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MASKING_FILTERING_ON_ACID_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MASKING_FILTERING_ON_MATERIALIZED_VIEWS_SOURCES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MASKING_COMPLEX_TYPE_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UPDATEDELETE_PARSE_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UPDATE_CANNOT_UPDATE_PART_VALUE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INSERT_CANNOT_CREATE_TEMP_FILE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_NONACID_TXNMGR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg VALUES_TABLE_CONSTRUCTOR_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_NONACID_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_NO_SORTED_BUCKETS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_TYPE_PARTIAL_PARTITION_SPEC_NO_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_PARTITION_CASCADE_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DROP_NATIVE_FUNCTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UPDATE_CANNOT_UPDATE_BUCKET_VALUE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg IMPORT_INTO_STRICT_REPL_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_LOCATION_NONEMPTY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTAS_CREATES_VOID_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TBL_SORTED_NOT_BUCKETED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCK_ACQUIRE_TIMEDOUT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COMPILE_LOCK_TIMED_OUT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_SERDE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_FILEFORMAT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CANNOT_REORDER_COLUMNS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_COLUMN_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPLACE_CANNOT_DROP_COLUMNS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPLACE_UNSUPPORTED_TYPE_CONVERSION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_NOMAPAGGR_MULTIGBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CANNOT_RETRIEVE_TABLE_METADATA;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_AST_TREE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ERROR_SERIALIZE_METASTORE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg IO_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ERROR_SERIALIZE_METADATA;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_LOAD_TABLE_FILE_WORK;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CLASSPATH_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg IMPORT_SEMANTIC_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_FK_SYNTAX;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_CSTR_SYNTAX;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_NOT_ENOUGH_HISTORY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_COLUMN_NAME;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_SET_OPERATOR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOCK_ACQUIRE_CANCELLED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NOT_RECOGNIZED_CONSTRAINT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_CONSTRAINT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPLACE_VIEW_WITH_MATERIALIZED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPLACE_MATERIALIZED_WITH_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UPDATE_DELETE_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MATERIALIZED_VIEW_DEF_EMPTY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MERGE_PREDIACTE_REQUIRED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MERGE_TOO_MANY_DELETE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg MERGE_TOO_MANY_UPDATE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_TARGET_COLUMN_IN_SET_CLAUSE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_FUNCTION_EXPR_NOT_IN_GROUPBY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NON_PARTITIONED_TABLE_CASCADE_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_SIZE_LIMIT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REBUILD_NO_MATERIALIZED_VIEW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOAD_DATA_ACID_FILE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_INSERTONLYTRAN_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg LOAD_DATA_LAUNCH_JOB_PARSE_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg RESOURCE_PLAN_ALREADY_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg RESOURCE_PLAN_NOT_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INCOMPATIBLE_STRUCT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OBJECTNAME_CONTAINS_DOT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg WITHIN_GROUP_NOT_ALLOWED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg WITHIN_GROUP_PARAMETER_MISMATCH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_STRUCT_ATTRIBUTE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OFFSET_NOT_SUPPORTED_IN_SUBQUERY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg WITH_COL_LIST_NUM_OVERFLOW;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NULL_TREATMENT_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DATACONNECTOR_ALREADY_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DATACONNECTOR_NOT_EXISTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TIME_TRAVEL_NOT_ALLOWED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_METADATA_TABLE_NAME;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg TABLE_META_REF_NOT_SUPPORTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COMPACTION_REFUSED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CBO_IS_REQUIRED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTLF_UNSUPPORTED_FORMAT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg NON_NATIVE_ACID_UPDATE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg READ_ONLY_DATABASE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_INIT_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_IO_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_GENERIC_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_CLOSING_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITIONS_TOO_MANY_PER_NODE_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_IN_IMPLICIT_TXN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_IN_TXN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_WITHOUT_TXN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACCESS_DENIED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg QUOTA_EXCEEDED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_PATH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg FILE_NOT_FOUND;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg WRONG_FILE_FORMAT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_FILE_MISSING_FROM_SRC_AND_CM_PATH;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_EXTERNAL_SERVICE_CONNECTION_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CLIENT_POLLING_OPSTATUS_INTERRUPTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTLF_FAILED_INFERENCE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTLF_CLASS_NOT_FOUND;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CTLF_MISSING_STORAGE_FORMAT_DESCRIPTOR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARQUET_FOOTER_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg PARQUET_UNHANDLED_TYPE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ORC_FOOTER_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_RT_EXCEPTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_NOT_OBTAINED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_INITIALIZATION_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_CONNECTION_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_PUBLISHING_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_CLOSING_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_PART_KEY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_PARTITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_PARSE_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg DROP_COMMAND_NOT_ALLOWED_FOR_PARTITION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_COLUMN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSAGGREGATOR_SOURCETASK_NULL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATSAGGREGATOR_CONNECTION_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg STATS_SKIPPING_BY_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg INVALID_FILE_FORMAT_IN_LOAD;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg SCHEMA_REQUIRED_TO_READ_ACID_TABLES;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_TABLES_MUST_BE_READ_WITH_ACID_READER;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_TABLES_MUST_BE_READ_WITH_HIVEINPUTFORMAT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg ACID_LOAD_DATA_INVALID_FILE_NAME;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_FILE_FORMAT;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_BUCKETED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_PARTITION_ARCHIVED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_NON_NATIVE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_NOT_MANAGED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_TRANSACTIONAL;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_FILE_SYSTEM_OPERATION_RETRY;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_SOURCE_DATABASE_NOT_FOUND;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_DATABASE_IS_TARGET_OF_REPLICATION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_DB_OR_TABLE_PATTERN;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_EVENTS_MISSING_IN_METASTORE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_BOOTSTRAP_LOAD_PATH_NOT_VALID;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_CONFIG_FOR_SERVICE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_INTERNAL_CONFIG_FOR_SERVICE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_RETRY_EXHAUSTED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_FAILED_WITH_NON_RECOVERABLE_ERROR;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_ARGUMENTS;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_ALTER_TABLE;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_PERMISSION_DENIED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_DISTCP_SNAPSHOT_EXCEPTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg RANGER_AUTHORIZATION_FAILED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg RANGER_AUTHENTICATION_FAILED;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_INCOMPATIBLE_EXCEPTION;
public static final enum org.apache.hadoop.hive.ql.ErrorMsg REPL_FAILOVER_TARGET_MODIFIED;
private int errorCode;
private java.lang.String mesg;
private java.lang.String sqlState;
private java.text.MessageFormat format;
private static final char SPACE;
private static final java.util.regex.Pattern ERROR_MESSAGE_PATTERN;
private static final java.util.regex.Pattern ERROR_CODE_PATTERN;
private static java.util.Map mesgToErrorMsgMap;
private static java.util.Map formatToErrorMsgMap;
private static int minMesgLength;
private static final org.apache.hadoop.hive.ql.ErrorMsg[] $VALUES;
public static org.apache.hadoop.hive.ql.ErrorMsg[] values()
{
java.lang.Object v;
org.apache.hadoop.hive.ql.ErrorMsg[] v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg[] $VALUES>;
v = virtualinvoke v.<java.lang.Object: java.lang.Object clone()>();
return v;
}
public static org.apache.hadoop.hive.ql.ErrorMsg valueOf(java.lang.String)
{
java.lang.String v;
java.lang.Enum v;
v := @parameter: java.lang.String;
v = staticinvoke <java.lang.Enum: java.lang.Enum valueOf(java.lang.Class,java.lang.String)>(class "Lorg/apache/hadoop/hive/ql/ErrorMsg;", v);
return v;
}
public static org.apache.hadoop.hive.ql.ErrorMsg getErrorMsg(java.lang.Exception)
{
java.lang.Exception v;
org.apache.hadoop.hive.ql.ErrorMsg v, v, v, v, v, v;
boolean v, v, v, v, v;
v := @parameter: java.lang.Exception;
v = v instanceof org.apache.hadoop.security.AccessControlException;
if v == 0 goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACCESS_DENIED>;
return v;
label:
v = v instanceof org.apache.hadoop.hdfs.protocol.NSQuotaExceededException;
if v == 0 goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg QUOTA_EXCEEDED>;
return v;
label:
v = v instanceof org.apache.hadoop.hdfs.protocol.DSQuotaExceededException;
if v == 0 goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg QUOTA_EXCEEDED>;
return v;
label:
v = v instanceof org.apache.hadoop.hdfs.protocol.UnresolvedPathException;
if v == 0 goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_PATH>;
return v;
label:
v = v instanceof java.io.FileNotFoundException;
if v == 0 goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg FILE_NOT_FOUND>;
return v;
label:
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_RT_EXCEPTION>;
return v;
}
public static org.apache.hadoop.hive.ql.ErrorMsg getErrorMsg(java.lang.String)
{
java.util.regex.Matcher v, v;
java.util.Map v, v, v;
int v, v, v, v;
java.lang.String v, v, v, v;
boolean v, v, v;
java.util.Iterator v;
java.util.Set v;
java.lang.Object v, v, v, v, v;
java.util.regex.Pattern v;
org.apache.hadoop.hive.ql.ErrorMsg v, v;
v := @parameter: java.lang.String;
if v != null goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg GENERIC_ERROR>;
return v;
label:
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map mesgToErrorMsgMap>;
v = interfaceinvoke v.<java.util.Map: java.lang.Object get(java.lang.Object)>(v);
if v == null goto label;
return v;
label:
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map formatToErrorMsgMap>;
v = interfaceinvoke v.<java.util.Map: java.util.Set entrySet()>();
v = interfaceinvoke v.<java.util.Set: java.util.Iterator iterator()>();
label:
v = interfaceinvoke v.<java.util.Iterator: boolean hasNext()>();
if v == 0 goto label;
v = interfaceinvoke v.<java.util.Iterator: java.lang.Object next()>();
v = interfaceinvoke v.<java.util.Map$Entry: java.lang.Object getKey()>();
v = virtualinvoke v.<java.util.regex.Pattern: java.util.regex.Matcher matcher(java.lang.CharSequence)>(v);
v = virtualinvoke v.<java.util.regex.Matcher: boolean matches()>();
if v == 0 goto label;
v = interfaceinvoke v.<java.util.Map$Entry: java.lang.Object getValue()>();
return v;
label:
v = virtualinvoke v.<java.lang.String: java.lang.String trim()>();
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.regex.Pattern ERROR_MESSAGE_PATTERN>;
v = virtualinvoke v.<java.util.regex.Pattern: java.util.regex.Matcher matcher(java.lang.CharSequence)>(v);
v = virtualinvoke v.<java.util.regex.Matcher: boolean matches()>();
if v == 0 goto label;
v = virtualinvoke v.<java.util.regex.Matcher: java.lang.String group(int)>(1);
label:
v = virtualinvoke v.<java.lang.String: int length()>();
v = <org.apache.hadoop.hive.ql.ErrorMsg: int minMesgLength>;
if v <= v goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map mesgToErrorMsgMap>;
v = virtualinvoke v.<java.lang.String: java.lang.String trim()>();
v = interfaceinvoke v.<java.util.Map: java.lang.Object get(java.lang.Object)>(v);
if v == null goto label;
return v;
label:
v = virtualinvoke v.<java.lang.String: int lastIndexOf(int)>(32);
v = (int) -1;
if v == v goto label;
v = virtualinvoke v.<java.lang.String: java.lang.String substring(int,int)>(0, v);
v = virtualinvoke v.<java.lang.String: java.lang.String trim()>();
goto label;
label:
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg GENERIC_ERROR>;
return v;
}
public static org.apache.hadoop.hive.ql.ErrorMsg getErrorMsg(int)
{
int v, v, v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
org.apache.hadoop.hive.ql.ErrorMsg[] v;
v := @parameter: int;
v = staticinvoke <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg[] values()>();
v = lengthof v;
v = 0;
label:
if v >= v goto label;
v = v[v];
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: int getErrorCode()>();
if v != v goto label;
return v;
label:
v = v + 1;
goto label;
label:
return null;
}
public static java.lang.String findSQLState(java.lang.String)
{
java.lang.String v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @parameter: java.lang.String;
v = staticinvoke <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg getErrorMsg(java.lang.String)>(v);
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String getSQLState()>();
return v;
}
private void <init>(java.lang.String, int, int, java.lang.String)
{
int v, v;
java.lang.String v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String;
v := @parameter: int;
v := @parameter: int;
v := @parameter: java.lang.String;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>(v, v, v, v, "42000", 0);
return;
}
private void <init>(java.lang.String, int, int, java.lang.String, boolean)
{
int v, v;
java.lang.String v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
boolean v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String;
v := @parameter: int;
v := @parameter: int;
v := @parameter: java.lang.String;
v := @parameter: boolean;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>(v, v, v, v, "42000", v);
return;
}
private void <init>(java.lang.String, int, int, java.lang.String, java.lang.String)
{
int v, v;
java.lang.String v, v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String;
v := @parameter: int;
v := @parameter: int;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>(v, v, v, v, v, 0);
return;
}
private void <init>(java.lang.String, int, int, java.lang.String, java.lang.String, boolean)
{
java.text.MessageFormat v;
int v, v;
java.lang.String v, v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
boolean v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String;
v := @parameter: int;
v := @parameter: int;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v := @parameter: boolean;
specialinvoke v.<java.lang.Enum: void <init>(java.lang.String,int)>(v, v);
v.<org.apache.hadoop.hive.ql.ErrorMsg: int errorCode> = v;
v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String mesg> = v;
v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String sqlState> = v;
if v == 0 goto label;
v = new java.text.MessageFormat;
specialinvoke v.<java.text.MessageFormat: void <init>(java.lang.String)>(v);
goto label;
label:
v = null;
label:
v.<org.apache.hadoop.hive.ql.ErrorMsg: java.text.MessageFormat format> = v;
return;
}
public java.lang.String getMsg(java.lang.String)
{
java.lang.String v, v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String mesg>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (java.lang.String,java.lang.String)>(v, v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("\u \u0001");
return v;
}
public java.lang.String format(java.lang.String)
{
java.lang.String[] v;
java.lang.String v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String;
v = newarray (java.lang.String)[1];
v[0] = v;
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String format(java.lang.String[])>(v);
return v;
}
public transient java.lang.String format(java.lang.String[])
{
java.lang.String[] v;
java.text.MessageFormat v, v;
int v, v, v, v, v;
java.lang.String v, v, v, v, v, v, v;
java.lang.StringBuilder v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String[];
if v == null goto label;
v = lengthof v;
if v != 0 goto label;
label:
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String getMsg()>();
return v;
label:
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.text.MessageFormat format>;
if v == null goto label;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.text.MessageFormat format>;
v = virtualinvoke v.<java.text.MessageFormat: java.lang.String format(java.lang.Object)>(v);
return v;
label:
v = lengthof v;
if v <= 1 goto label;
v = new java.lang.StringBuilder;
specialinvoke v.<java.lang.StringBuilder: void <init>()>();
v = lengthof v;
v = 0;
label:
if v >= v goto label;
v = v[v];
if v == null goto label;
v = virtualinvoke v.<java.lang.StringBuilder: int length()>();
if v <= 0 goto label;
virtualinvoke v.<java.lang.StringBuilder: java.lang.StringBuilder append(java.lang.String)>(" ");
label:
virtualinvoke v.<java.lang.StringBuilder: java.lang.StringBuilder append(java.lang.String)>(v);
label:
v = v + 1;
goto label;
label:
v = virtualinvoke v.<java.lang.StringBuilder: java.lang.String toString()>();
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String getMsg(java.lang.String)>(v);
return v;
label:
v = v[0];
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String getMsg(java.lang.String)>(v);
return v;
}
public java.lang.String getErrorCodedMsg()
{
int v;
java.lang.String v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: int errorCode>;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String mesg>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int,java.lang.String)>(v, v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("[Error \u]: \u0001");
return v;
}
public transient java.lang.String getErrorCodedMsg(java.lang.String[])
{
int v;
java.lang.String[] v;
java.lang.String v, v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v := @parameter: java.lang.String[];
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: int errorCode>;
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String format(java.lang.String[])>(v);
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int,java.lang.String)>(v, v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("[Error \u]: \u0001");
return v;
}
public static java.util.regex.Pattern getErrorCodePattern()
{
java.util.regex.Pattern v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.regex.Pattern ERROR_CODE_PATTERN>;
return v;
}
public java.lang.String getMsg()
{
java.lang.String v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String mesg>;
return v;
}
public java.lang.String getSQLState()
{
java.lang.String v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String sqlState>;
return v;
}
public int getErrorCode()
{
int v;
org.apache.hadoop.hive.ql.ErrorMsg v;
v := @this: org.apache.hadoop.hive.ql.ErrorMsg;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: int errorCode>;
return v;
}
private static org.apache.hadoop.hive.ql.ErrorMsg[] $values()
{
org.apache.hadoop.hive.ql.ErrorMsg[] v;
org.apache.hadoop.hive.ql.ErrorMsg v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
v = newarray (org.apache.hadoop.hive.ql.ErrorMsg)[349];
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg GENERIC_ERROR>;
v[0] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE>;
v[1] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_COLUMN>;
v[2] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_OR_COLUMN>;
v[3] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_TABLE_OR_COLUMN>;
v[4] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_PARTITION>;
v[5] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_COLUMN>;
v[6] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_TABLE_ALIAS>;
v[7] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_ALIAS>;
v[8] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_TABLE_ALIAS>;
v[9] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FUNCTION>;
v[10] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FUNCTION_SIGNATURE>;
v[11] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_OPERATOR_SIGNATURE>;
v[12] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT>;
v[13] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT_LENGTH>;
v[14] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT_TYPE>;
v[15] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_1>;
v[16] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_2>;
v[17] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_3>;
v[18] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TRANSFORM>;
v[19] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_MULTIPLE_DISTINCTS>;
v[20] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_SUBQUERY_ALIAS>;
v[21] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_INSERT_INSUBQUERY>;
v[22] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_KEY_EXPR_IN_GROUPBY>;
v[23] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_XPATH>;
v[24] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_PATH>;
v[25] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ILLEGAL_PATH>;
v[26] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_NUMERICAL_CONSTANT>;
v[27] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARRAYINDEX_TYPE>;
v[28] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPINDEX_CONSTANT>;
v[29] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPINDEX_TYPE>;
v[30] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_COLLECTION_TYPE>;
v[31] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SELECT_DISTINCT_WITH_GROUPBY>;
v[32] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMN_REPEATED_IN_PARTITIONING_COLS>;
v[33] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DUPLICATE_COLUMN_NAMES>;
v[34] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_BUCKET_NUMBER>;
v[35] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMN_REPEATED_IN_CLUSTER_SORT>;
v[36] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SAMPLE_RESTRICTION>;
v[37] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SAMPLE_COLUMN_NOT_FOUND>;
v[38] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_PARTITION_PREDICATE>;
v[39] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_DOT>;
v[40] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TBL_DDL_SERDE>;
v[41] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TARGET_TABLE_COLUMN_MISMATCH>;
v[42] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_ALIAS_NOT_ALLOWED>;
v[43] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_DISTRIBUTEBY_CONFLICT>;
v[44] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ORDERBY_DISTRIBUTEBY_CONFLICT>;
v[45] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_SORTBY_CONFLICT>;
v[46] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ORDERBY_SORTBY_CONFLICT>;
v[47] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_ORDERBY_CONFLICT>;
v[48] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_LIMIT_WITH_ORDERBY>;
v[49] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNION_NOTIN_SUBQ>;
v[50] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_INPUT_FORMAT_TYPE>;
v[51] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_OUTPUT_FORMAT_TYPE>;
v[52] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_VALID_PARTN>;
v[53] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_OUTER_MAPJOIN>;
v[54] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPJOIN_HINT>;
v[55] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPJOIN_TABLE>;
v[56] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_BUCKETED_TABLE>;
v[57] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BUCKETED_NUMERATOR_BIGGER_DENOMINATOR>;
v[58] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NEED_PARTITION_ERROR>;
v[59] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_CTLT_COEXISTENCE>;
v[60] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LINES_TERMINATED_BY_NON_NEWLINE>;
v[61] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_COLLST_COEXISTENCE>;
v[62] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLT_COLLST_COEXISTENCE>;
v[63] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_SELECT_SCHEMA>;
v[64] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_PARCOL_COEXISTENCE>;
v[65] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_MULTI_LOADFILE>;
v[66] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_EXTTBL_COEXISTENCE>;
v[67] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_EXTERNAL_TABLE>;
v[68] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATABASE_NOT_EXISTS>;
v[69] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_ALREADY_EXISTS>;
v[70] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMN_ALIAS_ALREADY_EXISTS>;
v[71] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_MULTIPLE_EXPR>;
v[72] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_REQUIRE_AS>;
v[73] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_GROUP_BY>;
v[74] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_SORT_BY>;
v[75] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_CLUSTER_BY>;
v[76] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_DISTRIBUTE_BY>;
v[77] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_INVALID_LOCATION>;
v[78] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_LATERAL_VIEW>;
v[79] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_ALIAS_MISMATCH>;
v[80] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDF_STATEFUL_INVALID_LOCATION>;
v[81] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LATERAL_VIEW_WITH_JOIN>;
v[82] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LATERAL_VIEW_INVALID_CHILD>;
v[83] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OUTPUT_SPECIFIED_MULTIPLE_TIMES>;
v[84] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_AS>;
v[85] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VIEW_COL_MISMATCH>;
v[86] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DML_AGAINST_VIEW>;
v[87] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ANALYZE_VIEW>;
v[88] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VIEW_PARTITION_TOTAL>;
v[89] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VIEW_PARTITION_MISMATCH>;
v[90] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_DYN_STA_ORDER>;
v[91] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITION_DISABLED>;
v[92] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITION_STRICT_MODE>;
v[93] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NONEXISTPARTCOL>;
v[94] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_TYPE>;
v[95] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CREATE_NON_NATIVE_AS>;
v[96] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_INTO_NON_NATIVE>;
v[97] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCKMGR_NOT_SPECIFIED>;
v[98] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCKMGR_NOT_INITIALIZED>;
v[99] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_CANNOT_BE_ACQUIRED>;
v[100] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ZOOKEEPER_CLIENT_COULD_NOT_BE_INITIALIZED>;
v[101] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OVERWRITE_ARCHIVED_PART>;
v[102] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_METHODS_DISABLED>;
v[103] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_ON_MULI_PARTS>;
v[104] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNARCHIVE_ON_MULI_PARTS>;
v[105] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_ON_TABLE>;
v[106] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RESERVED_PART_VAL>;
v[107] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OFFLINE_TABLE_OR_PARTITION>;
v[108] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NEED_PARTITION_SPECIFICATION>;
v[109] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_METADATA>;
v[110] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NEED_TABLE_SPECIFICATION>;
v[111] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_EXISTS>;
v[112] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_DATA_EXISTS>;
v[113] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INCOMPATIBLE_SCHEMA>;
v[114] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg EXIM_FOR_NON_NATIVE>;
v[115] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_BUCKETIZED_TABLE>;
v[116] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTSPEC_DIFFER_FROM_SCHEMA>;
v[117] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_COLUMN_NON_PRIMITIVE>;
v[118] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_DYNAMICPARTITION_IFNOTEXISTS>;
v[119] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDAF_INVALID_LOCATION>;
v[120] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DROP_PARTITION_NON_STRING_PARTCOLS_NONEQUALITY>;
v[121] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_COMMAND_FOR_VIEWS>;
v[122] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_COMMAND_FOR_TABLES>;
v[123] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_VIEW_DISALLOWED_OP>;
v[124] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NON_NATIVE>;
v[125] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SORTMERGE_MAPJOIN_FAILED>;
v[126] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BUCKET_MAPJOIN_NOT_POSSIBLE>;
v[127] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BUCKETED_TABLE_METADATA_INCORRECT>;
v[128] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg JOINNODE_OUTERJOIN_MORETHAN_16>;
v[129] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JDO_FILTER_EXPRESSION>;
v[130] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_BUCKETNUM_NONBUCKETIZED_TBL>;
v[131] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_FOR_NON_MANAGED_TABLE>;
v[132] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_FOR_NON_NATIVE_TABLE>;
v[133] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTSPEC_FOR_NON_PARTITIONED_TABLE>;
v[134] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_IN_ON_CLAUSE_OF_MERGE>;
v[135] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_INTO_STORED_AS_DIR>;
v[136] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_STOREDASDIR_NOT_SKEWED>;
v[137] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_SKEWED_LOC_NO_LOC>;
v[138] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_SKEWED_LOC_NO_MAP>;
v[139] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_NO_COLUMN_NAME>;
v[140] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_NO_COLUMN_VALUE>;
v[141] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_DUPLICATE_COLUMN_NAMES>;
v[142] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_INVALID_COLUMN>;
v[143] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_1>;
v[144] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_2>;
v[145] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_3>;
v[146] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NOT_ALLOWED_RENAME_SKEWED_COLUMN>;
v[147] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_NOMAPAGGR>;
v[148] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_EXPRESSION_INVALID>;
v[149] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_EMPTY>;
v[150] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_UNION_REMOVE_OPTIMIZATION_NEEDS_SUBDIRECTORIES>;
v[151] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_EXPR_NOT_IN_GROUPBY>;
v[152] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_PARTITION_SPEC>;
v[153] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_UNSET_NON_EXIST_PROPERTY>;
v[154] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_VIEW_AS_SELECT_NOT_EXIST>;
v[155] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_VIEW_WITH_PARTITION>;
v[156] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg EXISTING_TABLE_IS_NOT_VIEW>;
v[157] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_SUPPORTED_ORDERBY_ALLCOLREF_POS>;
v[158] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_POSITION_ALIAS_IN_GROUPBY>;
v[159] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_POSITION_ALIAS_IN_ORDERBY>;
v[160] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_SKEW>;
v[161] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_DISTINCTS>;
v[162] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OPERATOR_NOT_ALLOWED_WITH_MAPJOIN>;
v[163] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ANALYZE_TABLE_NOSCAN_NON_NATIVE>;
v[164] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_VALUE_NOT_CONTINUOUS>;
v[165] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLES_INCOMPATIBLE_SCHEMAS>;
v[166] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg EXCHANGE_PARTITION_NOT_ALLOWED_WITH_TRANSACTIONAL_TABLES>;
v[167] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_COLUMN_NOT_RC>;
v[168] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_COLUMN_ARCHIVED>;
v[169] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_BUCKETED_COLUMN>;
v[170] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_LIST_BUCKETED_COLUMN>;
v[171] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_NOT_PARTITIONED>;
v[172] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATABASE_ALREADY_EXISTS>;
v[173] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_REPLACE_COLUMNS>;
v[174] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BAD_LOCATION_VALUE>;
v[175] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_ALTER_TBL_OP>;
v[176] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_BIGTABLE_MAPJOIN>;
v[177] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MISSING_OVER_CLAUSE>;
v[178] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_SPEC_TYPE_MISMATCH>;
v[179] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_SUBQUERY_EXPRESSION>;
v[180] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_SUBQUERY_EXPRESSION>;
v[181] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_HDFS_URI>;
v[182] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_DIR>;
v[183] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_VALID_LOCATIONS>;
v[184] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_PRINCIPAL_TYPE_GROUP>;
v[185] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_NAME>;
v[186] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_IMMUTABLE_TABLE>;
v[187] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_GLOBAL>;
v[188] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_COLUMN>;
v[189] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_SPECIFIED>;
v[190] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_INSTANTIATED>;
v[191] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXN_NO_SUCH_TRANSACTION>;
v[192] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXN_ABORTED>;
v[193] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DBTXNMGR_REQUIRES_CONCURRENCY>;
v[194] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_ACID>;
v[195] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_NO_SUCH_LOCK>;
v[196] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_REQUEST_UNSUPPORTED>;
v[197] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg METASTORE_COMMUNICATION_FAILED>;
v[198] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg METASTORE_COULD_NOT_INITIATE>;
v[199] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_COMPACTION_TYPE>;
v[200] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_COMPACTION_PARTITION>;
v[201] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TOO_MANY_COMPACTION_PARTITIONS>;
v[202] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DISTINCT_NOT_SUPPORTED>;
v[203] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NONACID_COMPACTION_NOT_SUPPORTED>;
v[204] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MASKING_FILTERING_ON_ACID_NOT_SUPPORTED>;
v[205] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MASKING_FILTERING_ON_MATERIALIZED_VIEWS_SOURCES>;
v[206] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MASKING_COMPLEX_TYPE_NOT_SUPPORTED>;
v[207] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATEDELETE_PARSE_ERROR>;
v[208] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATE_CANNOT_UPDATE_PART_VALUE>;
v[209] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_CANNOT_CREATE_TEMP_FILE>;
v[210] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_NONACID_TXNMGR>;
v[211] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VALUES_TABLE_CONSTRUCTOR_NOT_SUPPORTED>;
v[212] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_NONACID_TABLE>;
v[213] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_NO_SORTED_BUCKETS>;
v[214] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_TYPE_PARTIAL_PARTITION_SPEC_NO_SUPPORTED>;
v[215] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_PARTITION_CASCADE_NOT_SUPPORTED>;
v[216] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DROP_NATIVE_FUNCTION>;
v[217] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATE_CANNOT_UPDATE_BUCKET_VALUE>;
v[218] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg IMPORT_INTO_STRICT_REPL_TABLE>;
v[219] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_LOCATION_NONEMPTY>;
v[220] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_CREATES_VOID_TYPE>;
v[221] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TBL_SORTED_NOT_BUCKETED>;
v[222] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_ACQUIRE_TIMEDOUT>;
v[223] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COMPILE_LOCK_TIMED_OUT>;
v[224] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_SERDE>;
v[225] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_FILEFORMAT>;
v[226] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_REORDER_COLUMNS>;
v[227] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_COLUMN_TYPE>;
v[228] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_CANNOT_DROP_COLUMNS>;
v[229] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_UNSUPPORTED_TYPE_CONVERSION>;
v[230] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_NOMAPAGGR_MULTIGBY>;
v[231] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_RETRIEVE_TABLE_METADATA>;
v[232] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_AST_TREE>;
v[233] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ERROR_SERIALIZE_METASTORE>;
v[234] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg IO_ERROR>;
v[235] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ERROR_SERIALIZE_METADATA>;
v[236] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_LOAD_TABLE_FILE_WORK>;
v[237] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLASSPATH_ERROR>;
v[238] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg IMPORT_SEMANTIC_ERROR>;
v[239] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FK_SYNTAX>;
v[240] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_CSTR_SYNTAX>;
v[241] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_NOT_ENOUGH_HISTORY>;
v[242] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_COLUMN_NAME>;
v[243] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_SET_OPERATOR>;
v[244] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_ACQUIRE_CANCELLED>;
v[245] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NOT_RECOGNIZED_CONSTRAINT>;
v[246] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_CONSTRAINT>;
v[247] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_VIEW_WITH_MATERIALIZED>;
v[248] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_MATERIALIZED_WITH_VIEW>;
v[249] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATE_DELETE_VIEW>;
v[250] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MATERIALIZED_VIEW_DEF_EMPTY>;
v[251] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MERGE_PREDIACTE_REQUIRED>;
v[252] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MERGE_TOO_MANY_DELETE>;
v[253] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MERGE_TOO_MANY_UPDATE>;
v[254] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION>;
v[255] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TARGET_COLUMN_IN_SET_CLAUSE>;
v[256] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_FUNCTION_EXPR_NOT_IN_GROUPBY>;
v[257] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NON_PARTITIONED_TABLE_CASCADE_NOT_SUPPORTED>;
v[258] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_SIZE_LIMIT>;
v[259] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REBUILD_NO_MATERIALIZED_VIEW>;
v[260] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_DATA_ACID_FILE>;
v[261] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_INSERTONLYTRAN_TABLE>;
v[262] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_DATA_LAUNCH_JOB_PARSE_ERROR>;
v[263] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RESOURCE_PLAN_ALREADY_EXISTS>;
v[264] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RESOURCE_PLAN_NOT_EXISTS>;
v[265] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INCOMPATIBLE_STRUCT>;
v[266] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OBJECTNAME_CONTAINS_DOT>;
v[267] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WITHIN_GROUP_NOT_ALLOWED>;
v[268] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WITHIN_GROUP_PARAMETER_MISMATCH>;
v[269] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_STRUCT_ATTRIBUTE>;
v[270] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OFFSET_NOT_SUPPORTED_IN_SUBQUERY>;
v[271] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WITH_COL_LIST_NUM_OVERFLOW>;
v[272] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NULL_TREATMENT_NOT_SUPPORTED>;
v[273] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATACONNECTOR_ALREADY_EXISTS>;
v[274] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATACONNECTOR_NOT_EXISTS>;
v[275] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TIME_TRAVEL_NOT_ALLOWED>;
v[276] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_METADATA_TABLE_NAME>;
v[277] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_META_REF_NOT_SUPPORTED>;
v[278] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COMPACTION_REFUSED>;
v[279] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CBO_IS_REQUIRED>;
v[280] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_UNSUPPORTED_FORMAT>;
v[281] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_NATIVE_ACID_UPDATE>;
v[282] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg READ_ONLY_DATABASE>;
v[283] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_INIT_ERROR>;
v[284] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_IO_ERROR>;
v[285] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_GENERIC_ERROR>;
v[286] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_CLOSING_ERROR>;
v[287] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITIONS_TOO_MANY_PER_NODE_ERROR>;
v[288] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_IN_IMPLICIT_TXN>;
v[289] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_IN_TXN>;
v[290] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_WITHOUT_TXN>;
v[291] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACCESS_DENIED>;
v[292] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg QUOTA_EXCEEDED>;
v[293] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_PATH>;
v[294] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg FILE_NOT_FOUND>;
v[295] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WRONG_FILE_FORMAT>;
v[296] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FILE_MISSING_FROM_SRC_AND_CM_PATH>;
v[297] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_EXTERNAL_SERVICE_CONNECTION_ERROR>;
v[298] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLIENT_POLLING_OPSTATUS_INTERRUPTED>;
v[299] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_FAILED_INFERENCE>;
v[300] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_CLASS_NOT_FOUND>;
v[301] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_MISSING_STORAGE_FORMAT_DESCRIPTOR>;
v[302] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARQUET_FOOTER_ERROR>;
v[303] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARQUET_UNHANDLED_TYPE>;
v[304] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ORC_FOOTER_ERROR>;
v[305] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_RT_EXCEPTION>;
v[306] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_NOT_OBTAINED>;
v[307] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_INITIALIZATION_ERROR>;
v[308] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_CONNECTION_ERROR>;
v[309] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_PUBLISHING_ERROR>;
v[310] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_CLOSING_ERROR>;
v[311] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_PART_KEY>;
v[312] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_PARTITION>;
v[313] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_PARSE_ERROR>;
v[314] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DROP_COMMAND_NOT_ALLOWED_FOR_PARTITION>;
v[315] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_COLUMN>;
v[316] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSAGGREGATOR_SOURCETASK_NULL>;
v[317] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSAGGREGATOR_CONNECTION_ERROR>;
v[318] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATS_SKIPPING_BY_ERROR>;
v[319] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FILE_FORMAT_IN_LOAD>;
v[320] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCHEMA_REQUIRED_TO_READ_ACID_TABLES>;
v[321] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_TABLES_MUST_BE_READ_WITH_ACID_READER>;
v[322] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_TABLES_MUST_BE_READ_WITH_HIVEINPUTFORMAT>;
v[323] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_LOAD_DATA_INVALID_FILE_NAME>;
v[324] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_FILE_FORMAT>;
v[325] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_BUCKETED>;
v[326] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_PARTITION_ARCHIVED>;
v[327] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_NON_NATIVE>;
v[328] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_NOT_MANAGED>;
v[329] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_TRANSACTIONAL>;
v[330] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FILE_SYSTEM_OPERATION_RETRY>;
v[331] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_SOURCE_DATABASE_NOT_FOUND>;
v[332] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_DATABASE_IS_TARGET_OF_REPLICATION>;
v[333] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_DB_OR_TABLE_PATTERN>;
v[334] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_EVENTS_MISSING_IN_METASTORE>;
v[335] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_BOOTSTRAP_LOAD_PATH_NOT_VALID>;
v[336] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_CONFIG_FOR_SERVICE>;
v[337] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_INTERNAL_CONFIG_FOR_SERVICE>;
v[338] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_RETRY_EXHAUSTED>;
v[339] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FAILED_WITH_NON_RECOVERABLE_ERROR>;
v[340] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_ARGUMENTS>;
v[341] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_ALTER_TABLE>;
v[342] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_PERMISSION_DENIED>;
v[343] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_DISTCP_SNAPSHOT_EXCEPTION>;
v[344] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RANGER_AUTHORIZATION_FAILED>;
v[345] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RANGER_AUTHENTICATION_FAILED>;
v[346] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INCOMPATIBLE_EXCEPTION>;
v[347] = v;
v = <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FAILOVER_TARGET_MODIFIED>;
v[348] = v;
return v;
}
static void <clinit>()
{
java.text.MessageFormat v;
int v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
java.lang.String v, v, v, v, v, v, v, v, v, v, v, v;
java.util.regex.Pattern v, v, v;
org.apache.hadoop.hive.conf.HiveConf$ConfVars v, v;
java.util.Map v, v;
org.apache.hadoop.hive.ql.ErrorMsg[] v, v;
org.apache.hadoop.hive.ql.ErrorMsg v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
java.util.HashMap v, v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40000;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("GENERIC_ERROR", 0, v, "Exception while processing");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg GENERIC_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("INVALID_TABLE", 1, 10001, "Table not found", "42S02");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_COLUMN", 2, 10002, "Invalid column reference");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_TABLE_OR_COLUMN", 3, 10004, "Invalid table alias or column reference");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_OR_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("AMBIGUOUS_TABLE_OR_COLUMN", 4, 10005, "Ambiguous table alias or column reference");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_TABLE_OR_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_PARTITION", 5, 10006, "Partition not found");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_PARTITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("AMBIGUOUS_COLUMN", 6, 10007, "Ambiguous column reference");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("AMBIGUOUS_TABLE_ALIAS", 7, 10008, "Ambiguous table alias");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_TABLE_ALIAS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_TABLE_ALIAS", 8, 10009, "Invalid table alias");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_ALIAS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_TABLE_ALIAS", 9, 10010, "No table alias");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_TABLE_ALIAS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_FUNCTION", 10, 10011, "Invalid function");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FUNCTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_FUNCTION_SIGNATURE", 11, 10012, "Function argument type mismatch");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FUNCTION_SIGNATURE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_OPERATOR_SIGNATURE", 12, 10013, "Operator argument type mismatch");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_OPERATOR_SIGNATURE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_ARGUMENT", 13, 10014, "Wrong arguments");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("INVALID_ARGUMENT_LENGTH", 14, 10015, "Arguments length mismatch", "21000");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT_LENGTH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_ARGUMENT_TYPE", 15, 10016, "Argument type mismatch");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARGUMENT_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_JOIN_CONDITION_1", 16, 10017, "Both left and right aliases encountered in JOIN");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_1> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_JOIN_CONDITION_2", 17, 10018, "Neither left nor right aliases encountered in JOIN");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_2> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_JOIN_CONDITION_3", 18, 10019, "OR not supported in JOIN currently");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION_3> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_TRANSFORM", 19, 10020, "TRANSFORM with other SELECT columns not supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TRANSFORM> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UNSUPPORTED_MULTIPLE_DISTINCTS", 20, 10022, "DISTINCT on different columns not supported with skew in data");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_MULTIPLE_DISTINCTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_SUBQUERY_ALIAS", 21, 10023, "No alias for subquery");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_SUBQUERY_ALIAS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_INSERT_INSUBQUERY", 22, 10024, "Cannot insert in a subquery. Inserting to table ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_INSERT_INSUBQUERY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NON_KEY_EXPR_IN_GROUPBY", 23, 10025, "Expression not in GROUP BY key");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_KEY_EXPR_IN_GROUPBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_XPATH", 24, 10026, "General . and [] operators are not supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_XPATH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_PATH", 25, 10027, "Invalid path");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_PATH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ILLEGAL_PATH", 26, 10028, "Path is not legal");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ILLEGAL_PATH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_NUMERICAL_CONSTANT", 27, 10029, "Invalid numerical constant");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_NUMERICAL_CONSTANT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_ARRAYINDEX_TYPE", 28, 10030, "Not proper type for index of ARRAY. Currently, only integer type is supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_ARRAYINDEX_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_MAPINDEX_CONSTANT", 29, 10031, "Non-constant expression for map indexes not supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPINDEX_CONSTANT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_MAPINDEX_TYPE", 30, 10032, "MAP key type does not match index expression type");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPINDEX_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NON_COLLECTION_TYPE", 31, 10033, "[] not valid on non-collection types");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_COLLECTION_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SELECT_DISTINCT_WITH_GROUPBY", 32, 10034, "SELECT DISTINCT and GROUP BY can not be in the same query");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SELECT_DISTINCT_WITH_GROUPBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("COLUMN_REPEATED_IN_PARTITIONING_COLS", 33, 10035, "Column repeated in partitioning columns");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMN_REPEATED_IN_PARTITIONING_COLS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DUPLICATE_COLUMN_NAMES", 34, 10036, "Duplicate column name:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DUPLICATE_COLUMN_NAMES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_BUCKET_NUMBER", 35, 10037, "Bucket number should be bigger than zero");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_BUCKET_NUMBER> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("COLUMN_REPEATED_IN_CLUSTER_SORT", 36, 10038, "Same column cannot appear in CLUSTER BY and SORT BY");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMN_REPEATED_IN_CLUSTER_SORT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SAMPLE_RESTRICTION", 37, 10039, "Cannot SAMPLE on more than two columns");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SAMPLE_RESTRICTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SAMPLE_COLUMN_NOT_FOUND", 38, 10040, "SAMPLE column not found");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SAMPLE_COLUMN_NOT_FOUND> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_PARTITION_PREDICATE", 39, 10041, "No partition predicate found");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_PARTITION_PREDICATE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_DOT", 40, 10042, ". Operator is only supported on struct or list of struct types");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_DOT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_TBL_DDL_SERDE", 41, 10043, "Either list of columns or a custom serializer should be specified");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TBL_DDL_SERDE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TARGET_TABLE_COLUMN_MISMATCH", 42, 10044, "Cannot insert into target table because column number/types are different");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TARGET_TABLE_COLUMN_MISMATCH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TABLE_ALIAS_NOT_ALLOWED", 43, 10045, "Table alias not allowed in sampling clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_ALIAS_NOT_ALLOWED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CLUSTERBY_DISTRIBUTEBY_CONFLICT", 44, 10046, "Cannot have both CLUSTER BY and DISTRIBUTE BY clauses");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_DISTRIBUTEBY_CONFLICT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ORDERBY_DISTRIBUTEBY_CONFLICT", 45, 10047, "Cannot have both ORDER BY and DISTRIBUTE BY clauses");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ORDERBY_DISTRIBUTEBY_CONFLICT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CLUSTERBY_SORTBY_CONFLICT", 46, 10048, "Cannot have both CLUSTER BY and SORT BY clauses");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_SORTBY_CONFLICT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ORDERBY_SORTBY_CONFLICT", 47, 10049, "Cannot have both ORDER BY and SORT BY clauses");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ORDERBY_SORTBY_CONFLICT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CLUSTERBY_ORDERBY_CONFLICT", 48, 10050, "Cannot have both CLUSTER BY and ORDER BY clauses");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLUSTERBY_ORDERBY_CONFLICT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_LIMIT_WITH_ORDERBY", 49, 10051, "In strict mode, if ORDER BY is specified, LIMIT must also be specified");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_LIMIT_WITH_ORDERBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UNION_NOTIN_SUBQ", 50, 10053, "Top level UNION is not supported currently; use a subquery for the UNION");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNION_NOTIN_SUBQ> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_INPUT_FORMAT_TYPE", 51, 10054, "Input format must implement InputFormat");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_INPUT_FORMAT_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_OUTPUT_FORMAT_TYPE", 52, 10055, "Output Format must implement HiveOutputFormat, otherwise it should be either IgnoreKeyTextOutputFormat or SequenceFileOutputFormat");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_OUTPUT_FORMAT_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = <org.apache.hadoop.hive.conf.HiveConf$StrictChecks: java.lang.String NO_PARTITIONLESS_MSG>;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_VALID_PARTN", 53, 10056, v);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_VALID_PARTN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_OUTER_MAPJOIN", 54, 10057, "MAPJOIN cannot be performed with OUTER JOIN");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_OUTER_MAPJOIN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_MAPJOIN_HINT", 55, 10058, "All tables are specified as map-table for join");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPJOIN_HINT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_MAPJOIN_TABLE", 56, 10059, "Result of a union cannot be a map table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_MAPJOIN_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NON_BUCKETED_TABLE", 57, 10060, "Sampling expression needed for non-bucketed table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_BUCKETED_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("BUCKETED_NUMERATOR_BIGGER_DENOMINATOR", 58, 10061, "Numerator should not be bigger than denominator in sample clause for table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BUCKETED_NUMERATOR_BIGGER_DENOMINATOR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NEED_PARTITION_ERROR", 59, 10062, "Need to specify partition columns because the destination table is partitioned");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NEED_PARTITION_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_CTLT_COEXISTENCE", 60, 10063, "Create table command does not allow LIKE and AS-SELECT in the same command");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_CTLT_COEXISTENCE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LINES_TERMINATED_BY_NON_NEWLINE", 61, 10064, "LINES TERMINATED BY only supports newline \'\\n\' right now");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LINES_TERMINATED_BY_NON_NEWLINE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_COLLST_COEXISTENCE", 62, 10065, "CREATE TABLE AS SELECT command cannot specify the list of columns for the target table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_COLLST_COEXISTENCE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTLT_COLLST_COEXISTENCE", 63, 10066, "CREATE TABLE LIKE command cannot specify the list of columns for the target table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLT_COLLST_COEXISTENCE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_SELECT_SCHEMA", 64, 10067, "Cannot derive schema from the select-clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_SELECT_SCHEMA> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_PARCOL_COEXISTENCE", 65, 10068, "CREATE-TABLE-AS-SELECT does not support partitioning in the target table ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_PARCOL_COEXISTENCE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_MULTI_LOADFILE", 66, 10069, "CREATE-TABLE-AS-SELECT results in multiple file load");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_MULTI_LOADFILE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_EXTTBL_COEXISTENCE", 67, 10070, "CREATE-TABLE-AS-SELECT cannot create external table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_EXTTBL_COEXISTENCE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INSERT_EXTERNAL_TABLE", 68, 10071, "Inserting into a external table is not allowed");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_EXTERNAL_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DATABASE_NOT_EXISTS", 69, 10072, "Database does not exist:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATABASE_NOT_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("TABLE_ALREADY_EXISTS", 70, 10073, "Table already exists:", "42S02");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_ALREADY_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("COLUMN_ALIAS_ALREADY_EXISTS", 71, 10074, "Column alias already exists:", "42S02");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMN_ALIAS_ALREADY_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_MULTIPLE_EXPR", 72, 10075, "Only a single expression in the SELECT clause is supported with UDTF\'s");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_MULTIPLE_EXPR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_REQUIRE_AS", 73, 10076, "UDTF\'s require an AS clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_REQUIRE_AS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_NO_GROUP_BY", 74, 10077, "GROUP BY is not supported with a UDTF in the SELECT clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_GROUP_BY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_NO_SORT_BY", 75, 10078, "SORT BY is not supported with a UDTF in the SELECT clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_SORT_BY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_NO_CLUSTER_BY", 76, 10079, "CLUSTER BY is not supported with a UDTF in the SELECT clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_CLUSTER_BY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_NO_DISTRIBUTE_BY", 77, 10080, "DISTRIBUTE BY is not supported with a UDTF in the SELECT clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_NO_DISTRIBUTE_BY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_INVALID_LOCATION", 78, 10081, "UDTF\'s are not supported outside the SELECT clause, nor nested in expressions");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_INVALID_LOCATION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_LATERAL_VIEW", 79, 10082, "UDTF\'s cannot be in a select expression when there is a lateral view");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_LATERAL_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDTF_ALIAS_MISMATCH", 80, 10083, "The number of aliases supplied in the AS clause does not match the number of columns output by the UDTF");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDTF_ALIAS_MISMATCH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDF_STATEFUL_INVALID_LOCATION", 81, 10084, "Stateful UDF\'s can only be invoked in the SELECT list");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDF_STATEFUL_INVALID_LOCATION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LATERAL_VIEW_WITH_JOIN", 82, 10085, "JOIN with a LATERAL VIEW is not supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LATERAL_VIEW_WITH_JOIN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LATERAL_VIEW_INVALID_CHILD", 83, 10086, "LATERAL VIEW AST with invalid child");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LATERAL_VIEW_INVALID_CHILD> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("OUTPUT_SPECIFIED_MULTIPLE_TIMES", 84, 10087, "The same output cannot be present multiple times: ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OUTPUT_SPECIFIED_MULTIPLE_TIMES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_AS", 85, 10088, "AS clause has an invalid number of aliases");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_AS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("VIEW_COL_MISMATCH", 86, 10089, "The number of columns produced by the SELECT clause does not match the number of column names specified by CREATE VIEW");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VIEW_COL_MISMATCH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DML_AGAINST_VIEW", 87, 10090, "A view cannot be used as target table for LOAD or INSERT");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DML_AGAINST_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ANALYZE_VIEW", 88, 10091, "ANALYZE is not supported for views");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ANALYZE_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("VIEW_PARTITION_TOTAL", 89, 10092, "At least one non-partitioning column must be present in view");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VIEW_PARTITION_TOTAL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("VIEW_PARTITION_MISMATCH", 90, 10093, "Rightmost columns in view output do not match PARTITIONED ON clause");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VIEW_PARTITION_MISMATCH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("PARTITION_DYN_STA_ORDER", 91, 10094, "Dynamic partition cannot be the parent of a static partition");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_DYN_STA_ORDER> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DYNAMIC_PARTITION_DISABLED", 92, 10095, "Dynamic partition is disabled. Either enable it by setting hive.exec.dynamic.partition=true or specify partition column values");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITION_DISABLED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DYNAMIC_PARTITION_STRICT_MODE", 93, 10096, "Dynamic partition strict mode requires at least one static partition column. To turn this off set hive.exec.dynamic.partition.mode=nonstrict");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITION_STRICT_MODE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NONEXISTPARTCOL", 94, 10098, "Non-Partition column appears in the partition specification: ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NONEXISTPARTCOL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UNSUPPORTED_TYPE", 95, 10099, "DATETIME type isn\'t supported yet. Please use DATE or TIMESTAMP instead");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CREATE_NON_NATIVE_AS", 96, 10100, "CREATE TABLE AS SELECT cannot be used for a non-native table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CREATE_NON_NATIVE_AS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOAD_INTO_NON_NATIVE", 97, 10101, "A non-native table cannot be used as target for LOAD");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_INTO_NON_NATIVE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOCKMGR_NOT_SPECIFIED", 98, 10102, "Lock manager not specified correctly, set hive.lock.manager");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCKMGR_NOT_SPECIFIED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOCKMGR_NOT_INITIALIZED", 99, 10103, "Lock manager could not be initialized, check hive.lock.manager ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCKMGR_NOT_INITIALIZED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOCK_CANNOT_BE_ACQUIRED", 100, 10104, "Locks on the underlying objects cannot be acquired, retry after some time.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_CANNOT_BE_ACQUIRED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ZOOKEEPER_CLIENT_COULD_NOT_BE_INITIALIZED", 101, 10105, "Check hive.zookeeper.quorum and hive.zookeeper.client.port");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ZOOKEEPER_CLIENT_COULD_NOT_BE_INITIALIZED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("OVERWRITE_ARCHIVED_PART", 102, 10106, "Cannot overwrite an archived partition. Unarchive before running this command");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OVERWRITE_ARCHIVED_PART> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ARCHIVE_METHODS_DISABLED", 103, 10107, "Archiving methods are currently disabled. Please see the Hive wiki for more information about enabling archiving");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_METHODS_DISABLED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ARCHIVE_ON_MULI_PARTS", 104, 10108, "ARCHIVE can only be run on a single partition");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_ON_MULI_PARTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UNARCHIVE_ON_MULI_PARTS", 105, 10109, "ARCHIVE can only be run on a single partition");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNARCHIVE_ON_MULI_PARTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ARCHIVE_ON_TABLE", 106, 10110, "ARCHIVE can only be run on partitions");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ARCHIVE_ON_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("RESERVED_PART_VAL", 107, 10111, "Partition value contains a reserved substring");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RESERVED_PART_VAL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("OFFLINE_TABLE_OR_PARTITION", 108, 10113, "Query against an offline table or partition");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OFFLINE_TABLE_OR_PARTITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NEED_PARTITION_SPECIFICATION", 109, 10115, "Table is partitioned and partition specification is needed");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NEED_PARTITION_SPECIFICATION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_METADATA", 110, 10116, "The metadata file could not be parsed ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_METADATA> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NEED_TABLE_SPECIFICATION", 111, 10117, "Table name could be determined; It should be specified ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NEED_TABLE_SPECIFICATION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("PARTITION_EXISTS", 112, 10118, "Partition already exists");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TABLE_DATA_EXISTS", 113, 10119, "Table exists and contains data files");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_DATA_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INCOMPATIBLE_SCHEMA", 114, 10120, "The existing table is not compatible with the Export/Import spec. ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INCOMPATIBLE_SCHEMA> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("EXIM_FOR_NON_NATIVE", 115, 10121, "Export/Import cannot be done for a non-native table.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg EXIM_FOR_NON_NATIVE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INSERT_INTO_BUCKETIZED_TABLE", 116, 10122, "Bucketized tables do not support INSERT INTO:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_BUCKETIZED_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("PARTSPEC_DIFFER_FROM_SCHEMA", 117, 10125, "Partition columns in partition specification are not the same as that defined in the table schema. The names and orders have to be exactly the same.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTSPEC_DIFFER_FROM_SCHEMA> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("PARTITION_COLUMN_NON_PRIMITIVE", 118, 10126, "Partition column must be of primitive type.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_COLUMN_NON_PRIMITIVE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INSERT_INTO_DYNAMICPARTITION_IFNOTEXISTS", 119, 10127, "Dynamic partitions do not support IF NOT EXISTS. Specified partitions with value :");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_DYNAMICPARTITION_IFNOTEXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UDAF_INVALID_LOCATION", 120, 10128, "Not yet supported place for UDAF");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UDAF_INVALID_LOCATION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DROP_PARTITION_NON_STRING_PARTCOLS_NONEQUALITY", 121, 10129, "Drop partitions for a non-string partition column is only allowed using equality");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DROP_PARTITION_NON_STRING_PARTCOLS_NONEQUALITY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_COMMAND_FOR_VIEWS", 122, 10131, "To alter a view you need to use the ALTER VIEW command.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_COMMAND_FOR_VIEWS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_COMMAND_FOR_TABLES", 123, 10132, "To alter a base table you need to use the ALTER TABLE command.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_COMMAND_FOR_TABLES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_VIEW_DISALLOWED_OP", 124, 10133, "Cannot use this form of ALTER on a view");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_VIEW_DISALLOWED_OP> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ALTER_TABLE_NON_NATIVE", 125, 10134, "ALTER TABLE can only be used for {0} to a non-native table  {1}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NON_NATIVE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SORTMERGE_MAPJOIN_FAILED", 126, 10135, "Sort merge bucketed join could not be performed. If you really want to perform the operation, either set hive.optimize.bucketmapjoin.sortedmerge=false, or set hive.enforce.sortmergebucketmapjoin=false.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SORTMERGE_MAPJOIN_FAILED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("BUCKET_MAPJOIN_NOT_POSSIBLE", 127, 10136, "Bucketed mapjoin cannot be performed. This can be due to multiple reasons:  . Join columns don\'t match bucketed columns.  . Number of buckets are not a multiple of each other. If you really want to perform the operation, either remove the mapjoin hint from your query or set hive.enforce.bucketmapjoin to false.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BUCKET_MAPJOIN_NOT_POSSIBLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("BUCKETED_TABLE_METADATA_INCORRECT", 128, 10141, "Bucketed table metadata is not correct. Fix the metadata or don\'t use bucketed mapjoin, by setting hive.enforce.bucketmapjoin to false.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BUCKETED_TABLE_METADATA_INCORRECT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("JOINNODE_OUTERJOIN_MORETHAN_16", 129, 10142, "Single join node containing outer join(s) cannot have more than 16 aliases");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg JOINNODE_OUTERJOIN_MORETHAN_16> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_JDO_FILTER_EXPRESSION", 130, 10143, "Invalid expression for JDO filter");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JDO_FILTER_EXPRESSION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_BUCKETNUM_NONBUCKETIZED_TBL", 131, 10145, "Table is not bucketized.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_BUCKETNUM_NONBUCKETIZED_TBL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TRUNCATE_FOR_NON_MANAGED_TABLE", 132, 10146, "Cannot truncate non-managed table {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_FOR_NON_MANAGED_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TRUNCATE_FOR_NON_NATIVE_TABLE", 133, 10147, "Cannot truncate non-native table {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_FOR_NON_NATIVE_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("PARTSPEC_FOR_NON_PARTITIONED_TABLE", 134, 10148, "Partition spec for non partitioned table {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTSPEC_FOR_NON_PARTITIONED_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_TABLE_IN_ON_CLAUSE_OF_MERGE", 135, 10149, "No columns from target table \'\'{0}\'\' found in ON clause \'\'{1}\'\' of MERGE statement.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_IN_ON_CLAUSE_OF_MERGE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOAD_INTO_STORED_AS_DIR", 136, 10195, "A stored-as-directories table cannot be used as target for LOAD");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_INTO_STORED_AS_DIR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_TBL_STOREDASDIR_NOT_SKEWED", 137, 10196, "This operation is only valid on skewed table.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_STOREDASDIR_NOT_SKEWED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_TBL_SKEWED_LOC_NO_LOC", 138, 10197, "Alter table skewed location doesn\'t have locations.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_SKEWED_LOC_NO_LOC> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_TBL_SKEWED_LOC_NO_MAP", 139, 10198, "Alter table skewed location doesn\'t have location map.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_SKEWED_LOC_NO_MAP> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_NO_COLUMN_NAME", 140, 10200, "No skewed column name.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_NO_COLUMN_NAME> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_NO_COLUMN_VALUE", 141, 10201, "No skewed values.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_NO_COLUMN_VALUE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_DUPLICATE_COLUMN_NAMES", 142, 10202, "Duplicate skewed column name:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_DUPLICATE_COLUMN_NAMES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_INVALID_COLUMN", 143, 10203, "Invalid skewed column name:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_INVALID_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_1", 144, 10204, "Skewed column name is empty but skewed value is not.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_1> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_2", 145, 10205, "Skewed column value is empty but skewed name is not.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_2> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_3", 146, 10206, "The number of skewed column names and the number of skewed column values are different: ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_3> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_TABLE_NOT_ALLOWED_RENAME_SKEWED_COLUMN", 147, 10207, " is a skewed column. It\'s not allowed to rename skewed column or change skewed column type.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NOT_ALLOWED_RENAME_SKEWED_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_AGGR_NOMAPAGGR", 148, 10209, "Grouping sets aggregations (with rollups or cubes) are not allowed if map-side  aggregation is turned off. Set hive.map.aggr=true if you want to use grouping sets");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_NOMAPAGGR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_AGGR_EXPRESSION_INVALID", 149, 10210, "Grouping sets aggregations (with rollups or cubes) are not allowed if aggregation function parameters overlap with the aggregation functions columns");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_EXPRESSION_INVALID> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_EMPTY", 150, 10211, "Empty grouping sets not allowed");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_EMPTY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_UNION_REMOVE_OPTIMIZATION_NEEDS_SUBDIRECTORIES", 151, 10212, "In order to use hive.optimize.union.remove, the hadoop version that you are using should support sub-directories for tables/partitions. If that is true, set hive.hadoop.supports.subdirectories to true. Otherwise, set hive.optimize.union.remove to false");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_UNION_REMOVE_OPTIMIZATION_NEEDS_SUBDIRECTORIES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_EXPR_NOT_IN_GROUPBY", 152, 10213, "Grouping sets expression is not in GROUP BY key");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_EXPR_NOT_IN_GROUPBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_PARTITION_SPEC", 153, 10214, "Invalid partition spec specified");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_PARTITION_SPEC> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_TBL_UNSET_NON_EXIST_PROPERTY", 154, 10215, "Please use the following syntax if not sure whether the property existed or not:\nALTER TABLE tableName UNSET TBLPROPERTIES IF EXISTS (key, key, ...)\n");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TBL_UNSET_NON_EXIST_PROPERTY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_VIEW_AS_SELECT_NOT_EXIST", 155, 10216, "Cannot ALTER VIEW AS SELECT if view currently does not exist\n");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_VIEW_AS_SELECT_NOT_EXIST> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPLACE_VIEW_WITH_PARTITION", 156, 10217, "Cannot replace a view with CREATE VIEW or REPLACE VIEW or ALTER VIEW AS SELECT if the view has partitions\n");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_VIEW_WITH_PARTITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("EXISTING_TABLE_IS_NOT_VIEW", 157, 10218, "Existing table is not a view\n");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg EXISTING_TABLE_IS_NOT_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_SUPPORTED_ORDERBY_ALLCOLREF_POS", 158, 10219, "Position in ORDER BY is not supported when using SELECT *");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_SUPPORTED_ORDERBY_ALLCOLREF_POS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_POSITION_ALIAS_IN_GROUPBY", 159, 10220, "Invalid position alias in Group By\n");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_POSITION_ALIAS_IN_GROUPBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_POSITION_ALIAS_IN_ORDERBY", 160, 10221, "Invalid position alias in Order By\n");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_POSITION_ALIAS_IN_ORDERBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_SKEW", 161, 10225, "An additional MR job is introduced since the number of rows created per input row due to grouping sets is more than hive.new.job.grouping.set.cardinality. There is no need to handle skew separately. set hive.groupby.skewindata to false.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_SKEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_DISTINCTS", 162, 10226, "An additional MR job is introduced since the cardinality of grouping sets is more than hive.new.job.grouping.set.cardinality. This functionality is not supported with distincts. Either set hive.new.job.grouping.set.cardinality to a high number (higher than the number of rows per input row due to grouping sets in the query), or rewrite the query to not use distincts.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_DISTINCTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("OPERATOR_NOT_ALLOWED_WITH_MAPJOIN", 163, 10227, "Not all clauses are supported with mapjoin hint. Please remove mapjoin hint.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OPERATOR_NOT_ALLOWED_WITH_MAPJOIN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ANALYZE_TABLE_NOSCAN_NON_NATIVE", 164, 10228, "ANALYZE TABLE NOSCAN cannot be used for a non-native table");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ANALYZE_TABLE_NOSCAN_NON_NATIVE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("PARTITION_VALUE_NOT_CONTINUOUS", 165, 10234, "Partition values specified are not continuous. A subpartition value is specified without specifying the parent partition\'s value");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_VALUE_NOT_CONTINUOUS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TABLES_INCOMPATIBLE_SCHEMAS", 166, 10235, "Tables have incompatible schemas and their partitions  cannot be exchanged.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLES_INCOMPATIBLE_SCHEMAS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("EXCHANGE_PARTITION_NOT_ALLOWED_WITH_TRANSACTIONAL_TABLES", 167, 10236, "Exchange partition is not allowed with transactional tables. Alternatively, shall use load data or insert overwrite to move partitions.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg EXCHANGE_PARTITION_NOT_ALLOWED_WITH_TRANSACTIONAL_TABLES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TRUNCATE_COLUMN_NOT_RC", 168, 10237, "Only RCFileFormat supports column truncation.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_COLUMN_NOT_RC> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TRUNCATE_COLUMN_ARCHIVED", 169, 10238, "Column truncation cannot be performed on archived partitions.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_COLUMN_ARCHIVED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TRUNCATE_BUCKETED_COLUMN", 170, 10239, "A column on which a partition/table is bucketed cannot be truncated.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_BUCKETED_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TRUNCATE_LIST_BUCKETED_COLUMN", 171, 10240, "A column on which a partition/table is list bucketed cannot be truncated.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TRUNCATE_LIST_BUCKETED_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TABLE_NOT_PARTITIONED", 172, 10241, "Table {0} is not a partitioned table", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_NOT_PARTITIONED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("DATABASE_ALREADY_EXISTS", 173, 10242, "Database {0} already exists", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATABASE_ALREADY_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CANNOT_REPLACE_COLUMNS", 174, 10243, "Replace columns is not supported for table {0}. SerDe may be incompatible.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_REPLACE_COLUMNS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("BAD_LOCATION_VALUE", 175, 10244, "{0}  is not absolute.  Please specify a complete absolute uri.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg BAD_LOCATION_VALUE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("UNSUPPORTED_ALTER_TBL_OP", 176, 10245, "{0} alter table options is not supported", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_ALTER_TBL_OP> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_BIGTABLE_MAPJOIN", 177, 10246, "{0} table chosen for streaming is not valid", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_BIGTABLE_MAPJOIN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("MISSING_OVER_CLAUSE", 178, 10247, "Missing over clause for function : ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MISSING_OVER_CLAUSE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("PARTITION_SPEC_TYPE_MISMATCH", 179, 10248, "Cannot add partition column {0} of type {1} as it cannot be converted to type {2}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARTITION_SPEC_TYPE_MISMATCH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UNSUPPORTED_SUBQUERY_EXPRESSION", 180, 10249, "Unsupported SubQuery Expression");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_SUBQUERY_EXPRESSION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_SUBQUERY_EXPRESSION", 181, 10250, "Invalid SubQuery expression");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_SUBQUERY_EXPRESSION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_HDFS_URI", 182, 10251, "{0} is not a hdfs uri", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_HDFS_URI> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_DIR", 183, 10252, "{0} is not a directory", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_DIR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("NO_VALID_LOCATIONS", 184, 10253, "Could not find any valid location to place the jars. Please update hive.jar.directory or hive.user.install.directory with a valid location", 0);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_VALID_LOCATIONS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("UNSUPPORTED_AUTHORIZATION_PRINCIPAL_TYPE_GROUP", 185, 10254, "Principal type GROUP is not supported in this authorization setting", "28000");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_PRINCIPAL_TYPE_GROUP> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_TABLE_NAME", 186, 10255, "Invalid table name {0}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TABLE_NAME> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INSERT_INTO_IMMUTABLE_TABLE", 187, 10256, "Inserting into a non-empty immutable table is not allowed");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_INTO_IMMUTABLE_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_GLOBAL", 188, 10257, "Resource type GLOBAL is not supported in this authorization setting", "28000");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_GLOBAL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_COLUMN", 189, 10258, "Resource type COLUMN is not supported in this authorization setting", "28000");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TXNMGR_NOT_SPECIFIED", 190, 10260, "Transaction manager not specified correctly, set hive.txn.manager");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_SPECIFIED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TXNMGR_NOT_INSTANTIATED", 191, 10261, "Transaction manager could not be instantiated, check hive.txn.manager");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_INSTANTIATED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TXN_NO_SUCH_TRANSACTION", 192, 10262, "No record of transaction {0} could be found, may have timed out", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXN_NO_SUCH_TRANSACTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TXN_ABORTED", 193, 10263, "Transaction manager has aborted the transaction {0}.  Reason: {1}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXN_ABORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DBTXNMGR_REQUIRES_CONCURRENCY", 194, 10264, "To use DbTxnManager you must set hive.support.concurrency=true");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DBTXNMGR_REQUIRES_CONCURRENCY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TXNMGR_NOT_ACID", 195, 10265, "This command is not allowed on an ACID table {0}.{1} with a non-ACID transaction manager", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TXNMGR_NOT_ACID> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("LOCK_NO_SUCH_LOCK", 196, 10270, "No record of lock {0} could be found, may have timed out", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_NO_SUCH_LOCK> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOCK_REQUEST_UNSUPPORTED", 197, 10271, "Current transaction manager does not support explicit lock requests.  Transaction manager:  ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_REQUEST_UNSUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("METASTORE_COMMUNICATION_FAILED", 198, 10280, "Error communicating with the metastore");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg METASTORE_COMMUNICATION_FAILED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("METASTORE_COULD_NOT_INITIATE", 199, 10281, "Unable to initiate connection to the metastore.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg METASTORE_COULD_NOT_INITIATE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_COMPACTION_TYPE", 200, 10282, "Invalid compaction type, supported values are \'major\' and \'minor\'");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_COMPACTION_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NO_COMPACTION_PARTITION", 201, 10283, "You must specify a partition to compact for partitioned tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NO_COMPACTION_PARTITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("TOO_MANY_COMPACTION_PARTITIONS", 202, 10284, "Compaction can only be requested on one partition at a time.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TOO_MANY_COMPACTION_PARTITIONS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DISTINCT_NOT_SUPPORTED", 203, 10285, "Distinct keyword is not support in current context");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DISTINCT_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("NONACID_COMPACTION_NOT_SUPPORTED", 204, 10286, "Compaction is not allowed on non-ACID table {0}.{1}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NONACID_COMPACTION_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("MASKING_FILTERING_ON_ACID_NOT_SUPPORTED", 205, 10287, "Detected {0}.{1} has row masking/column filtering enabled, which is not supported for query involving ACID operations", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MASKING_FILTERING_ON_ACID_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("MASKING_FILTERING_ON_MATERIALIZED_VIEWS_SOURCES", 206, 10288, "Querying directly materialized view contents is not supported since we detected {0}.{1} used by materialized view has row masking/column filtering enabled", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MASKING_FILTERING_ON_MATERIALIZED_VIEWS_SOURCES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("MASKING_COMPLEX_TYPE_NOT_SUPPORTED", 207, 10289, "Masking complex types is not supported, found a masking expression {0} over column {1}:{2}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MASKING_COMPLEX_TYPE_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UPDATEDELETE_PARSE_ERROR", 208, 10290, "Encountered parse error while parsing rewritten merge/update or delete query");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATEDELETE_PARSE_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UPDATE_CANNOT_UPDATE_PART_VALUE", 209, 10292, "Updating values of partition columns is not supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATE_CANNOT_UPDATE_PART_VALUE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INSERT_CANNOT_CREATE_TEMP_FILE", 210, 10293, "Unable to create temp file for insert values ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INSERT_CANNOT_CREATE_TEMP_FILE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ACID_OP_ON_NONACID_TXNMGR", 211, 10294, "Attempt to do update or delete using transaction manager that does not support these operations.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_NONACID_TXNMGR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("VALUES_TABLE_CONSTRUCTOR_NOT_SUPPORTED", 212, 10296, "Values clause with table constructor not yet supported");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg VALUES_TABLE_CONSTRUCTOR_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ACID_OP_ON_NONACID_TABLE", 213, 10297, "Attempt to do update or delete on table {0} that is not transactional", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_NONACID_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ACID_NO_SORTED_BUCKETS", 214, 10298, "ACID insert, update, delete not supported on tables that are sorted, table {0}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_NO_SORTED_BUCKETS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ALTER_TABLE_TYPE_PARTIAL_PARTITION_SPEC_NO_SUPPORTED", 215, 10299, "Alter table partition type {0} does not allow partial partition spec", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_TYPE_PARTIAL_PARTITION_SPEC_NO_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ALTER_TABLE_PARTITION_CASCADE_NOT_SUPPORTED", 216, 10300, "Alter table partition type {0} does not support cascade", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_PARTITION_CASCADE_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DROP_NATIVE_FUNCTION", 217, 10301, "Cannot drop native function");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DROP_NATIVE_FUNCTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("UPDATE_CANNOT_UPDATE_BUCKET_VALUE", 218, 10302, "Updating values of bucketing columns is not supported.  Column {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATE_CANNOT_UPDATE_BUCKET_VALUE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("IMPORT_INTO_STRICT_REPL_TABLE", 219, 10303, "Non-repl import disallowed against table that is a destination of replication.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg IMPORT_INTO_STRICT_REPL_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_LOCATION_NONEMPTY", 220, 10304, "CREATE-TABLE-AS-SELECT cannot create table with location to a non-empty directory.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_LOCATION_NONEMPTY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTAS_CREATES_VOID_TYPE", 221, 10305, "CREATE-TABLE-AS-SELECT creates a VOID type, please use CAST to specify the type, near field: ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTAS_CREATES_VOID_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TBL_SORTED_NOT_BUCKETED", 222, 10306, "Destination table {0} found to be sorted but not bucketed.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TBL_SORTED_NOT_BUCKETED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("LOCK_ACQUIRE_TIMEDOUT", 223, 10307, "Lock acquisition for {0} timed out after {1}ms.  {2}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_ACQUIRE_TIMEDOUT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("COMPILE_LOCK_TIMED_OUT", 224, 10308, "Attempt to acquire compile lock timed out.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COMPILE_LOCK_TIMED_OUT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CANNOT_CHANGE_SERDE", 225, 10309, "Changing SerDe (from {0}) is not supported for table {1}. File format may be incompatible", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_SERDE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CANNOT_CHANGE_FILEFORMAT", 226, 10310, "Changing file format (from {0}) is not supported for table {1}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_FILEFORMAT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CANNOT_REORDER_COLUMNS", 227, 10311, "Reordering columns is not supported for table {0}. SerDe may be incompatible", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_REORDER_COLUMNS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CANNOT_CHANGE_COLUMN_TYPE", 228, 10312, "Changing from type {0} to {1} is not supported for column {2}. SerDe may be incompatible", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_CHANGE_COLUMN_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPLACE_CANNOT_DROP_COLUMNS", 229, 10313, "Replacing columns cannot drop columns for table {0}. SerDe may be incompatible", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_CANNOT_DROP_COLUMNS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPLACE_UNSUPPORTED_TYPE_CONVERSION", 230, 10314, "Replacing columns with unsupported type conversion (from {0} to {1}) for column {2}. SerDe may be incompatible", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_UNSUPPORTED_TYPE_CONVERSION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_AGGR_NOMAPAGGR_MULTIGBY", 231, 10315, "Grouping sets aggregations (with rollups or cubes) are not allowed when HIVE_MULTI_GROUPBY_SINGLE_REDUCER is turned on. Set hive.multigroupby.singlereducer=false if you want to use grouping sets");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_AGGR_NOMAPAGGR_MULTIGBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CANNOT_RETRIEVE_TABLE_METADATA", 232, 10316, "Error while retrieving table metadata");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CANNOT_RETRIEVE_TABLE_METADATA> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_AST_TREE", 233, 10318, "Internal error : Invalid AST");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_AST_TREE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ERROR_SERIALIZE_METASTORE", 234, 10319, "Error while serializing the metastore objects");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ERROR_SERIALIZE_METASTORE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("IO_ERROR", 235, 10320, "Error while performing IO operation ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg IO_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ERROR_SERIALIZE_METADATA", 236, 10321, "Error while serializing the metadata");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ERROR_SERIALIZE_METADATA> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_LOAD_TABLE_FILE_WORK", 237, 10322, "Invalid Load Table Work or Load File Work");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_LOAD_TABLE_FILE_WORK> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CLASSPATH_ERROR", 238, 10323, "Classpath error");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLASSPATH_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("IMPORT_SEMANTIC_ERROR", 239, 10324, "Import Semantic Analyzer Error");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg IMPORT_SEMANTIC_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_FK_SYNTAX", 240, 10325, "Invalid Foreign Key syntax");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FK_SYNTAX> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_CSTR_SYNTAX", 241, 10326, "Invalid Constraint syntax");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_CSTR_SYNTAX> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ACID_NOT_ENOUGH_HISTORY", 242, 10327, "Not enough history available for ({0},{1}).  Oldest available base: {2}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_NOT_ENOUGH_HISTORY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_COLUMN_NAME", 243, 10328, "Invalid column name");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_COLUMN_NAME> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UNSUPPORTED_SET_OPERATOR", 244, 10329, "Unsupported set operator");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNSUPPORTED_SET_OPERATOR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOCK_ACQUIRE_CANCELLED", 245, 10330, "Query was cancelled while acquiring locks on the underlying objects. ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOCK_ACQUIRE_CANCELLED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NOT_RECOGNIZED_CONSTRAINT", 246, 10331, "Constraint not recognized");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NOT_RECOGNIZED_CONSTRAINT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_CONSTRAINT", 247, 10332, "Invalid constraint definition");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_CONSTRAINT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPLACE_VIEW_WITH_MATERIALIZED", 248, 10400, "Attempt to replace view {0} with materialized view", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_VIEW_WITH_MATERIALIZED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPLACE_MATERIALIZED_WITH_VIEW", 249, 10401, "Attempt to replace materialized view {0} with view", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPLACE_MATERIALIZED_WITH_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("UPDATE_DELETE_VIEW", 250, 10402, "You cannot update or delete records in a view");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UPDATE_DELETE_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("MATERIALIZED_VIEW_DEF_EMPTY", 251, 10403, "Query for the materialized view rebuild could not be retrieved");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MATERIALIZED_VIEW_DEF_EMPTY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("MERGE_PREDIACTE_REQUIRED", 252, 10404, "MERGE statement with both UPDATE and DELETE clauses requires \"AND <boolean>\" on the 1st WHEN MATCHED clause of <{0}>", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MERGE_PREDIACTE_REQUIRED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("MERGE_TOO_MANY_DELETE", 253, 10405, "MERGE statement can have at most 1 WHEN MATCHED ... DELETE clause: <{0}>", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MERGE_TOO_MANY_DELETE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("MERGE_TOO_MANY_UPDATE", 254, 10406, "MERGE statement can have at most 1 WHEN MATCHED ... UPDATE clause: <{0}>", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg MERGE_TOO_MANY_UPDATE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_JOIN_CONDITION", 255, 10407, "Error parsing condition in join");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_JOIN_CONDITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_TARGET_COLUMN_IN_SET_CLAUSE", 256, 10408, "Target column \"{0}\" of set clause is not found in table \"{1}\".", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_TARGET_COLUMN_IN_SET_CLAUSE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_FUNCTION_EXPR_NOT_IN_GROUPBY", 257, 10409, "Expression in GROUPING function not present in GROUP BY");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_FUNCTION_EXPR_NOT_IN_GROUPBY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ALTER_TABLE_NON_PARTITIONED_TABLE_CASCADE_NOT_SUPPORTED", 258, 10410, "Alter table with non-partitioned table does not support cascade");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ALTER_TABLE_NON_PARTITIONED_TABLE_CASCADE_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("HIVE_GROUPING_SETS_SIZE_LIMIT", 259, 10411, "Grouping sets size cannot be greater than 64");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg HIVE_GROUPING_SETS_SIZE_LIMIT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REBUILD_NO_MATERIALIZED_VIEW", 260, 10412, "Rebuild command only valid for materialized views");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REBUILD_NO_MATERIALIZED_VIEW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("LOAD_DATA_ACID_FILE", 261, 10413, "\"{0}\" was created by Acid write - it cannot be loaded into anther Acid table", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_DATA_ACID_FILE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ACID_OP_ON_INSERTONLYTRAN_TABLE", 262, 10414, "Attempt to do update or delete on table {0} that is insert-only transactional", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_OP_ON_INSERTONLYTRAN_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("LOAD_DATA_LAUNCH_JOB_PARSE_ERROR", 263, 10416, "Encountered parse error while parsing rewritten load data into insert query");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg LOAD_DATA_LAUNCH_JOB_PARSE_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("RESOURCE_PLAN_ALREADY_EXISTS", 264, 10417, "Resource plan {0} already exists", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RESOURCE_PLAN_ALREADY_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("RESOURCE_PLAN_NOT_EXISTS", 265, 10418, "Resource plan {0} does not exist", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RESOURCE_PLAN_NOT_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INCOMPATIBLE_STRUCT", 266, 10419, "Incompatible structs.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INCOMPATIBLE_STRUCT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("OBJECTNAME_CONTAINS_DOT", 267, 10420, "Table or database name may not contain dot(.) character", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OBJECTNAME_CONTAINS_DOT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("WITHIN_GROUP_NOT_ALLOWED", 268, 10421, "Not an ordered-set aggregate function: {0}. WITHIN GROUP clause is not allowed.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WITHIN_GROUP_NOT_ALLOWED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("WITHIN_GROUP_PARAMETER_MISMATCH", 269, 10422, "The number of hypothetical direct arguments ({0}) must match the number of ordering columns ({1})", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WITHIN_GROUP_PARAMETER_MISMATCH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("AMBIGUOUS_STRUCT_ATTRIBUTE", 270, 10423, "Attribute \"{0}\" specified more than once in structured type.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg AMBIGUOUS_STRUCT_ATTRIBUTE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("OFFSET_NOT_SUPPORTED_IN_SUBQUERY", 271, 10424, "OFFSET is not supported in subquery of exists", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OFFSET_NOT_SUPPORTED_IN_SUBQUERY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("WITH_COL_LIST_NUM_OVERFLOW", 272, 10425, "WITH-clause query {0} returns {1} columns, but {2} labels were specified. The number of column labels must be smaller or equal to the number of expressions returned by the query.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WITH_COL_LIST_NUM_OVERFLOW> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("NULL_TREATMENT_NOT_SUPPORTED", 273, 10426, "Function {0} does not support null treatment.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NULL_TREATMENT_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("DATACONNECTOR_ALREADY_EXISTS", 274, 10427, "Dataconnector {0} already exists", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATACONNECTOR_ALREADY_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DATACONNECTOR_NOT_EXISTS", 275, 10428, "Dataconnector does not exist:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DATACONNECTOR_NOT_EXISTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TIME_TRAVEL_NOT_ALLOWED", 276, 10429, "Time travel is not allowed for {0}. Please choose a storage format which supports the feature.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TIME_TRAVEL_NOT_ALLOWED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("INVALID_METADATA_TABLE_NAME", 277, 10430, "Invalid metadata table name {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_METADATA_TABLE_NAME> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("TABLE_META_REF_NOT_SUPPORTED", 278, 10431, "Table Meta Ref extension is not supported for table {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg TABLE_META_REF_NOT_SUPPORTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("COMPACTION_REFUSED", 279, 10432, "Compaction request for {0}.{1}{2} is refused, details: {3}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COMPACTION_REFUSED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = <org.apache.hadoop.hive.conf.HiveConf$ConfVars: org.apache.hadoop.hive.conf.HiveConf$ConfVars HIVE_CBO_ENABLED>;
v = v.<org.apache.hadoop.hive.conf.HiveConf$ConfVars: java.lang.String varname>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (java.lang.String)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("The following functionality requires CBO (\u): {0}");
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CBO_IS_REQUIRED", 280, 10433, v, 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CBO_IS_REQUIRED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CTLF_UNSUPPORTED_FORMAT", 281, 10434, "CREATE TABLE LIKE FILE is not supported by the \'\'{0}\'\' file format", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_UNSUPPORTED_FORMAT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = <org.apache.hadoop.hive.conf.HiveConf$ConfVars: org.apache.hadoop.hive.conf.HiveConf$ConfVars SPLIT_UPDATE>;
v = v.<org.apache.hadoop.hive.conf.HiveConf$ConfVars: java.lang.String varname>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (java.lang.String)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Update and Merge to a non-native ACID table in \"merge-on-read\" mode is only supported when \"\u0001\"=\"true\"");
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("NON_NATIVE_ACID_UPDATE", 282, 10435, v);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg NON_NATIVE_ACID_UPDATE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("READ_ONLY_DATABASE", 283, 10436, "Database {0} is read-only", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg READ_ONLY_DATABASE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SCRIPT_INIT_ERROR", 284, 20000, "Unable to initialize custom script.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_INIT_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SCRIPT_IO_ERROR", 285, 20001, "An error occurred while reading or writing to your custom script. It may have crashed with an error.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_IO_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SCRIPT_GENERIC_ERROR", 286, 20002, "Hive encountered some unknown error while running your custom script.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_GENERIC_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SCRIPT_CLOSING_ERROR", 287, 20003, "An error occurred when trying to close the Operator running your custom script.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCRIPT_CLOSING_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DYNAMIC_PARTITIONS_TOO_MANY_PER_NODE_ERROR", 288, 20004, "Fatal error occurred when node tried to create too many dynamic partitions. The maximum number of dynamic partitions is controlled by hive.exec.max.dynamic.partitions and hive.exec.max.dynamic.partitions.pernode. ");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DYNAMIC_PARTITIONS_TOO_MANY_PER_NODE_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("OP_NOT_ALLOWED_IN_IMPLICIT_TXN", 289, 20006, "Operation {0} is not allowed in an implicit transaction ({1}).", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_IN_IMPLICIT_TXN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("OP_NOT_ALLOWED_IN_TXN", 290, 20007, "Operation {0} is not allowed in a transaction ({1},queryId={2}).", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_IN_TXN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("OP_NOT_ALLOWED_WITHOUT_TXN", 291, 20008, "Operation {0} is not allowed without an active transaction", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg OP_NOT_ALLOWED_WITHOUT_TXN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>("ACCESS_DENIED", 292, 20009, "Access denied: {0}", "42000", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACCESS_DENIED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>("QUOTA_EXCEEDED", 293, 20010, "Quota exceeded: {0}", "64000", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg QUOTA_EXCEEDED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>("UNRESOLVED_PATH", 294, 20011, "Unresolved path: {0}", "64000", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_PATH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>("FILE_NOT_FOUND", 295, 20012, "File not found: {0}", "64000", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg FILE_NOT_FOUND> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>("WRONG_FILE_FORMAT", 296, 20013, "Wrong file format. Please check the file\'s format.", "64000", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg WRONG_FILE_FORMAT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPL_FILE_MISSING_FROM_SRC_AND_CM_PATH", 297, 20016, "File is missing from both source and cm path.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FILE_MISSING_FROM_SRC_AND_CM_PATH> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_EXTERNAL_SERVICE_CONNECTION_ERROR", 298, 20017, "Failed to connect to {0} service. Error code {1}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_EXTERNAL_SERVICE_CONNECTION_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String)>("CLIENT_POLLING_OPSTATUS_INTERRUPTED", 299, 20018, "Interrupted while polling on the operation status", "70100");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CLIENT_POLLING_OPSTATUS_INTERRUPTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CTLF_FAILED_INFERENCE", 300, 20019, "Failed to infer schema:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_FAILED_INFERENCE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CTLF_CLASS_NOT_FOUND", 301, 20020, "Failed to find SerDe class ({0}) for \'\'{1}\'\'", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_CLASS_NOT_FOUND> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("CTLF_MISSING_STORAGE_FORMAT_DESCRIPTOR", 302, 20021, "Failed to find StorageFormatDescriptor for file format \'\'{0}\'\'", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CTLF_MISSING_STORAGE_FORMAT_DESCRIPTOR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("PARQUET_FOOTER_ERROR", 303, 20022, "Failed to read parquet footer:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARQUET_FOOTER_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("PARQUET_UNHANDLED_TYPE", 304, 20023, "Unhandled type {0}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg PARQUET_UNHANDLED_TYPE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ORC_FOOTER_ERROR", 305, 20024, "Failed to read orc footer:");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ORC_FOOTER_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,java.lang.String,boolean)>("UNRESOLVED_RT_EXCEPTION", 306, 29999, "Runtime Error: {0}", "58004", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg UNRESOLVED_RT_EXCEPTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("STATSPUBLISHER_NOT_OBTAINED", 307, 30000, "StatsPublisher cannot be obtained. There was a error to retrieve the StatsPublisher, and retrying might help. If you don\'t want the query to fail because accurate statistics could not be collected, set hive.stats.reliable=false");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_NOT_OBTAINED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("STATSPUBLISHER_INITIALIZATION_ERROR", 308, 30001, "StatsPublisher cannot be initialized. There was a error in the initialization of StatsPublisher, and retrying might help. If you don\'t want the query to fail because accurate statistics could not be collected, set hive.stats.reliable=false");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_INITIALIZATION_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("STATSPUBLISHER_CONNECTION_ERROR", 309, 30002, "StatsPublisher cannot be connected to.There was a error while connecting to the StatsPublisher, and retrying might help. If you don\'t want the query to fail because accurate statistics could not be collected, set hive.stats.reliable=false");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_CONNECTION_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("STATSPUBLISHER_PUBLISHING_ERROR", 310, 30003, "Error in publishing stats. There was an error in publishing stats via StatsPublisher, and retrying might help. If you don\'t want the query to fail because accurate statistics could not be collected, set hive.stats.reliable=false");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_PUBLISHING_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("STATSPUBLISHER_CLOSING_ERROR", 311, 30004, "StatsPublisher cannot be closed.There was a error while closing the StatsPublisher, and retrying might help. If you don\'t want the query to fail because accurate statistics could not be collected, set hive.stats.reliable=false");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSPUBLISHER_CLOSING_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("COLUMNSTATSCOLLECTOR_INVALID_PART_KEY", 312, 30005, "Invalid partitioning key specified in ANALYZE statement");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_PART_KEY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("COLUMNSTATSCOLLECTOR_INVALID_PARTITION", 313, 30007, "Invalid partitioning key/value specified in ANALYZE statement");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_PARTITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("COLUMNSTATSCOLLECTOR_PARSE_ERROR", 314, 30009, "Encountered parse error while parsing rewritten query");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_PARSE_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("DROP_COMMAND_NOT_ALLOWED_FOR_PARTITION", 315, 30011, "Partition protected from being dropped");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg DROP_COMMAND_NOT_ALLOWED_FOR_PARTITION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("COLUMNSTATSCOLLECTOR_INVALID_COLUMN", 316, 30012, "Column statistics are not supported for partition columns");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg COLUMNSTATSCOLLECTOR_INVALID_COLUMN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("STATSAGGREGATOR_SOURCETASK_NULL", 317, 30014, "SourceTask of StatsTask should not be null");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSAGGREGATOR_SOURCETASK_NULL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("STATSAGGREGATOR_CONNECTION_ERROR", 318, 30015, "Stats aggregator of type {0} cannot be connected to", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATSAGGREGATOR_CONNECTION_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("STATS_SKIPPING_BY_ERROR", 319, 30017, "Skipping stats aggregation by error {0}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg STATS_SKIPPING_BY_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("INVALID_FILE_FORMAT_IN_LOAD", 320, 30019, "The file that you are trying to load does not match the file format of the destination table.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg INVALID_FILE_FORMAT_IN_LOAD> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("SCHEMA_REQUIRED_TO_READ_ACID_TABLES", 321, 30020, "Neither the configuration variables schema.evolution.columns / schema.evolution.columns.types nor the columns / columns.types are set.  Table schema information is required to read ACID tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg SCHEMA_REQUIRED_TO_READ_ACID_TABLES> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ACID_TABLES_MUST_BE_READ_WITH_ACID_READER", 322, 30021, "An ORC ACID reader required to read ACID tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_TABLES_MUST_BE_READ_WITH_ACID_READER> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("ACID_TABLES_MUST_BE_READ_WITH_HIVEINPUTFORMAT", 323, 30022, "Must use HiveInputFormat to read ACID tables (set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat)");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_TABLES_MUST_BE_READ_WITH_HIVEINPUTFORMAT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("ACID_LOAD_DATA_INVALID_FILE_NAME", 324, 30023, "{0} file name is not valid in Load Data into Acid table {1}.  Examples of valid names are: 00000_0, 00000_0_copy_1", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg ACID_LOAD_DATA_INVALID_FILE_NAME> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CONCATENATE_UNSUPPORTED_FILE_FORMAT", 325, 30030, "Concatenate/Merge only supported for RCFile and ORCFile formats");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_FILE_FORMAT> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CONCATENATE_UNSUPPORTED_TABLE_BUCKETED", 326, 30031, "Concatenate/Merge can not be performed on bucketed tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_BUCKETED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CONCATENATE_UNSUPPORTED_PARTITION_ARCHIVED", 327, 30032, "Concatenate/Merge can not be performed on archived partitions");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_PARTITION_ARCHIVED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CONCATENATE_UNSUPPORTED_TABLE_NON_NATIVE", 328, 30033, "Concatenate/Merge can not be performed on non-native tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_NON_NATIVE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CONCATENATE_UNSUPPORTED_TABLE_NOT_MANAGED", 329, 30034, "Concatenate/Merge can only be performed on managed tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_NOT_MANAGED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("CONCATENATE_UNSUPPORTED_TABLE_TRANSACTIONAL", 330, 30035, "Concatenate/Merge can not be performed on transactional tables");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg CONCATENATE_UNSUPPORTED_TABLE_TRANSACTIONAL> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_FILE_SYSTEM_OPERATION_RETRY", 331, 30047, "Replication file system operation retry expired. Error {0}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FILE_SYSTEM_OPERATION_RETRY> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_SOURCE_DATABASE_NOT_FOUND", 332, 30048, "Cannot dump database {0} as it does not exist", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_SOURCE_DATABASE_NOT_FOUND> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40003;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPL_DATABASE_IS_TARGET_OF_REPLICATION", 333, v, "Cannot dump database as it is a Target of replication.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_DATABASE_IS_TARGET_OF_REPLICATION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40005;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPL_INVALID_DB_OR_TABLE_PATTERN", 334, v, "Invalid pattern for the DB or table name in the replication policy. It should be a valid regex enclosed within single or double quotes.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_DB_OR_TABLE_PATTERN> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40006;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPL_EVENTS_MISSING_IN_METASTORE", 335, v, "Notification events are missing in the meta store.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_EVENTS_MISSING_IN_METASTORE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40007;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_BOOTSTRAP_LOAD_PATH_NOT_VALID", 336, v, "Load path {0} not valid as target database is bootstrapped from some other path : {1}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_BOOTSTRAP_LOAD_PATH_NOT_VALID> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40008;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_INVALID_CONFIG_FOR_SERVICE", 337, v, "Invalid config error : {0} for {1} service.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_CONFIG_FOR_SERVICE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40009;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_INVALID_INTERNAL_CONFIG_FOR_SERVICE", 338, v, "Invalid internal config error : {0} for {1} service.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_INTERNAL_CONFIG_FOR_SERVICE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40010;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_RETRY_EXHAUSTED", 339, v, "Retry exhausted for retryable error code {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_RETRY_EXHAUSTED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40011;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPL_FAILED_WITH_NON_RECOVERABLE_ERROR", 340, v, "Replication failed with non recoverable error. Needs manual intervention");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FAILED_WITH_NON_RECOVERABLE_ERROR> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40012;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_INVALID_ARGUMENTS", 341, v, "Invalid arguments error : {0}.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_ARGUMENTS> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40013;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_INVALID_ALTER_TABLE", 342, v, "{0}Unable to alter table{1}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INVALID_ALTER_TABLE> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40014;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_PERMISSION_DENIED", 343, v, "{0}org.apache.hadoop.security.AccessControlException{1}", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_PERMISSION_DENIED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40015;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_DISTCP_SNAPSHOT_EXCEPTION", 344, v, "SNAPSHOT_ERROR", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_DISTCP_SNAPSHOT_EXCEPTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40016;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("RANGER_AUTHORIZATION_FAILED", 345, v, "Authorization Failure while communicating to Ranger admin", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RANGER_AUTHORIZATION_FAILED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40017;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("RANGER_AUTHENTICATION_FAILED", 346, v, "Authentication Failure while communicating to Ranger admin", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg RANGER_AUTHENTICATION_FAILED> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40018;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String,boolean)>("REPL_INCOMPATIBLE_EXCEPTION", 347, v, "Cannot load into database {0} as it is replication incompatible.", 1);
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_INCOMPATIBLE_EXCEPTION> = v;
v = new org.apache.hadoop.hive.ql.ErrorMsg;
v = (int) 40019;
specialinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: void <init>(java.lang.String,int,int,java.lang.String)>("REPL_FAILOVER_TARGET_MODIFIED", 348, v, "Database event id changed post table diff generation.");
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg REPL_FAILOVER_TARGET_MODIFIED> = v;
v = staticinvoke <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg[] $values()>();
<org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg[] $VALUES> = v;
v = staticinvoke <java.util.regex.Pattern: java.util.regex.Pattern compile(java.lang.String)>(".*Line [0-9]+:[0-9]+ (.*)");
<org.apache.hadoop.hive.ql.ErrorMsg: java.util.regex.Pattern ERROR_MESSAGE_PATTERN> = v;
v = staticinvoke <java.util.regex.Pattern: java.util.regex.Pattern compile(java.lang.String)>("HiveException:\\s+\\[Error ([0-9]+)\\]: (.*)");
<org.apache.hadoop.hive.ql.ErrorMsg: java.util.regex.Pattern ERROR_CODE_PATTERN> = v;
v = new java.util.HashMap;
specialinvoke v.<java.util.HashMap: void <init>()>();
<org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map mesgToErrorMsgMap> = v;
v = new java.util.HashMap;
specialinvoke v.<java.util.HashMap: void <init>()>();
<org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map formatToErrorMsgMap> = v;
v = (int) -1;
<org.apache.hadoop.hive.ql.ErrorMsg: int minMesgLength> = v;
v = staticinvoke <org.apache.hadoop.hive.ql.ErrorMsg: org.apache.hadoop.hive.ql.ErrorMsg[] values()>();
v = lengthof v;
v = 0;
label:
if v >= v goto label;
v = v[v];
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.text.MessageFormat format>;
if v == null goto label;
v = v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String mesg>;
v = virtualinvoke v.<java.lang.String: java.lang.String replaceAll(java.lang.String,java.lang.String)>("\\{[0-9]+\\}", ".*");
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map formatToErrorMsgMap>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (java.lang.String)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("^\u0001$");
v = staticinvoke <java.util.regex.Pattern: java.util.regex.Pattern compile(java.lang.String,int)>(v, 32);
interfaceinvoke v.<java.util.Map: java.lang.Object put(java.lang.Object,java.lang.Object)>(v, v);
goto label;
label:
v = <org.apache.hadoop.hive.ql.ErrorMsg: java.util.Map mesgToErrorMsgMap>;
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String getMsg()>();
v = virtualinvoke v.<java.lang.String: java.lang.String trim()>();
interfaceinvoke v.<java.util.Map: java.lang.Object put(java.lang.Object,java.lang.Object)>(v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.ql.ErrorMsg: java.lang.String getMsg()>();
v = virtualinvoke v.<java.lang.String: java.lang.String trim()>();
v = virtualinvoke v.<java.lang.String: int length()>();
v = <org.apache.hadoop.hive.ql.ErrorMsg: int minMesgLength>;
v = (int) -1;
if v == v goto label;
v = <org.apache.hadoop.hive.ql.ErrorMsg: int minMesgLength>;
if v >= v goto label;
label:
<org.apache.hadoop.hive.ql.ErrorMsg: int minMesgLength> = v;
label:
v = v + 1;
goto label;
label:
return;
}
}