public class org.apache.hadoop.hive.druid.serde.TestDruidSerDe extends java.lang.Object
{
private static final java.lang.String TIMESERIES_QUERY;
private static final java.lang.String TIMESERIES_QUERY_RESULTS;
private byte[] tsQueryResults;
private byte[] topNQueryResults;
private byte[] groupByQueryResults;
private byte[] groupByTimeExtractQueryResults;
private byte[] selectQueryResults;
private byte[] groupByMonthExtractQueryResults;
private byte[] scanQueryResults;
private static final java.lang.Object[][] TIMESERIES_QUERY_RESULTS_RECORDS;
private static final java.lang.String TIMESERIES_COLUMN_NAMES;
private static final java.lang.String TIMESERIES_COLUMN_TYPES;
private static final java.lang.String TOPN_QUERY;
private static final java.lang.String TOPN_QUERY_RESULTS;
private static final java.lang.Object[][] TOPN_QUERY_RESULTS_RECORDS;
private static final java.lang.String TOPN_COLUMN_NAMES;
private static final java.lang.String TOPN_COLUMN_TYPES;
private static final java.lang.String GROUP_BY_QUERY;
private static final java.lang.String GROUP_BY_QUERY_RESULTS;
private static final java.lang.String GB_TIME_EXTRACTIONS;
private static final java.lang.String GB_TIME_EXTRACTIONS_RESULTS;
private static final java.lang.String GB_MONTH_EXTRACTIONS_RESULTS;
private static final java.lang.String GB_MONTH_EXTRACTIONS;
private static final java.lang.Object[][] GROUP_BY_QUERY_EXTRACTION_RESULTS_RECORDS;
private static final java.lang.Object[][] GROUP_BY_QUERY_RESULTS_RECORDS;
private static final java.lang.Object[][] GB_MONTH_EXTRACTION_RESULTS_RECORDS;
private static final java.lang.String GROUP_BY_COLUMN_NAMES;
private static final java.lang.String GROUP_BY_COLUMN_TYPES;
private static final java.lang.String GB_TIME_EXTRACTIONS_COLUMN_NAMES;
private static final java.lang.String GB_TIME_EXTRACTIONS_COLUMN_TYPES;
private static final java.lang.String GB_MONTH_EXTRACTIONS_COLUMN_NAMES;
private static final java.lang.String GB_MONTH_EXTRACTIONS_COLUMN_TYPES;
private static final java.lang.String SCAN_COLUMN_NAMES;
private static final java.lang.String SCAN_COLUMN_TYPES;
private static final java.lang.Object[][] SCAN_QUERY_RESULTS_RECORDS;
private static final java.lang.String SCAN_QUERY;
private static final java.lang.String SCAN_QUERY_RESULTS;
private static final java.lang.String COLUMN_NAMES;
private static final java.lang.String COLUMN_TYPES;
private static final java.lang.Object[] ROW_OBJECT;
private static final org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE;
public org.junit.rules.ExpectedException expectedEx;
private static final java.lang.Object[] ROW_OBJECT_2;
private static final org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE_2;
public void <init>()
{
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
org.junit.rules.ExpectedException v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
specialinvoke v.<java.lang.Object: void <init>()>();
v = staticinvoke <org.junit.rules.ExpectedException: org.junit.rules.ExpectedException none()>();
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.junit.rules.ExpectedException expectedEx> = v;
return;
}
public void setup() throws java.io.IOException
{
byte[] v, v, v, v, v, v;
org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe$6 v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe$5 v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe$4 v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe$3 v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe$2 v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe$1 v;
java.lang.Object v, v, v, v, v, v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = new org.apache.hadoop.hive.druid.serde.TestDruidSerDe$1;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe$1: void <init>(org.apache.hadoop.hive.druid.serde.TestDruidSerDe)>(v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: java.lang.Object readValue(java.lang.String,org.apache.hive.druid.com.fasterxml.jackson.core.type.TypeReference)>("[  {    \"timestamp\": \"2012-01-01T00:00:00.000Z\",    \"result\": { \"sample_name1\": 0, \"sample_name2\": 1.0, \"sample_divide\": 2.2222 }   },  {    \"timestamp\": \"2012-01-02T00:00:00.000Z\",    \"result\": { \"sample_name1\": 2, \"sample_name2\": 3.32, \"sample_divide\": 4 }  }]", v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: byte[] writeValueAsBytes(java.lang.Object)>(v);
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] tsQueryResults> = v;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = new org.apache.hadoop.hive.druid.serde.TestDruidSerDe$2;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe$2: void <init>(org.apache.hadoop.hive.druid.serde.TestDruidSerDe)>(v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: java.lang.Object readValue(java.lang.String,org.apache.hive.druid.com.fasterxml.jackson.core.type.TypeReference)>("[  {    \"timestamp\": \"2013-08-31T00:00:00.000Z\",    \"result\": [      {        \"sample_dim\": \"dim1_val\",        \"count\": 111,        \"some_metric\": 10669,        \"sample_divide\": 96.11711711711712      },      {        \"sample_dim\": \"another_dim1_val\",        \"count\": 88,        \"some_metric\": 28344,        \"sample_divide\": 322.09090909090907      },      {        \"sample_dim\": \"dim1_val3\",        \"count\": 70,        \"some_metric\": 871,        \"sample_divide\": 12.442857142857143      },      {        \"sample_dim\": \"dim1_val4\",        \"count\": 62,        \"some_metric\": 815,        \"sample_divide\": 13.14516129032258      },      {        \"sample_dim\": \"dim1_val5\",        \"count\": 60,        \"some_metric\": 2787,        \"sample_divide\": 46.45      }    ]  }]", v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: byte[] writeValueAsBytes(java.lang.Object)>(v);
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] topNQueryResults> = v;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = new org.apache.hadoop.hive.druid.serde.TestDruidSerDe$3;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe$3: void <init>(org.apache.hadoop.hive.druid.serde.TestDruidSerDe)>(v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: java.lang.Object readValue(java.lang.String,org.apache.hive.druid.com.fasterxml.jackson.core.type.TypeReference)>("[   {    \"version\" : \"v1\",    \"timestamp\" : \"2012-01-01T00:00:00.000Z\",    \"event\" : {      \"country\" : \"India\",      \"device\" : \"phone\",      \"total_usage\" : 88,      \"data_transfer\" : 29.91233453,      \"avg_usage\" : 60.32    }  },   {    \"version\" : \"v1\",    \"timestamp\" : \"2012-01-01T00:00:12.000Z\",    \"event\" : {      \"country\" : \"Spain\",      \"device\" : \"pc\",      \"total_usage\" : 16,      \"data_transfer\" : 172.93494959,      \"avg_usage\" : 6.333333    }  }]", v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: byte[] writeValueAsBytes(java.lang.Object)>(v);
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] groupByQueryResults> = v;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = new org.apache.hadoop.hive.druid.serde.TestDruidSerDe$4;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe$4: void <init>(org.apache.hadoop.hive.druid.serde.TestDruidSerDe)>(v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: java.lang.Object readValue(java.lang.String,org.apache.hive.druid.com.fasterxml.jackson.core.type.TypeReference)>("[   {    \"version\" : \"v1\",    \"timestamp\" : \"2012-01-01T00:00:00.000Z\",    \"event\" : {      \"extract\" : \"2012-01-01T00:00:00.000Z\",      \"$f\" : 200  }  },   {    \"version\" : \"v1\",    \"timestamp\" : \"2012-01-01T00:00:12.000Z\",    \"event\" : {      \"extract\" : \"2012-01-01T00:00:12.000Z\",      \"$f\" : 400  }   }]", v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: byte[] writeValueAsBytes(java.lang.Object)>(v);
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] groupByTimeExtractQueryResults> = v;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = new org.apache.hadoop.hive.druid.serde.TestDruidSerDe$5;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe$5: void <init>(org.apache.hadoop.hive.druid.serde.TestDruidSerDe)>(v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: java.lang.Object readValue(java.lang.String,org.apache.hive.druid.com.fasterxml.jackson.core.type.TypeReference)>("[   {    \"version\" : \"v1\",    \"timestamp\" : \"2012-01-01T00:00:00.000Z\",    \"event\" : {      \"extract_month\" : \"01\",      \"$f\" : 200  }  },   {    \"version\" : \"v1\",    \"timestamp\" : \"2012-01-01T00:00:12.000Z\",    \"event\" : {      \"extract_month\" : \"01\",      \"$f\" : 400  }   }]", v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: byte[] writeValueAsBytes(java.lang.Object)>(v);
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] groupByMonthExtractQueryResults> = v;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = new org.apache.hadoop.hive.druid.serde.TestDruidSerDe$6;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe$6: void <init>(org.apache.hadoop.hive.druid.serde.TestDruidSerDe)>(v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: java.lang.Object readValue(java.lang.String,org.apache.hive.druid.com.fasterxml.jackson.core.type.TypeReference)>("[{\"segmentId\":\"wikipedia_2012-12-29T00:00:00.000Z_2013-01-10T08:00:00.000Z_2013-01-10T08:13:47.830Z_v9\",\"columns\":[\"__time\",\"robot\",\"namespace\",\"anonymous\",\"unpatrolled\",\"page\",\"language\",\"newpage\",\"user\",\"count\",\"added\",\"delta\",\"variation\",\"deleted\"],\"events\":[[\"2013-01-01T00:00:00.000Z\", 1,\"article\",\"0\",\"0\",\"11._korpus_(NOVJ)\",\"sl\",\"0\",\"EmausBot\",1.0,39.0,39.0,39.0,0.0],[\"2013-01-01T00:00:00.000Z\", 0,\"article\",\"0\",\"0\",\"112_U.S._580\",\"en\",\"1\",\"MZMcBride\",1.0,70.0,70.0,70.0,0.0],[\"2013-01-01T00:00:12.000Z\", 0,\"article\",\"0\",\"0\",\"113_U.S._243\",\"en\",\"1\",\"MZMcBride\",1.0,77.0,77.0,77.0,0.0],[\"2013-01-01T00:00:12.000Z\", 0,\"article\",\"0\",\"0\",\"113_U.S._73\",\"en\",\"1\",\"MZMcBride\",1.0,70.0,70.0,70.0,0.0],[\"2013-01-01T00:00:12.000Z\", 0,\"article\",\"0\",\"0\",\"113_U.S._756\",\"en\",\"1\",\"MZMcBride\",1.0,68.0,68.0,68.0,0.0]]}]", v);
v = virtualinvoke v.<org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper: byte[] writeValueAsBytes(java.lang.Object)>(v);
v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] scanQueryResults> = v;
return;
}
public void testDruidDeserializer() throws org.apache.hadoop.hive.serde.SerDeException, java.lang.NoSuchFieldException, java.lang.SecurityException, java.lang.IllegalArgumentException, java.lang.IllegalAccessException, java.io.IOException, java.lang.InterruptedException, java.lang.NoSuchMethodException, java.lang.reflect.InvocationTargetException
{
byte[] v, v, v, v, v, v;
java.lang.Object[][] v, v, v, v, v, v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
org.apache.hadoop.conf.Configuration v;
org.apache.hadoop.hive.druid.QTestDruidSerDe v;
java.util.Properties v, v, v, v, v, v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
v = new org.apache.hadoop.hive.druid.QTestDruidSerDe;
specialinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void <init>()>();
v = new org.apache.hadoop.conf.Configuration;
specialinvoke v.<org.apache.hadoop.conf.Configuration: void <init>()>();
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesQuery(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.lang.String)>("sample_datasource", "timeseries", "{  \"queryType\": \"timeseries\",  \"dataSource\": \"sample_datasource\",  \"granularity\": \"day\",  \"descending\": \"true\",  \"filter\": {    \"type\": \"and\",    \"fields\": [      { \"type\": \"selector\", \"dimension\": \"sample_dimension1\", \"value\": \"sample_value1\" },      { \"type\": \"or\",        \"fields\": [          { \"type\": \"selector\", \"dimension\": \"sample_dimension2\", \"value\": \"sample_value2\" },          { \"type\": \"selector\", \"dimension\": \"sample_dimension3\", \"value\": \"sample_value3\" }        ]      }    ]  },  \"aggregations\": [    { \"type\": \"longSum\", \"name\": \"sample_name1\", \"fieldName\": \"sample_fieldName1\" },    { \"type\": \"doubleSum\", \"name\": \"sample_name2\", \"fieldName\": \"sample_fieldName2\" }  ],  \"postAggregations\": [    { \"type\": \"arithmetic\",      \"name\": \"sample_divide\",      \"fn\": \"/\",      \"fields\": [        { \"type\": \"fieldAccess\", \"name\": \"postAgg__sample_name1\", \"fieldName\": \"sample_name1\" },        { \"type\": \"fieldAccess\", \"name\": \"postAgg__sample_name2\", \"fieldName\": \"sample_name2\" }      ]    }  ],  \"intervals\": [ \"2012-01-01T00:00:00.000/2012-01-03T00:00:00.000\" ]}", "timestamp,sample_name,sample_name,sample_divide", "timestamp with local time zone,bigint,float,float");
virtualinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] tsQueryResults>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] TIMESERIES_QUERY_RESULTS_RECORDS>;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.String,java.lang.String,byte[],java.lang.Object[][])>(v, "timeseries", "{  \"queryType\": \"timeseries\",  \"dataSource\": \"sample_datasource\",  \"granularity\": \"day\",  \"descending\": \"true\",  \"filter\": {    \"type\": \"and\",    \"fields\": [      { \"type\": \"selector\", \"dimension\": \"sample_dimension1\", \"value\": \"sample_value1\" },      { \"type\": \"or\",        \"fields\": [          { \"type\": \"selector\", \"dimension\": \"sample_dimension2\", \"value\": \"sample_value2\" },          { \"type\": \"selector\", \"dimension\": \"sample_dimension3\", \"value\": \"sample_value3\" }        ]      }    ]  },  \"aggregations\": [    { \"type\": \"longSum\", \"name\": \"sample_name1\", \"fieldName\": \"sample_fieldName1\" },    { \"type\": \"doubleSum\", \"name\": \"sample_name2\", \"fieldName\": \"sample_fieldName2\" }  ],  \"postAggregations\": [    { \"type\": \"arithmetic\",      \"name\": \"sample_divide\",      \"fn\": \"/\",      \"fields\": [        { \"type\": \"fieldAccess\", \"name\": \"postAgg__sample_name1\", \"fieldName\": \"sample_name1\" },        { \"type\": \"fieldAccess\", \"name\": \"postAgg__sample_name2\", \"fieldName\": \"sample_name2\" }      ]    }  ],  \"intervals\": [ \"2012-01-01T00:00:00.000/2012-01-03T00:00:00.000\" ]}", v, v);
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesQuery(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.lang.String)>("sample_data", "topN", "{  \"queryType\": \"topN\",  \"dataSource\": \"sample_data\",  \"dimension\": \"sample_dim\",  \"threshold\": 5,  \"metric\": \"count\",  \"granularity\": \"all\",  \"filter\": {    \"type\": \"and\",    \"fields\": [      {        \"type\": \"selector\",        \"dimension\": \"dim1\",        \"value\": \"some_value\"      },      {        \"type\": \"selector\",        \"dimension\": \"dim2\",        \"value\": \"some_other_val\"      }    ]  },  \"aggregations\": [    {      \"type\": \"longSum\",      \"name\": \"count\",      \"fieldName\": \"count\"    },    {      \"type\": \"doubleSum\",      \"name\": \"some_metric\",      \"fieldName\": \"some_metric\"    }  ],  \"postAggregations\": [    {      \"type\": \"arithmetic\",      \"name\": \"sample_divide\",      \"fn\": \"/\",      \"fields\": [        {          \"type\": \"fieldAccess\",          \"name\": \"some_metric\",          \"fieldName\": \"some_metric\"        },        {          \"type\": \"fieldAccess\",          \"name\": \"count\",          \"fieldName\": \"count\"        }      ]    }  ],  \"intervals\": [    \"2013-08-31T00:00:00.000/2013-09-03T00:00:00.000\"  ]}", "timestamp,sample_dim,count,some_metric,sample_divide", "timestamp with local time zone,string,bigint,float,float");
virtualinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] topNQueryResults>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] TOPN_QUERY_RESULTS_RECORDS>;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.String,java.lang.String,byte[],java.lang.Object[][])>(v, "topN", "{  \"queryType\": \"topN\",  \"dataSource\": \"sample_data\",  \"dimension\": \"sample_dim\",  \"threshold\": 5,  \"metric\": \"count\",  \"granularity\": \"all\",  \"filter\": {    \"type\": \"and\",    \"fields\": [      {        \"type\": \"selector\",        \"dimension\": \"dim1\",        \"value\": \"some_value\"      },      {        \"type\": \"selector\",        \"dimension\": \"dim2\",        \"value\": \"some_other_val\"      }    ]  },  \"aggregations\": [    {      \"type\": \"longSum\",      \"name\": \"count\",      \"fieldName\": \"count\"    },    {      \"type\": \"doubleSum\",      \"name\": \"some_metric\",      \"fieldName\": \"some_metric\"    }  ],  \"postAggregations\": [    {      \"type\": \"arithmetic\",      \"name\": \"sample_divide\",      \"fn\": \"/\",      \"fields\": [        {          \"type\": \"fieldAccess\",          \"name\": \"some_metric\",          \"fieldName\": \"some_metric\"        },        {          \"type\": \"fieldAccess\",          \"name\": \"count\",          \"fieldName\": \"count\"        }      ]    }  ],  \"intervals\": [    \"2013-08-31T00:00:00.000/2013-09-03T00:00:00.000\"  ]}", v, v);
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesQuery(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.lang.String)>("sample_datasource", "groupBy", "{  \"queryType\": \"groupBy\",  \"dataSource\": \"sample_datasource\",  \"granularity\": \"day\",  \"dimensions\": [\"country\", \"device\"],  \"limitSpec\": { \"type\": \"default\", \"limit\": 5000, \"columns\": [\"country\", \"data_transfer\"] },  \"filter\": {    \"type\": \"and\",    \"fields\": [      { \"type\": \"selector\", \"dimension\": \"carrier\", \"value\": \"AT&T\" },      { \"type\": \"or\",         \"fields\": [          { \"type\": \"selector\", \"dimension\": \"make\", \"value\": \"Apple\" },          { \"type\": \"selector\", \"dimension\": \"make\", \"value\": \"Samsung\" }        ]      }    ]  },  \"aggregations\": [    { \"type\": \"longSum\", \"name\": \"total_usage\", \"fieldName\": \"user_count\" },    { \"type\": \"doubleSum\", \"name\": \"data_transfer\", \"fieldName\": \"data_transfer\" }  ],  \"postAggregations\": [    { \"type\": \"arithmetic\",      \"name\": \"avg_usage\",      \"fn\": \"/\",      \"fields\": [        { \"type\": \"fieldAccess\", \"fieldName\": \"data_transfer\" },        { \"type\": \"fieldAccess\", \"fieldName\": \"total_usage\" }      ]    }  ],  \"intervals\": [ \"2012-01-01T00:00:00.000/2012-01-03T00:00:00.000\" ],  \"having\": {    \"type\": \"greaterThan\",    \"aggregation\": \"total_usage\",    \"value\": 100  }}", "timestamp,country,device,total_usage,data_transfer,avg_usage", "timestamp with local time zone,string,string,bigint,double,float");
virtualinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] groupByQueryResults>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] GROUP_BY_QUERY_RESULTS_RECORDS>;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.String,java.lang.String,byte[],java.lang.Object[][])>(v, "groupBy", "{  \"queryType\": \"groupBy\",  \"dataSource\": \"sample_datasource\",  \"granularity\": \"day\",  \"dimensions\": [\"country\", \"device\"],  \"limitSpec\": { \"type\": \"default\", \"limit\": 5000, \"columns\": [\"country\", \"data_transfer\"] },  \"filter\": {    \"type\": \"and\",    \"fields\": [      { \"type\": \"selector\", \"dimension\": \"carrier\", \"value\": \"AT&T\" },      { \"type\": \"or\",         \"fields\": [          { \"type\": \"selector\", \"dimension\": \"make\", \"value\": \"Apple\" },          { \"type\": \"selector\", \"dimension\": \"make\", \"value\": \"Samsung\" }        ]      }    ]  },  \"aggregations\": [    { \"type\": \"longSum\", \"name\": \"total_usage\", \"fieldName\": \"user_count\" },    { \"type\": \"doubleSum\", \"name\": \"data_transfer\", \"fieldName\": \"data_transfer\" }  ],  \"postAggregations\": [    { \"type\": \"arithmetic\",      \"name\": \"avg_usage\",      \"fn\": \"/\",      \"fields\": [        { \"type\": \"fieldAccess\", \"fieldName\": \"data_transfer\" },        { \"type\": \"fieldAccess\", \"fieldName\": \"total_usage\" }      ]    }  ],  \"intervals\": [ \"2012-01-01T00:00:00.000/2012-01-03T00:00:00.000\" ],  \"having\": {    \"type\": \"greaterThan\",    \"aggregation\": \"total_usage\",    \"value\": 100  }}", v, v);
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesQuery(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.lang.String)>("sample_datasource", "groupBy", "{\"queryType\":\"groupBy\",\"dataSource\":\"sample_datasource\",\"granularity\":\"all\",\"dimensions\":[{\"type\":\"extraction\",\"dimension\":\"__time\",\"outputName\":\"extract\",\"extractionFn\":{\"type\":\"timeFormat\",\"format\":\"yyyy-MM-dd\'T\'HH:mm:ss.SSS\'Z\'\",\"timeZone\":\"UTC\"}}],\"limitSpec\":{\"type\":\"default\"},\"aggregations\":[{\"type\":\"count\",\"name\":\"$f\"}],\"intervals\":[\"1900-01-01T00:00:00.000/3000-01-01T00:00:00.000\"]}", "timestamp,extract,$f", "timestamp with local time zone,timestamp with local time zone,bigint");
virtualinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] groupByTimeExtractQueryResults>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] GROUP_BY_QUERY_EXTRACTION_RESULTS_RECORDS>;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.String,java.lang.String,byte[],java.lang.Object[][])>(v, "groupBy", "{\"queryType\":\"groupBy\",\"dataSource\":\"sample_datasource\",\"granularity\":\"all\",\"dimensions\":[{\"type\":\"extraction\",\"dimension\":\"__time\",\"outputName\":\"extract\",\"extractionFn\":{\"type\":\"timeFormat\",\"format\":\"yyyy-MM-dd\'T\'HH:mm:ss.SSS\'Z\'\",\"timeZone\":\"UTC\"}}],\"limitSpec\":{\"type\":\"default\"},\"aggregations\":[{\"type\":\"count\",\"name\":\"$f\"}],\"intervals\":[\"1900-01-01T00:00:00.000/3000-01-01T00:00:00.000\"]}", v, v);
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesQuery(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.lang.String)>("sample_datasource", "groupBy", "{\"queryType\":\"groupBy\",\"dataSource\":\"sample_datasource\",\"granularity\":\"all\",\"dimensions\":[{\"type\":\"extraction\",\"dimension\":\"__time\",\"outputName\":\"extract_month\",\"extractionFn\":{\"type\":\"timeFormat\",\"format\":\"M\",\"timeZone\":\"UTC\",\"locale\":\"en-US\"}}],\"limitSpec\":{\"type\":\"default\"},\"aggregations\":[{\"type\":\"count\",\"name\":\"$f\"}],\"intervals\":[\"1900-01-01T00:00:00.000/3000-01-01T00:00:00.000\"]}", "timestamp,extract_month,$f", "timestamp with local time zone,int,bigint");
virtualinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] groupByMonthExtractQueryResults>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] GB_MONTH_EXTRACTION_RESULTS_RECORDS>;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.String,java.lang.String,byte[],java.lang.Object[][])>(v, "groupBy", "{\"queryType\":\"groupBy\",\"dataSource\":\"sample_datasource\",\"granularity\":\"all\",\"dimensions\":[{\"type\":\"extraction\",\"dimension\":\"__time\",\"outputName\":\"extract_month\",\"extractionFn\":{\"type\":\"timeFormat\",\"format\":\"M\",\"timeZone\":\"UTC\",\"locale\":\"en-US\"}}],\"limitSpec\":{\"type\":\"default\"},\"aggregations\":[{\"type\":\"count\",\"name\":\"$f\"}],\"intervals\":[\"1900-01-01T00:00:00.000/3000-01-01T00:00:00.000\"]}", v, v);
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesQuery(java.lang.String,java.lang.String,java.lang.String,java.lang.String,java.lang.String)>("wikipedia", "scan", "{   \"queryType\": \"scan\",   \"dataSource\": \"wikipedia\",   \"descending\": \"false\",   \"columns\":[\"robot\",\"namespace\",\"anonymous\",\"unpatrolled\",\"page\",\"language\",\"newpage\",\"user\",\"count\",\"added\",\"delta\",\"variation\",\"deleted\"],   \"granularity\": \"all\",   \"intervals\": [     \"2013-01-01/2013-01-02\"   ], \"resultFormat\": \"compactedList\", \"limit\": 5}", "__time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted", "timestamp with local time zone,boolean,string,string,string,string,string,string,string,double,double,float,float,float");
virtualinvoke v.<org.apache.hadoop.hive.druid.QTestDruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: byte[] scanQueryResults>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] SCAN_QUERY_RESULTS_RECORDS>;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.String,java.lang.String,byte[],java.lang.Object[][])>(v, "scan", "{   \"queryType\": \"scan\",   \"dataSource\": \"wikipedia\",   \"descending\": \"false\",   \"columns\":[\"robot\",\"namespace\",\"anonymous\",\"unpatrolled\",\"page\",\"language\",\"newpage\",\"user\",\"count\",\"added\",\"delta\",\"variation\",\"deleted\"],   \"granularity\": \"all\",   \"intervals\": [     \"2013-01-01/2013-01-02\"   ], \"resultFormat\": \"compactedList\", \"limit\": 5}", v, v);
return;
}
private static java.util.Properties createPropertiesQuery(java.lang.String, java.lang.String, java.lang.String, java.lang.String, java.lang.String)
{
java.util.Properties v;
java.lang.String v, v, v, v, v;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v = new java.util.Properties;
specialinvoke v.<java.util.Properties: void <init>()>();
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("druid.datasource", v);
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("druid.query.json", v);
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("druid.query.type", v);
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("druid.fieldNames", v);
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("druid.fieldTypes", v);
return v;
}
private void deserializeQueryResults(org.apache.hadoop.hive.druid.serde.DruidSerDe, java.lang.String, java.lang.String, byte[], java.lang.Object[][]) throws org.apache.hadoop.hive.serde.SerDeException, java.io.IOException, java.lang.NoSuchFieldException, java.lang.SecurityException, java.lang.IllegalArgumentException, java.lang.IllegalAccessException, java.lang.InterruptedException, java.lang.NoSuchMethodException, java.lang.reflect.InvocationTargetException
{
byte[] v;
org.apache.hadoop.io.NullWritable v;
org.apache.hadoop.conf.Configuration v;
org.apache.hadoop.hive.serde.objectinspector.ObjectInspector v;
org.apache.hadoop.hive.druid.serde.DruidSerDe v;
java.lang.Object[] v, v;
java.lang.String v, v, v, v, v, v;
org.mockito.stubbing.OngoingStubbing v, v;
org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper v, v, v, v;
java.io.ByteArrayInputStream v, v;
org.apache.hadoop.fs.Path v;
boolean v, v;
java.util.List v;
org.apache.hadoop.hive.druid.io.HiveDruidSplit v;
org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader v, v;
java.lang.String[] v;
java.lang.Object[][] v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
int v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hive.druid.com.google.common.util.concurrent.ListenableFuture v, v;
org.apache.hadoop.hive.druid.serde.DruidWritable v, v;
org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture v, v;
java.lang.Class v, v, v, v;
java.lang.Object v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
v := @parameter: org.apache.hadoop.hive.druid.serde.DruidSerDe;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v := @parameter: byte[];
v := @parameter: java.lang.Object[][];
v = staticinvoke <org.mockito.Mockito: java.lang.Object mock(java.lang.Class)>(class "Lorg/apache/hive/druid/org/apache/druid/java/util/http/client/HttpClient;");
v = staticinvoke <org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture: org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture create()>();
v = new java.io.ByteArrayInputStream;
specialinvoke v.<java.io.ByteArrayInputStream: void <init>(byte[])>(v);
virtualinvoke v.<org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture: boolean set(java.lang.Object)>(v);
v = staticinvoke <org.mockito.ArgumentMatchers: java.lang.Object anyObject()>();
v = staticinvoke <org.mockito.ArgumentMatchers: java.lang.Object any(java.lang.Class)>(class "Lorg/apache/hive/druid/org/apache/druid/java/util/http/client/response/HttpResponseHandler;");
v = interfaceinvoke v.<org.apache.hive.druid.org.apache.druid.java.util.http.client.HttpClient: org.apache.hive.druid.com.google.common.util.concurrent.ListenableFuture go(org.apache.hive.druid.org.apache.druid.java.util.http.client.Request,org.apache.hive.druid.org.apache.druid.java.util.http.client.response.HttpResponseHandler)>(v, v);
v = staticinvoke <org.mockito.Mockito: org.mockito.stubbing.OngoingStubbing when(java.lang.Object)>(v);
interfaceinvoke v.<org.mockito.stubbing.OngoingStubbing: org.mockito.stubbing.OngoingStubbing thenReturn(java.lang.Object)>(v);
v = staticinvoke <org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat: org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader getDruidQueryReader(java.lang.String)>(v);
v = new org.apache.hadoop.hive.druid.io.HiveDruidSplit;
v = new org.apache.hadoop.fs.Path;
specialinvoke v.<org.apache.hadoop.fs.Path: void <init>(java.lang.String)>("empty");
v = newarray (java.lang.String)[1];
v[0] = "testing_host";
specialinvoke v.<org.apache.hadoop.hive.druid.io.HiveDruidSplit: void <init>(java.lang.String,org.apache.hadoop.fs.Path,java.lang.String[])>(v, v, v);
v = new org.apache.hadoop.conf.Configuration;
specialinvoke v.<org.apache.hadoop.conf.Configuration: void <init>()>();
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader: void initialize(org.apache.hadoop.mapreduce.InputSplit,org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper,org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper,org.apache.hive.druid.org.apache.druid.java.util.http.client.HttpClient,org.apache.hadoop.conf.Configuration)>(v, v, v, v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: org.apache.hadoop.hive.serde.objectinspector.ObjectInspector getObjectInspector()>();
v = virtualinvoke v.<org.apache.hadoop.hive.serde.objectinspector.StructObjectInspector: java.util.List getAllStructFieldRefs()>();
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader: org.apache.hadoop.hive.druid.serde.DruidWritable createValue()>();
v = 0;
label:
v = staticinvoke <org.apache.hadoop.io.NullWritable: org.apache.hadoop.io.NullWritable get()>();
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader: boolean next(org.apache.hadoop.io.NullWritable,org.apache.hadoop.hive.druid.serde.DruidWritable)>(v, v);
if v == 0 goto label;
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: java.lang.Object deserialize(org.apache.hadoop.io.Writable)>(v);
v = v[v];
v = lengthof v;
v = interfaceinvoke v.<java.util.List: int size()>();
staticinvoke <org.junit.Assert: void assertEquals(long,long)>(v, v);
v = 0;
label:
v = interfaceinvoke v.<java.util.List: int size()>();
if v >= v goto label;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Field \u type");
v = v[v];
v = virtualinvoke v.<java.lang.Object: java.lang.Class getClass()>();
v = interfaceinvoke v.<java.util.List: java.lang.Object get(int)>(v);
v = virtualinvoke v.<java.lang.Object: java.lang.Class getClass()>();
staticinvoke <org.junit.Assert: void assertEquals(java.lang.String,java.lang.Object,java.lang.Object)>(v, v, v);
v = interfaceinvoke v.<java.util.List: java.lang.Object get(int)>(v);
v = virtualinvoke v.<org.apache.hadoop.hive.serde.objectinspector.StructObjectInspector: java.lang.Object getStructFieldData(java.lang.Object,org.apache.hadoop.hive.serde.objectinspector.StructField)>(v, v);
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Field \u0001");
v = v[v];
staticinvoke <org.junit.Assert: void assertEquals(java.lang.String,java.lang.Object,java.lang.Object)>(v, v, v);
v = v + 1;
goto label;
label:
v = v + 1;
goto label;
label:
v = lengthof v;
staticinvoke <org.junit.Assert: void assertEquals(long,long)>(v, v);
v = staticinvoke <org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture: org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture create()>();
v = new java.io.ByteArrayInputStream;
specialinvoke v.<java.io.ByteArrayInputStream: void <init>(byte[])>(v);
virtualinvoke v.<org.apache.hive.druid.com.google.common.util.concurrent.SettableFuture: boolean set(java.lang.Object)>(v);
v = staticinvoke <org.mockito.ArgumentMatchers: java.lang.Object anyObject()>();
v = staticinvoke <org.mockito.ArgumentMatchers: java.lang.Object any(java.lang.Class)>(class "Lorg/apache/hive/druid/org/apache/druid/java/util/http/client/response/HttpResponseHandler;");
v = interfaceinvoke v.<org.apache.hive.druid.org.apache.druid.java.util.http.client.HttpClient: org.apache.hive.druid.com.google.common.util.concurrent.ListenableFuture go(org.apache.hive.druid.org.apache.druid.java.util.http.client.Request,org.apache.hive.druid.org.apache.druid.java.util.http.client.response.HttpResponseHandler)>(v, v);
v = staticinvoke <org.mockito.Mockito: org.mockito.stubbing.OngoingStubbing when(java.lang.Object)>(v);
interfaceinvoke v.<org.mockito.stubbing.OngoingStubbing: org.mockito.stubbing.OngoingStubbing thenReturn(java.lang.Object)>(v);
v = staticinvoke <org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat: org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader getDruidQueryReader(java.lang.String)>(v);
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper JSON_MAPPER>;
v = <org.apache.hadoop.hive.druid.DruidStorageHandlerUtils: org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper SMILE_MAPPER>;
virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader: void initialize(org.apache.hadoop.mapreduce.InputSplit,org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper,org.apache.hive.druid.com.fasterxml.jackson.databind.ObjectMapper,org.apache.hive.druid.org.apache.druid.java.util.http.client.HttpClient,org.apache.hadoop.conf.Configuration)>(v, v, v, v, v);
v = 0;
label:
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader: boolean nextKeyValue()>();
if v == 0 goto label;
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidQueryRecordReader: org.apache.hadoop.hive.druid.serde.DruidWritable getCurrentValue()>();
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: java.lang.Object deserialize(org.apache.hadoop.io.Writable)>(v);
v = v[v];
v = lengthof v;
v = interfaceinvoke v.<java.util.List: int size()>();
staticinvoke <org.junit.Assert: void assertEquals(long,long)>(v, v);
v = 0;
label:
v = interfaceinvoke v.<java.util.List: int size()>();
if v >= v goto label;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Field \u type");
v = v[v];
v = virtualinvoke v.<java.lang.Object: java.lang.Class getClass()>();
v = interfaceinvoke v.<java.util.List: java.lang.Object get(int)>(v);
v = virtualinvoke v.<java.lang.Object: java.lang.Class getClass()>();
staticinvoke <org.junit.Assert: void assertEquals(java.lang.String,java.lang.Object,java.lang.Object)>(v, v, v);
v = interfaceinvoke v.<java.util.List: java.lang.Object get(int)>(v);
v = virtualinvoke v.<org.apache.hadoop.hive.serde.objectinspector.StructObjectInspector: java.lang.Object getStructFieldData(java.lang.Object,org.apache.hadoop.hive.serde.objectinspector.StructField)>(v, v);
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Field \u0001");
v = v[v];
staticinvoke <org.junit.Assert: void assertEquals(java.lang.String,java.lang.Object,java.lang.Object)>(v, v, v);
v = v + 1;
goto label;
label:
v = v + 1;
goto label;
label:
v = lengthof v;
staticinvoke <org.junit.Assert: void assertEquals(long,long)>(v, v);
return;
}
public void testDruidObjectSerializer() throws org.apache.hadoop.hive.serde.SerDeException, java.lang.NoSuchFieldException, java.lang.SecurityException, java.lang.IllegalArgumentException, java.lang.IllegalAccessException, java.io.IOException, java.lang.InterruptedException, java.lang.NoSuchMethodException, java.lang.reflect.InvocationTargetException
{
java.lang.Object[] v;
java.util.Properties v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
org.apache.hadoop.hive.druid.serde.DruidWritable v;
org.apache.hadoop.conf.Configuration v;
org.apache.hadoop.hive.druid.serde.DruidSerDe v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
v = new org.apache.hadoop.hive.druid.serde.DruidSerDe;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: void <init>()>();
v = new org.apache.hadoop.conf.Configuration;
specialinvoke v.<org.apache.hadoop.conf.Configuration: void <init>()>();
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesSource(java.lang.String,java.lang.String)>("__time,c,c,c,c,c,c,c,c,c8", "timestamp with local time zone,string,char(6),varchar(8),double,float,bigint,int,smallint,tinyint");
virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[] ROW_OBJECT>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE>;
staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void serializeObject(java.util.Properties,org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.Object[],org.apache.hadoop.hive.druid.serde.DruidWritable)>(v, v, v, v);
return;
}
public void testDruidObjectSerializerwithNullTimestamp() throws java.lang.Exception
{
org.apache.hadoop.hive.serde.io.ByteWritable v;
java.lang.Object[] v;
org.apache.hadoop.io.Text v;
org.apache.hadoop.io.LongWritable v;
org.apache.hadoop.hive.serde.io.HiveCharWritable v;
org.apache.hadoop.hive.common.type.HiveChar v;
org.apache.hadoop.hive.common.type.HiveVarchar v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
org.apache.hadoop.conf.Configuration v;
org.apache.hadoop.hive.serde.io.DoubleWritable v;
org.apache.hadoop.hive.serde.io.ShortWritable v;
org.junit.rules.ExpectedException v, v;
org.apache.hadoop.io.IntWritable v;
org.apache.hadoop.hive.druid.serde.DruidSerDe v;
java.util.Properties v;
org.apache.hadoop.hive.serde.io.HiveVarcharWritable v;
org.apache.hadoop.hive.druid.serde.DruidWritable v;
org.apache.hadoop.io.FloatWritable v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
v = new org.apache.hadoop.hive.druid.serde.DruidSerDe;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: void <init>()>();
v = new org.apache.hadoop.conf.Configuration;
specialinvoke v.<org.apache.hadoop.conf.Configuration: void <init>()>();
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesSource(java.lang.String,java.lang.String)>("__time,c,c,c,c,c,c,c,c,c8", "timestamp with local time zone,string,char(6),varchar(8),double,float,bigint,int,smallint,tinyint");
virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = newarray (java.lang.Object)[11];
v[0] = null;
v = new org.apache.hadoop.io.Text;
specialinvoke v.<org.apache.hadoop.io.Text: void <init>(java.lang.String)>("dim1_val");
v[1] = v;
v = new org.apache.hadoop.hive.serde.io.HiveCharWritable;
v = new org.apache.hadoop.hive.common.type.HiveChar;
specialinvoke v.<org.apache.hadoop.hive.common.type.HiveChar: void <init>(java.lang.String,int)>("dim2_v", 6);
specialinvoke v.<org.apache.hadoop.hive.serde.io.HiveCharWritable: void <init>(org.apache.hadoop.hive.common.type.HiveChar)>(v);
v[2] = v;
v = new org.apache.hadoop.hive.serde.io.HiveVarcharWritable;
v = new org.apache.hadoop.hive.common.type.HiveVarchar;
specialinvoke v.<org.apache.hadoop.hive.common.type.HiveVarchar: void <init>(java.lang.String,int)>("dim3_val", 8);
specialinvoke v.<org.apache.hadoop.hive.serde.io.HiveVarcharWritable: void <init>(org.apache.hadoop.hive.common.type.HiveVarchar)>(v);
v[3] = v;
v = new org.apache.hadoop.hive.serde.io.DoubleWritable;
specialinvoke v.<org.apache.hadoop.hive.serde.io.DoubleWritable: void <init>(double)>(10669.3);
v[4] = v;
v = new org.apache.hadoop.io.FloatWritable;
specialinvoke v.<org.apache.hadoop.io.FloatWritable: void <init>(float)>(10669.45F);
v[5] = v;
v = new org.apache.hadoop.io.LongWritable;
specialinvoke v.<org.apache.hadoop.io.LongWritable: void <init>(long)>(1113939L);
v[6] = v;
v = new org.apache.hadoop.io.IntWritable;
specialinvoke v.<org.apache.hadoop.io.IntWritable: void <init>(int)>(1112123);
v[7] = v;
v = new org.apache.hadoop.hive.serde.io.ShortWritable;
specialinvoke v.<org.apache.hadoop.hive.serde.io.ShortWritable: void <init>(short)>(12);
v[8] = v;
v = new org.apache.hadoop.hive.serde.io.ByteWritable;
specialinvoke v.<org.apache.hadoop.hive.serde.io.ByteWritable: void <init>(byte)>(0);
v[9] = v;
v[10] = null;
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.junit.rules.ExpectedException expectedEx>;
virtualinvoke v.<org.junit.rules.ExpectedException: void expect(java.lang.Class)>(class "Ljava/lang/NullPointerException;");
v = v.<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.junit.rules.ExpectedException expectedEx>;
virtualinvoke v.<org.junit.rules.ExpectedException: void expectMessage(java.lang.String)>("Timestamp column cannot have null value");
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE>;
staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void serializeObject(java.util.Properties,org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.Object[],org.apache.hadoop.hive.druid.serde.DruidWritable)>(v, v, v, v);
return;
}
private static java.util.Properties createPropertiesSource(java.lang.String, java.lang.String)
{
java.lang.String v, v;
java.util.Properties v;
v := @parameter: java.lang.String;
v := @parameter: java.lang.String;
v = new java.util.Properties;
specialinvoke v.<java.util.Properties: void <init>()>();
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("columns", v);
virtualinvoke v.<java.util.Properties: java.lang.Object setProperty(java.lang.String,java.lang.String)>("columns.types", v);
return v;
}
private static void serializeObject(java.util.Properties, org.apache.hadoop.hive.druid.serde.DruidSerDe, java.lang.Object[], org.apache.hadoop.hive.druid.serde.DruidWritable) throws org.apache.hadoop.hive.serde.SerDeException
{
org.apache.hadoop.io.Writable v;
org.apache.hadoop.hive.serde.objectinspector.StandardStructObjectInspector v;
java.util.Map v, v, v, v;
java.util.stream.Collector v, v;
org.apache.hadoop.hive.serde.typeinfo.PrimitiveTypeInfo v;
boolean v;
org.apache.hadoop.hive.druid.serde.DruidSerDe v;
java.util.Set v;
java.util.List v, v;
java.util.stream.Stream v, v, v, v;
java.lang.Object[] v;
java.util.function.Function v, v;
java.util.ArrayList v;
int v, v;
java.util.Properties v;
java.util.Iterator v;
org.apache.hadoop.hive.druid.serde.DruidWritable v;
java.lang.Object v, v, v, v, v, v;
v := @parameter: java.util.Properties;
v := @parameter: org.apache.hadoop.hive.druid.serde.DruidSerDe;
v := @parameter: java.lang.Object[];
v := @parameter: org.apache.hadoop.hive.druid.serde.DruidWritable;
v = new java.util.ArrayList;
v = staticinvoke <org.apache.hadoop.hive.ql.exec.Utilities: java.util.List getColumnNames(java.util.Properties)>(v);
specialinvoke v.<java.util.ArrayList: void <init>(java.util.Collection)>(v);
interfaceinvoke v.<java.util.List: boolean add(java.lang.Object)>("__time_granularity");
v = staticinvoke <org.apache.hadoop.hive.ql.exec.Utilities: java.util.List getColumnTypes(java.util.Properties)>(v);
v = interfaceinvoke v.<java.util.List: java.util.stream.Stream stream()>();
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe$getPrimitiveTypeInfo__7: java.util.function.Function bootstrap$()>();
v = interfaceinvoke v.<java.util.stream.Stream: java.util.stream.Stream map(java.util.function.Function)>(v);
v = staticinvoke <java.util.stream.Collectors: java.util.stream.Collector toList()>();
v = interfaceinvoke v.<java.util.stream.Stream: java.lang.Object collect(java.util.stream.Collector)>(v);
v = staticinvoke <org.apache.hadoop.hive.serde.typeinfo.TypeInfoFactory: org.apache.hadoop.hive.serde.typeinfo.PrimitiveTypeInfo getPrimitiveTypeInfo(java.lang.String)>("timestamp");
interfaceinvoke v.<java.util.List: boolean add(java.lang.Object)>(v);
v = interfaceinvoke v.<java.util.List: java.util.stream.Stream stream()>();
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe$getPrimitiveWritableObjectInspector__8: java.util.function.Function bootstrap$()>();
v = interfaceinvoke v.<java.util.stream.Stream: java.util.stream.Stream map(java.util.function.Function)>(v);
v = staticinvoke <java.util.stream.Collectors: java.util.stream.Collector toList()>();
v = interfaceinvoke v.<java.util.stream.Stream: java.lang.Object collect(java.util.stream.Collector)>(v);
v = staticinvoke <org.apache.hadoop.hive.serde.objectinspector.ObjectInspectorFactory: org.apache.hadoop.hive.serde.objectinspector.StandardStructObjectInspector getStandardStructObjectInspector(java.util.List,java.util.List)>(v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: org.apache.hadoop.io.Writable serialize(java.lang.Object,org.apache.hadoop.hive.serde.objectinspector.ObjectInspector)>(v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidWritable: java.util.Map getValue()>();
v = interfaceinvoke v.<java.util.Map: int size()>();
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidWritable: java.util.Map getValue()>();
v = interfaceinvoke v.<java.util.Map: int size()>();
staticinvoke <org.junit.Assert: void assertEquals(long,long)>(v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidWritable: java.util.Map getValue()>();
v = interfaceinvoke v.<java.util.Map: java.util.Set entrySet()>();
v = interfaceinvoke v.<java.util.Set: java.util.Iterator iterator()>();
label:
v = interfaceinvoke v.<java.util.Iterator: boolean hasNext()>();
if v == 0 goto label;
v = interfaceinvoke v.<java.util.Iterator: java.lang.Object next()>();
v = interfaceinvoke v.<java.util.Map$Entry: java.lang.Object getValue()>();
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidWritable: java.util.Map getValue()>();
v = interfaceinvoke v.<java.util.Map$Entry: java.lang.Object getKey()>();
v = interfaceinvoke v.<java.util.Map: java.lang.Object get(java.lang.Object)>(v);
staticinvoke <org.junit.Assert: void assertEquals(java.lang.Object,java.lang.Object)>(v, v);
goto label;
label:
return;
}
public void testDruidObjectDeserializer() throws org.apache.hadoop.hive.serde.SerDeException, java.lang.NoSuchFieldException, java.lang.SecurityException, java.lang.IllegalArgumentException, java.lang.IllegalAccessException, java.io.IOException, java.lang.InterruptedException, java.lang.NoSuchMethodException, java.lang.reflect.InvocationTargetException
{
java.lang.Object[] v;
java.util.Properties v;
org.apache.hadoop.hive.druid.serde.TestDruidSerDe v;
org.apache.hadoop.hive.druid.serde.DruidWritable v;
org.apache.hadoop.conf.Configuration v;
org.apache.hadoop.hive.druid.serde.DruidSerDe v;
v := @this: org.apache.hadoop.hive.druid.serde.TestDruidSerDe;
v = new org.apache.hadoop.hive.druid.serde.DruidSerDe;
specialinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: void <init>()>();
v = new org.apache.hadoop.conf.Configuration;
specialinvoke v.<org.apache.hadoop.conf.Configuration: void <init>()>();
v = staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.util.Properties createPropertiesSource(java.lang.String,java.lang.String)>("__time,c,c,c,c,c,c,c,c,c8", "timestamp with local time zone,string,char(6),varchar(8),double,float,bigint,int,smallint,tinyint");
virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: void initialize(org.apache.hadoop.conf.Configuration,java.util.Properties,java.util.Properties)>(v, v, null);
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[] ROW_OBJECT_2>;
v = <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE_2>;
staticinvoke <org.apache.hadoop.hive.druid.serde.TestDruidSerDe: void deserializeObject(org.apache.hadoop.hive.druid.serde.DruidSerDe,java.lang.Object[],org.apache.hadoop.hive.druid.serde.DruidWritable)>(v, v, v);
return;
}
private static void deserializeObject(org.apache.hadoop.hive.druid.serde.DruidSerDe, java.lang.Object[], org.apache.hadoop.hive.druid.serde.DruidWritable) throws org.apache.hadoop.hive.serde.SerDeException
{
java.lang.Object[] v;
org.apache.hadoop.hive.druid.serde.DruidWritable v;
int v, v, v, v;
java.lang.Class v, v;
java.lang.Object v, v, v, v, v;
org.apache.hadoop.hive.druid.serde.DruidSerDe v;
v := @parameter: org.apache.hadoop.hive.druid.serde.DruidSerDe;
v := @parameter: java.lang.Object[];
v := @parameter: org.apache.hadoop.hive.druid.serde.DruidWritable;
v = virtualinvoke v.<org.apache.hadoop.hive.druid.serde.DruidSerDe: java.lang.Object deserialize(org.apache.hadoop.io.Writable)>(v);
v = lengthof v;
v = interfaceinvoke v.<java.util.List: int size()>();
staticinvoke <org.junit.Assert: void assertEquals(long,long)>(v, v);
v = 0;
label:
v = lengthof v;
if v >= v goto label;
v = v[v];
v = virtualinvoke v.<java.lang.Object: java.lang.Class getClass()>();
v = interfaceinvoke v.<java.util.List: java.lang.Object get(int)>(v);
v = virtualinvoke v.<java.lang.Object: java.lang.Class getClass()>();
staticinvoke <org.junit.Assert: void assertEquals(java.lang.Object,java.lang.Object)>(v, v);
v = v[v];
v = interfaceinvoke v.<java.util.List: java.lang.Object get(int)>(v);
staticinvoke <org.junit.Assert: void assertEquals(java.lang.Object,java.lang.Object)>(v, v);
v = v + 1;
goto label;
label:
return;
}
static void <clinit>()
{
java.time.ZonedDateTime v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
java.lang.Integer v, v, v, v, v;
java.lang.Long v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
java.time.ZoneOffset v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.io.IntWritable v;
java.lang.Object[] v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.hive.serde.io.HiveCharWritable v;
org.apache.hadoop.io.FloatWritable v;
org.apache.hadoop.io.LongWritable v;
java.lang.Short v, v, v;
org.apache.hive.druid.com.google.common.collect.ImmutableMap v, v;
org.apache.hadoop.hive.serde.io.HiveVarcharWritable v;
java.time.Instant v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.hive.common.type.HiveVarchar v, v;
java.lang.Boolean v, v, v, v, v;
java.lang.Float v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.io.Text v;
org.apache.hadoop.hive.common.type.HiveChar v, v;
java.lang.Double v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.hive.serde.io.DoubleWritable v;
org.apache.hadoop.hive.common.type.TimestampTZ v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.hive.serde.io.TimestampWritableV2 v;
java.lang.Byte v, v, v;
org.apache.hadoop.hive.serde.io.ShortWritable v;
org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v, v;
org.apache.hadoop.hive.serde.io.TimestampLocalTZWritable v;
org.apache.hadoop.hive.serde.io.ByteWritable v;
java.lang.Object[][] v, v, v, v, v, v;
org.apache.hadoop.hive.common.type.Timestamp v;
org.apache.hadoop.hive.druid.serde.DruidWritable v, v;
v = newarray (java.lang.Object[])[2];
v = newarray (java.lang.Object)[4];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376000000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(0L);
v[1] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(1.0F);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(2.2222F);
v[3] = v;
v[0] = v;
v = newarray (java.lang.Object)[4];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325462400000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(2L);
v[1] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(3.32F);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(4.0F);
v[3] = v;
v[1] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] TIMESERIES_QUERY_RESULTS_RECORDS> = v;
v = newarray (java.lang.Object[])[5];
v = newarray (java.lang.Object)[5];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "dim1_val";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(111L);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(10669.0F);
v[3] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(96.11712F);
v[4] = v;
v[0] = v;
v = newarray (java.lang.Object)[5];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "another_dim1_val";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(88L);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(28344.0F);
v[3] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(322.0909F);
v[4] = v;
v[1] = v;
v = newarray (java.lang.Object)[5];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "dim1_val3";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(70L);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(871.0F);
v[3] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(12.442857F);
v[4] = v;
v[2] = v;
v = newarray (java.lang.Object)[5];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "dim1_val4";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(62L);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(815.0F);
v[3] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(13.145162F);
v[4] = v;
v[3] = v;
v = newarray (java.lang.Object)[5];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "dim1_val5";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(60L);
v[2] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(2787.0F);
v[3] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(46.45F);
v[4] = v;
v[4] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] TOPN_QUERY_RESULTS_RECORDS> = v;
v = newarray (java.lang.Object[])[2];
v = newarray (java.lang.Object)[3];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376000000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376000000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[1] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(200L);
v[2] = v;
v[0] = v;
v = newarray (java.lang.Object)[3];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376012000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376012000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[1] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(400L);
v[2] = v;
v[1] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] GROUP_BY_QUERY_EXTRACTION_RESULTS_RECORDS> = v;
v = newarray (java.lang.Object[])[2];
v = newarray (java.lang.Object)[6];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376000000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "India";
v[2] = "phone";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(88L);
v[3] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(29.91233453);
v[4] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(60.32F);
v[5] = v;
v[0] = v;
v = newarray (java.lang.Object)[6];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376012000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "Spain";
v[2] = "pc";
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(16L);
v[3] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(172.93494959);
v[4] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(6.333333F);
v[5] = v;
v[1] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] GROUP_BY_QUERY_RESULTS_RECORDS> = v;
v = newarray (java.lang.Object[])[2];
v = newarray (java.lang.Object)[3];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376000000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(1);
v[1] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(200L);
v[2] = v;
v[0] = v;
v = newarray (java.lang.Object)[3];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1325376012000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(1);
v[1] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(400L);
v[2] = v;
v[1] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] GB_MONTH_EXTRACTION_RESULTS_RECORDS> = v;
v = newarray (java.lang.Object[])[5];
v = newarray (java.lang.Object)[14];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1356998400000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = <java.lang.Boolean: java.lang.Boolean TRUE>;
v[1] = v;
v[2] = "article";
v[3] = "0";
v[4] = "0";
v[5] = "11._korpus_(NOVJ)";
v[6] = "sl";
v[7] = "0";
v[8] = "EmausBot";
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(1.0);
v[9] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(39.0);
v[10] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(39.0F);
v[11] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(39.0F);
v[12] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(0.0F);
v[13] = v;
v[0] = v;
v = newarray (java.lang.Object)[14];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1356998400000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = <java.lang.Boolean: java.lang.Boolean FALSE>;
v[1] = v;
v[2] = "article";
v[3] = "0";
v[4] = "0";
v[5] = "112_U.S._580";
v[6] = "en";
v[7] = "1";
v[8] = "MZMcBride";
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(1.0);
v[9] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(70.0);
v[10] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(70.0F);
v[11] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(70.0F);
v[12] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(0.0F);
v[13] = v;
v[1] = v;
v = newarray (java.lang.Object)[14];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1356998412000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = <java.lang.Boolean: java.lang.Boolean FALSE>;
v[1] = v;
v[2] = "article";
v[3] = "0";
v[4] = "0";
v[5] = "113_U.S._243";
v[6] = "en";
v[7] = "1";
v[8] = "MZMcBride";
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(1.0);
v[9] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(77.0);
v[10] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(77.0F);
v[11] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(77.0F);
v[12] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(0.0F);
v[13] = v;
v[2] = v;
v = newarray (java.lang.Object)[14];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1356998412000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = <java.lang.Boolean: java.lang.Boolean FALSE>;
v[1] = v;
v[2] = "article";
v[3] = "0";
v[4] = "0";
v[5] = "113_U.S._73";
v[6] = "en";
v[7] = "1";
v[8] = "MZMcBride";
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(1.0);
v[9] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(70.0);
v[10] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(70.0F);
v[11] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(70.0F);
v[12] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(0.0F);
v[13] = v;
v[3] = v;
v = newarray (java.lang.Object)[14];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1356998412000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v = <java.lang.Boolean: java.lang.Boolean FALSE>;
v[1] = v;
v[2] = "article";
v[3] = "0";
v[4] = "0";
v[5] = "113_U.S._756";
v[6] = "en";
v[7] = "1";
v[8] = "MZMcBride";
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(1.0);
v[9] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(68.0);
v[10] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(68.0F);
v[11] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(68.0F);
v[12] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(0.0F);
v[13] = v;
v[4] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[][] SCAN_QUERY_RESULTS_RECORDS> = v;
v = newarray (java.lang.Object)[11];
v = new org.apache.hadoop.hive.serde.io.TimestampLocalTZWritable;
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
specialinvoke v.<org.apache.hadoop.hive.serde.io.TimestampLocalTZWritable: void <init>(org.apache.hadoop.hive.common.type.TimestampTZ)>(v);
v[0] = v;
v = new org.apache.hadoop.io.Text;
specialinvoke v.<org.apache.hadoop.io.Text: void <init>(java.lang.String)>("dim1_val");
v[1] = v;
v = new org.apache.hadoop.hive.serde.io.HiveCharWritable;
v = new org.apache.hadoop.hive.common.type.HiveChar;
specialinvoke v.<org.apache.hadoop.hive.common.type.HiveChar: void <init>(java.lang.String,int)>("dim2_v", 6);
specialinvoke v.<org.apache.hadoop.hive.serde.io.HiveCharWritable: void <init>(org.apache.hadoop.hive.common.type.HiveChar)>(v);
v[2] = v;
v = new org.apache.hadoop.hive.serde.io.HiveVarcharWritable;
v = new org.apache.hadoop.hive.common.type.HiveVarchar;
specialinvoke v.<org.apache.hadoop.hive.common.type.HiveVarchar: void <init>(java.lang.String,int)>("dim3_val", 8);
specialinvoke v.<org.apache.hadoop.hive.serde.io.HiveVarcharWritable: void <init>(org.apache.hadoop.hive.common.type.HiveVarchar)>(v);
v[3] = v;
v = new org.apache.hadoop.hive.serde.io.DoubleWritable;
specialinvoke v.<org.apache.hadoop.hive.serde.io.DoubleWritable: void <init>(double)>(10669.3);
v[4] = v;
v = new org.apache.hadoop.io.FloatWritable;
specialinvoke v.<org.apache.hadoop.io.FloatWritable: void <init>(float)>(10669.45F);
v[5] = v;
v = new org.apache.hadoop.io.LongWritable;
specialinvoke v.<org.apache.hadoop.io.LongWritable: void <init>(long)>(1113939L);
v[6] = v;
v = new org.apache.hadoop.io.IntWritable;
specialinvoke v.<org.apache.hadoop.io.IntWritable: void <init>(int)>(1112123);
v[7] = v;
v = new org.apache.hadoop.hive.serde.io.ShortWritable;
specialinvoke v.<org.apache.hadoop.hive.serde.io.ShortWritable: void <init>(short)>(12);
v[8] = v;
v = new org.apache.hadoop.hive.serde.io.ByteWritable;
specialinvoke v.<org.apache.hadoop.hive.serde.io.ByteWritable: void <init>(byte)>(0);
v[9] = v;
v = new org.apache.hadoop.hive.serde.io.TimestampWritableV2;
v = staticinvoke <org.apache.hadoop.hive.common.type.Timestamp: org.apache.hadoop.hive.common.type.Timestamp ofEpochSecond(long)>(1377907200L);
specialinvoke v.<org.apache.hadoop.hive.serde.io.TimestampWritableV2: void <init>(org.apache.hadoop.hive.common.type.Timestamp)>(v);
v[10] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[] ROW_OBJECT> = v;
v = new org.apache.hadoop.hive.druid.serde.DruidWritable;
v = staticinvoke <org.apache.hive.druid.com.google.common.collect.ImmutableMap: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder builder()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(1377907200000L);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("__time", v);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c0", "dim1_val");
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c1", "dim2_v");
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c2", "dim3_val");
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(10669.3);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c3", v);
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(10669.45F);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c4", v);
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(1113939L);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c5", v);
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(1112123);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c6", v);
v = staticinvoke <java.lang.Short: java.lang.Short valueOf(short)>(12);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c7", v);
v = staticinvoke <java.lang.Byte: java.lang.Byte valueOf(byte)>(0);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c8", v);
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(1377907200000L);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("__time_granularity", v);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap build()>();
specialinvoke v.<org.apache.hadoop.hive.druid.serde.DruidWritable: void <init>(java.util.Map)>(v);
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE> = v;
v = newarray (java.lang.Object)[10];
v = new org.apache.hadoop.hive.common.type.TimestampTZ;
v = staticinvoke <java.time.Instant: java.time.Instant ofEpochMilli(long)>(1377907200000L);
v = <java.time.ZoneOffset: java.time.ZoneOffset UTC>;
v = virtualinvoke v.<java.time.Instant: java.time.ZonedDateTime atZone(java.time.ZoneId)>(v);
specialinvoke v.<org.apache.hadoop.hive.common.type.TimestampTZ: void <init>(java.time.ZonedDateTime)>(v);
v[0] = v;
v[1] = "dim1_val";
v = new org.apache.hadoop.hive.common.type.HiveChar;
specialinvoke v.<org.apache.hadoop.hive.common.type.HiveChar: void <init>(java.lang.String,int)>("dim2_v", 6);
v[2] = v;
v = new org.apache.hadoop.hive.common.type.HiveVarchar;
specialinvoke v.<org.apache.hadoop.hive.common.type.HiveVarchar: void <init>(java.lang.String,int)>("dim3_val", 8);
v[3] = v;
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(10669.3);
v[4] = v;
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(10669.45F);
v[5] = v;
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(1113939L);
v[6] = v;
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(1112123);
v[7] = v;
v = staticinvoke <java.lang.Short: java.lang.Short valueOf(short)>(12);
v[8] = v;
v = staticinvoke <java.lang.Byte: java.lang.Byte valueOf(byte)>(0);
v[9] = v;
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: java.lang.Object[] ROW_OBJECT_2> = v;
v = new org.apache.hadoop.hive.druid.serde.DruidWritable;
v = staticinvoke <org.apache.hive.druid.com.google.common.collect.ImmutableMap: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder builder()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(1377907200000L);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("__time", v);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c0", "dim1_val");
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c1", "dim2_v");
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c2", "dim3_val");
v = staticinvoke <java.lang.Double: java.lang.Double valueOf(double)>(10669.3);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c3", v);
v = staticinvoke <java.lang.Float: java.lang.Float valueOf(float)>(10669.45F);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c4", v);
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(1113939L);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c5", v);
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(1112123);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c6", v);
v = staticinvoke <java.lang.Short: java.lang.Short valueOf(short)>(12);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c7", v);
v = staticinvoke <java.lang.Byte: java.lang.Byte valueOf(byte)>(0);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder put(java.lang.Object,java.lang.Object)>("c8", v);
v = virtualinvoke v.<org.apache.hive.druid.com.google.common.collect.ImmutableMap$Builder: org.apache.hive.druid.com.google.common.collect.ImmutableMap build()>();
specialinvoke v.<org.apache.hadoop.hive.druid.serde.DruidWritable: void <init>(java.util.Map)>(v);
<org.apache.hadoop.hive.druid.serde.TestDruidSerDe: org.apache.hadoop.hive.druid.serde.DruidWritable DRUID_WRITABLE_2> = v;
return;
}
}