public class org.apache.hadoop.hive.druid.QTestDruidQueryBasedInputFormatToAddFaultyHost extends org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat
{
public void <init>()
{
org.apache.hadoop.hive.druid.QTestDruidQueryBasedInputFormatToAddFaultyHost v;
v := @this: org.apache.hadoop.hive.druid.QTestDruidQueryBasedInputFormatToAddFaultyHost;
specialinvoke v.<org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat: void <init>()>();
return;
}
protected org.apache.hadoop.hive.druid.io.HiveDruidSplit[] getInputSplits(org.apache.hadoop.conf.Configuration) throws java.io.IOException
{
org.apache.hadoop.hive.druid.io.HiveDruidSplit[] v, v;
java.lang.Object[] v, v;
java.lang.String[] v, v;
java.util.ArrayList v, v;
org.apache.hadoop.conf.Configuration v;
int v, v, v, v;
org.apache.hadoop.hive.druid.QTestDruidQueryBasedInputFormatToAddFaultyHost v;
org.apache.hadoop.fs.Path v;
java.lang.String v;
java.util.List v;
org.apache.hadoop.hive.druid.io.HiveDruidSplit v, v;
v := @this: org.apache.hadoop.hive.druid.QTestDruidQueryBasedInputFormatToAddFaultyHost;
v := @parameter: org.apache.hadoop.conf.Configuration;
v = specialinvoke v.<org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat: org.apache.hadoop.hive.druid.io.HiveDruidSplit[] getInputSplits(org.apache.hadoop.conf.Configuration)>(v);
v = new java.util.ArrayList;
specialinvoke v.<java.util.ArrayList: void <init>()>();
v = lengthof v;
v = 0;
label:
if v >= v goto label;
v = v[v];
v = virtualinvoke v.<org.apache.hadoop.hive.druid.io.HiveDruidSplit: java.lang.String[] getLocations()>();
v = lengthof v;
v = v + 1;
v = staticinvoke <org.apache.hive.druid.com.google.common.collect.Lists: java.util.ArrayList newArrayListWithCapacity(int)>(v);
interfaceinvoke v.<java.util.List: boolean add(java.lang.Object)>("localhost:8081");
v = staticinvoke <java.util.Arrays: java.util.List asList(java.lang.Object[])>(v);
interfaceinvoke v.<java.util.List: boolean addAll(java.util.Collection)>(v);
v = new org.apache.hadoop.hive.druid.io.HiveDruidSplit;
v = virtualinvoke v.<org.apache.hadoop.hive.druid.io.HiveDruidSplit: java.lang.String getDruidQuery()>();
v = virtualinvoke v.<org.apache.hadoop.hive.druid.io.HiveDruidSplit: org.apache.hadoop.fs.Path getPath()>();
v = newarray (java.lang.String)[0];
v = interfaceinvoke v.<java.util.List: java.lang.Object[] toArray(java.lang.Object[])>(v);
specialinvoke v.<org.apache.hadoop.hive.druid.io.HiveDruidSplit: void <init>(java.lang.String,org.apache.hadoop.fs.Path,java.lang.String[])>(v, v, v);
interfaceinvoke v.<java.util.List: boolean add(java.lang.Object)>(v);
v = v + 1;
goto label;
label:
v = newarray (org.apache.hadoop.hive.druid.io.HiveDruidSplit)[0];
v = interfaceinvoke v.<java.util.List: java.lang.Object[] toArray(java.lang.Object[])>(v);
return v;
}
}