public class org.apache.hadoop.hive.kafka.KafkaInputSplit extends org.apache.hadoop.mapred.FileSplit implements org.apache.hadoop.mapred.InputSplit
{
private java.lang.String topic;
private long startOffset;
private int partition;
private long endOffset;
static final boolean $assertionsDisabled;
public void <init>()
{
java.lang.String[] v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = (java.lang.String[]) null;
specialinvoke v.<org.apache.hadoop.mapred.FileSplit: void <init>(org.apache.hadoop.fs.Path,long,long,java.lang.String[])>(null, 0L, 0L, v);
return;
}
public void <init>(java.lang.String, int, long, long, org.apache.hadoop.fs.Path)
{
java.lang.String[] v;
long v, v;
byte v, v;
int v;
org.apache.hadoop.fs.Path v;
java.lang.String v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
boolean v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v := @parameter: java.lang.String;
v := @parameter: int;
v := @parameter: long;
v := @parameter: long;
v := @parameter: org.apache.hadoop.fs.Path;
v = (java.lang.String[]) null;
specialinvoke v.<org.apache.hadoop.mapred.FileSplit: void <init>(org.apache.hadoop.fs.Path,long,long,java.lang.String[])>(v, 0L, 0L, v);
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic> = v;
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset> = v;
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition> = v;
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset> = v;
v = v cmp 0L;
if v < 0 goto label;
v = v cmp v;
if v > 0 goto label;
v = 1;
goto label;
label:
v = 0;
label:
staticinvoke <com.google.common.base.Preconditions: void checkArgument(boolean,java.lang.String,long,long)>(v, "start [%s] has to be positive and >= end [%]", v, v);
return;
}
public long getLength()
{
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
return 0L;
}
public java.lang.String[] getLocations()
{
java.lang.String[] v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = newarray (java.lang.String)[0];
return v;
}
public void write(java.io.DataOutput) throws java.io.IOException
{
java.io.DataOutput v;
int v;
long v, v;
java.lang.String v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v := @parameter: java.io.DataOutput;
specialinvoke v.<org.apache.hadoop.mapred.FileSplit: void write(java.io.DataOutput)>(v);
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
interfaceinvoke v.<java.io.DataOutput: void writeUTF(java.lang.String)>(v);
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
interfaceinvoke v.<java.io.DataOutput: void writeInt(int)>(v);
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset>;
interfaceinvoke v.<java.io.DataOutput: void writeLong(long)>(v);
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset>;
interfaceinvoke v.<java.io.DataOutput: void writeLong(long)>(v);
return;
}
public void readFields(java.io.DataInput) throws java.io.IOException
{
long v, v, v, v, v, v, v;
byte v, v;
int v;
java.io.DataInput v;
java.lang.String v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
boolean v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v := @parameter: java.io.DataInput;
specialinvoke v.<org.apache.hadoop.mapred.FileSplit: void readFields(java.io.DataInput)>(v);
v = interfaceinvoke v.<java.io.DataInput: java.lang.String readUTF()>();
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic> = v;
v = interfaceinvoke v.<java.io.DataInput: int readInt()>();
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition> = v;
v = interfaceinvoke v.<java.io.DataInput: long readLong()>();
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset> = v;
v = interfaceinvoke v.<java.io.DataInput: long readLong()>();
v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset> = v;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset>;
v = v cmp 0L;
if v < 0 goto label;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset>;
v = v cmp v;
if v > 0 goto label;
v = 1;
goto label;
label:
v = 0;
label:
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset>;
staticinvoke <com.google.common.base.Preconditions: void checkArgument(boolean,java.lang.String,long,long)>(v, "start [%s] has to be positive and >= end [%]", v, v);
return;
}
public java.lang.String getTopic()
{
java.lang.String v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
return v;
}
public int getPartition()
{
int v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
return v;
}
public long getStartOffset()
{
long v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset>;
return v;
}
public long getEndOffset()
{
long v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset>;
return v;
}
public static org.apache.hadoop.hive.kafka.KafkaInputSplit intersectRange(org.apache.hadoop.hive.kafka.KafkaInputSplit, org.apache.hadoop.hive.kafka.KafkaInputSplit)
{
long v, v, v, v, v, v;
java.lang.AssertionError v, v;
byte v;
int v, v, v;
org.apache.hadoop.fs.Path v;
java.lang.String v, v, v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v, v, v;
boolean v, v, v;
v := @parameter: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v := @parameter: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = <org.apache.hadoop.hive.kafka.KafkaInputSplit: boolean $assertionsDisabled>;
if v != 0 goto label;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = virtualinvoke v.<java.lang.String: boolean equals(java.lang.Object)>(v);
if v != 0 goto label;
v = new java.lang.AssertionError;
specialinvoke v.<java.lang.AssertionError: void <init>()>();
throw v;
label:
v = <org.apache.hadoop.hive.kafka.KafkaInputSplit: boolean $assertionsDisabled>;
if v != 0 goto label;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
if v == v goto label;
v = new java.lang.AssertionError;
specialinvoke v.<java.lang.AssertionError: void <init>()>();
throw v;
label:
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = staticinvoke <java.lang.Math: long max(long,long)>(v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = staticinvoke <java.lang.Math: long min(long,long)>(v, v);
v = v cmp v;
if v <= 0 goto label;
return null;
label:
v = new org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: org.apache.hadoop.fs.Path getPath()>();
specialinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: void <init>(java.lang.String,int,long,long,org.apache.hadoop.fs.Path)>(v, v, v, v, v);
return v;
}
public static org.apache.hadoop.hive.kafka.KafkaInputSplit unionRange(org.apache.hadoop.hive.kafka.KafkaInputSplit, org.apache.hadoop.hive.kafka.KafkaInputSplit)
{
long v, v, v, v, v, v;
java.lang.AssertionError v, v;
int v, v, v;
org.apache.hadoop.fs.Path v;
java.lang.String v, v, v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v, v, v;
boolean v, v, v;
v := @parameter: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v := @parameter: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = <org.apache.hadoop.hive.kafka.KafkaInputSplit: boolean $assertionsDisabled>;
if v != 0 goto label;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = virtualinvoke v.<java.lang.String: boolean equals(java.lang.Object)>(v);
if v != 0 goto label;
v = new java.lang.AssertionError;
specialinvoke v.<java.lang.AssertionError: void <init>()>();
throw v;
label:
v = <org.apache.hadoop.hive.kafka.KafkaInputSplit: boolean $assertionsDisabled>;
if v != 0 goto label;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
if v == v goto label;
v = new java.lang.AssertionError;
specialinvoke v.<java.lang.AssertionError: void <init>()>();
throw v;
label:
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = staticinvoke <java.lang.Math: long min(long,long)>(v, v);
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = staticinvoke <java.lang.Math: long max(long,long)>(v, v);
v = new org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: org.apache.hadoop.fs.Path getPath()>();
specialinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: void <init>(java.lang.String,int,long,long,org.apache.hadoop.fs.Path)>(v, v, v, v, v);
return v;
}
public boolean equals(java.lang.Object)
{
long v, v, v, v;
java.lang.Integer v, v;
java.lang.Long v, v, v, v;
int v, v;
java.lang.String v, v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
boolean v, v, v, v, v, v;
java.lang.Object v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v := @parameter: java.lang.Object;
if v != v goto label;
return 1;
label:
v = v instanceof org.apache.hadoop.hive.kafka.KafkaInputSplit;
if v != 0 goto label;
return 0;
label:
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String getTopic()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String getTopic()>();
v = staticinvoke <com.google.common.base.Objects: boolean equal(java.lang.Object,java.lang.Object)>(v, v);
if v == 0 goto label;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
v = staticinvoke <com.google.common.base.Objects: boolean equal(java.lang.Object,java.lang.Object)>(v, v);
if v == 0 goto label;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int getPartition()>();
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(v);
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int getPartition()>();
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(v);
v = staticinvoke <com.google.common.base.Objects: boolean equal(java.lang.Object,java.lang.Object)>(v, v);
if v == 0 goto label;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
v = staticinvoke <com.google.common.base.Objects: boolean equal(java.lang.Object,java.lang.Object)>(v, v);
if v == 0 goto label;
v = 1;
goto label;
label:
v = 0;
label:
return v;
}
public int hashCode()
{
java.lang.Object[] v;
long v, v;
java.lang.Integer v;
java.lang.Long v, v;
int v, v;
java.lang.String v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = newarray (java.lang.Object)[4];
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String getTopic()>();
v[0] = v;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
v[1] = v;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int getPartition()>();
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(v);
v[2] = v;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
v[3] = v;
v = staticinvoke <com.google.common.base.Objects: int hashCode(java.lang.Object[])>(v);
return v;
}
public java.lang.String toString()
{
long v, v;
int v;
org.apache.hadoop.fs.Path v;
java.lang.String v, v, v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v;
v := @this: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long startOffset>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long endOffset>;
v = specialinvoke v.<org.apache.hadoop.mapred.FileSplit: org.apache.hadoop.fs.Path getPath()>();
v = virtualinvoke v.<org.apache.hadoop.fs.Path: java.lang.String toString()>();
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (java.lang.String,long,int,long,java.lang.String)>(v, v, v, v, v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("KafkaInputSplit{topic=\'\u0001\', startOffset=\u, partition=\u, endOffset=\u, path=\u0001}");
return v;
}
public static org.apache.hadoop.hive.kafka.KafkaInputSplit copyOf(org.apache.hadoop.hive.kafka.KafkaInputSplit)
{
int v;
org.apache.hadoop.fs.Path v;
long v, v;
java.lang.String v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v, v;
v := @parameter: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = new org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String getTopic()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int getPartition()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: org.apache.hadoop.fs.Path getPath()>();
specialinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: void <init>(java.lang.String,int,long,long,org.apache.hadoop.fs.Path)>(v, v, v, v, v);
return v;
}
public static java.util.List slice(long, org.apache.hadoop.hive.kafka.KafkaInputSplit)
{
long v, v, v, v, v, v, v, v, v, v, v, v;
byte v, v, v;
com.google.common.collect.ImmutableList v;
int v, v;
org.apache.hadoop.fs.Path v, v;
java.lang.String v, v;
org.apache.hadoop.hive.kafka.KafkaInputSplit v, v, v, v;
java.util.List v;
com.google.common.collect.ImmutableList$Builder v;
v := @parameter: long;
v := @parameter: org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
v = v - v;
v = v cmp v;
if v <= 0 goto label;
v = staticinvoke <com.google.common.collect.ImmutableList: com.google.common.collect.ImmutableList$Builder builder()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getStartOffset()>();
label:
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = v - v;
v = v cmp v;
if v >= 0 goto label;
v = new org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = v + v;
v = v + 1L;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: org.apache.hadoop.fs.Path getPath()>();
specialinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: void <init>(java.lang.String,int,long,long,org.apache.hadoop.fs.Path)>(v, v, v, v, v);
virtualinvoke v.<com.google.common.collect.ImmutableList$Builder: com.google.common.collect.ImmutableList$Builder add(java.lang.Object)>(v);
v = v + 1L;
v = v + v;
goto label;
label:
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = v cmp v;
if v >= 0 goto label;
v = new org.apache.hadoop.hive.kafka.KafkaInputSplit;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: java.lang.String topic>;
v = v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: int partition>;
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: long getEndOffset()>();
v = virtualinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: org.apache.hadoop.fs.Path getPath()>();
specialinvoke v.<org.apache.hadoop.hive.kafka.KafkaInputSplit: void <init>(java.lang.String,int,long,long,org.apache.hadoop.fs.Path)>(v, v, v, v, v);
virtualinvoke v.<com.google.common.collect.ImmutableList$Builder: com.google.common.collect.ImmutableList$Builder add(java.lang.Object)>(v);
label:
v = virtualinvoke v.<com.google.common.collect.ImmutableList$Builder: com.google.common.collect.ImmutableList build()>();
return v;
label:
v = staticinvoke <org.apache.hadoop.hive.kafka.KafkaInputSplit: org.apache.hadoop.hive.kafka.KafkaInputSplit copyOf(org.apache.hadoop.hive.kafka.KafkaInputSplit)>(v);
v = staticinvoke <java.util.Collections: java.util.List singletonList(java.lang.Object)>(v);
return v;
}
static void <clinit>()
{
java.lang.Class v;
boolean v, v;
v = class "Lorg/apache/hadoop/hive/kafka/KafkaInputSplit;";
v = virtualinvoke v.<java.lang.Class: boolean desiredAssertionStatus()>();
if v != 0 goto label;
v = 1;
goto label;
label:
v = 0;
label:
<org.apache.hadoop.hive.kafka.KafkaInputSplit: boolean $assertionsDisabled> = v;
return;
}
}