public abstract class oadd.org.apache.hadoop.fs.FSInputChecker extends oadd.org.apache.hadoop.fs.FSInputStream
{
public static final org.slf4j.Logger LOG;
protected oadd.org.apache.hadoop.fs.Path file;
private java.util.zip.Checksum sum;
private boolean verifyChecksum;
private int maxChunkSize;
private byte[] buf;
private byte[] checksum;
private java.nio.IntBuffer checksumInts;
private int pos;
private int count;
private int numOfRetries;
private long chunkPos;
private static final int CHUNKS_PER_READ;
protected static final int CHECKSUM_SIZE;
static final boolean $assertionsDisabled;
protected void <init>(oadd.org.apache.hadoop.fs.Path, int)
{
int v;
oadd.org.apache.hadoop.fs.Path v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: oadd.org.apache.hadoop.fs.Path;
v := @parameter: int;
specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputStream: void <init>()>();
v.<oadd.org.apache.hadoop.fs.FSInputChecker: boolean verifyChecksum> = 1;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos> = 0L;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: oadd.org.apache.hadoop.fs.Path file> = v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int numOfRetries> = v;
return;
}
protected void <init>(oadd.org.apache.hadoop.fs.Path, int, boolean, java.util.zip.Checksum, int, int)
{
java.util.zip.Checksum v;
int v, v, v;
oadd.org.apache.hadoop.fs.Path v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
boolean v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: oadd.org.apache.hadoop.fs.Path;
v := @parameter: int;
v := @parameter: boolean;
v := @parameter: java.util.zip.Checksum;
v := @parameter: int;
v := @parameter: int;
specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void <init>(oadd.org.apache.hadoop.fs.Path,int)>(v, v);
virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void set(boolean,java.util.zip.Checksum,int,int)>(v, v, v, v);
return;
}
protected abstract int readChunk(long, byte[], int, int, byte[]) throws java.io.IOException;
protected abstract long getChunkPosition(long);
protected synchronized boolean needChecksum()
{
java.util.zip.Checksum v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
boolean v, v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: boolean verifyChecksum>;
if v == 0 goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum>;
if v == null goto label;
v = 1;
goto label;
label:
v = 0;
label:
return v;
}
public synchronized int read() throws java.io.IOException
{
byte[] v;
short v;
byte v;
int v, v, v, v, v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v < v goto label;
specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void fill()>();
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v < v goto label;
v = (int) -1;
return v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] buf>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v + 1;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = v;
v = v[v];
v = v & 255;
return v;
}
public synchronized int read(byte[], int, int) throws java.io.IOException
{
byte[] v;
java.lang.IndexOutOfBoundsException v;
int v, v, v, v, v, v, v, v, v, v, v, v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: byte[];
v := @parameter: int;
v := @parameter: int;
v = v | v;
v = v + v;
v = v | v;
v = lengthof v;
v = v + v;
v = v - v;
v = v | v;
if v >= 0 goto label;
v = new java.lang.IndexOutOfBoundsException;
specialinvoke v.<java.lang.IndexOutOfBoundsException: void <init>()>();
throw v;
label:
if v != 0 goto label;
return 0;
label:
v = 0;
label:
v = v + v;
v = v - v;
v = specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: int read1(byte[],int,int)>(v, v, v);
if v > 0 goto label;
if v != 0 goto label;
v = v;
goto label;
label:
v = v;
label:
return v;
label:
v = v + v;
if v < v goto label;
return v;
}
private void fill() throws java.io.IOException
{
byte[] v;
java.lang.AssertionError v;
int v, v, v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
boolean v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v = <oadd.org.apache.hadoop.fs.FSInputChecker: boolean $assertionsDisabled>;
if v != 0 goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v >= v goto label;
v = new java.lang.AssertionError;
specialinvoke v.<java.lang.AssertionError: void <init>()>();
throw v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] buf>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
v = specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: int readChecksumChunk(byte[],int,int)>(v, 0, v);
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count> = v;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v >= 0 goto label;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count> = 0;
label:
return;
}
protected final synchronized int readAndDiscard(int) throws java.io.IOException
{
byte[] v;
int v, v, v, v, v, v, v, v, v, v, v, v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: int;
v = 0;
label:
if v >= v goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v < v goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] buf>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
v = specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: int readChecksumChunk(byte[],int,int)>(v, 0, v);
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count> = v;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v <= 0 goto label;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v - v;
v = v - v;
v = staticinvoke <java.lang.Math: int min(int,int)>(v, v);
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v + v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = v;
v = v + v;
goto label;
label:
return v;
}
private int read1(byte[], int, int) throws java.io.IOException
{
byte[] v, v;
int v, v, v, v, v, v, v, v, v, v, v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: byte[];
v := @parameter: int;
v := @parameter: int;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v - v;
if v > 0 goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
if v < v goto label;
v = specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: int readChecksumChunk(byte[],int,int)>(v, v, v);
return v;
label:
specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void fill()>();
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
if v > 0 goto label;
v = (int) -1;
return v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
label:
if v >= v goto label;
v = v;
goto label;
label:
v = v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] buf>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
staticinvoke <java.lang.System: void arraycopy(java.lang.Object,int,java.lang.Object,int,int)>(v, v, v, v, v);
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v + v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = v;
return v;
}
private int readChecksumChunk(byte[], int, int) throws java.io.IOException
{
byte[] v, v;
long v, v, v, v, v;
int v, v, v, v, v, v;
java.lang.String v, v;
boolean v, v, v;
org.slf4j.Logger v;
oadd.org.apache.hadoop.fs.ChecksumException v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: byte[];
v := @parameter: int;
v := @parameter: int;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = 0;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count> = 0;
v = 0;
v = 1;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int numOfRetries>;
label:
v = v - 1;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] checksum>;
v = virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: int readChunk(long,byte[],int,int,byte[])>(v, v, v, v, v);
if v <= 0 goto label;
v = virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: boolean needChecksum()>();
if v == 0 goto label;
specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void verifySums(byte[],int,int)>(v, v, v);
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v + v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos> = v;
label:
v = 0;
label:
goto label;
label:
v := @caughtexception;
v = <oadd.org.apache.hadoop.fs.FSInputChecker: org.slf4j.Logger LOG>;
v = v + v;
v = v + v;
v = staticinvoke <oadd.org.apache.hadoop.util.StringUtils: java.lang.String byteToHexString(byte[],int,int)>(v, v, v);
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (int,int,java.lang.String)>(v, v, v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Found checksum error: b[\u, \u]=\u0001");
interfaceinvoke v.<org.slf4j.Logger: void info(java.lang.String,java.lang.Throwable)>(v, v);
if v != 0 goto label;
throw v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: boolean seekToNewSource(long)>(v);
if v == 0 goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void seek(long)>(v);
goto label;
label:
throw v;
label:
if v != 0 goto label;
return v;
catch oadd.org.apache.hadoop.fs.ChecksumException from label to label with label;
}
private void verifySums(byte[], int, int) throws oadd.org.apache.hadoop.fs.ChecksumException
{
byte[] v;
long v, v, v;
java.util.zip.Checksum v, v, v;
java.nio.IntBuffer v, v, v;
int v, v, v, v, v, v, v, v, v, v, v, v, v, v;
java.lang.String v;
oadd.org.apache.hadoop.fs.ChecksumException v;
oadd.org.apache.hadoop.fs.Path v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: byte[];
v := @parameter: int;
v := @parameter: int;
v = v;
v = 0;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.nio.IntBuffer checksumInts>;
virtualinvoke v.<java.nio.IntBuffer: java.nio.IntBuffer rewind()>();
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.nio.IntBuffer checksumInts>;
v = v - 1;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
v = v / v;
v = v + 1;
virtualinvoke v.<java.nio.IntBuffer: java.nio.IntBuffer limit(int)>(v);
label:
if v <= 0 goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum>;
v = v + v;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
v = staticinvoke <java.lang.Math: int min(int,int)>(v, v);
interfaceinvoke v.<java.util.zip.Checksum: void update(byte[],int,int)>(v, v, v);
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.nio.IntBuffer checksumInts>;
v = virtualinvoke v.<java.nio.IntBuffer: int get()>();
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum>;
v = interfaceinvoke v.<java.util.zip.Checksum: long getValue()>();
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum>;
interfaceinvoke v.<java.util.zip.Checksum: void reset()>();
if v == v goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v + v;
v = new oadd.org.apache.hadoop.fs.ChecksumException;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: oadd.org.apache.hadoop.fs.Path file>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (oadd.org.apache.hadoop.fs.Path,long,int,int)>(v, v, v, v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("Checksum error: \u at \u exp: \u got: \u0001");
specialinvoke v.<oadd.org.apache.hadoop.fs.ChecksumException: void <init>(java.lang.String,long)>(v, v);
throw v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
v = v - v;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize>;
v = v + v;
goto label;
label:
return;
}
public static long checksum2long(byte[])
{
byte[] v;
long v, v, v;
byte v;
int v, v, v, v, v, v;
v := @parameter: byte[];
v = 0L;
v = 0;
label:
v = lengthof v;
if v >= v goto label;
v = v[v];
v = 255L & v;
v = lengthof v;
v = v - v;
v = v - 1;
v = v * 8;
v = v << v;
v = v | v;
v = v + 1;
goto label;
label:
return v;
}
public synchronized long getPos() throws java.io.IOException
{
int v, v, v;
long v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v - v;
v = staticinvoke <java.lang.Math: long max(long,long)>(0L, v);
v = v - v;
return v;
}
public synchronized int available() throws java.io.IOException
{
int v, v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos>;
v = v - v;
v = staticinvoke <java.lang.Math: int max(int,int)>(0, v);
return v;
}
public synchronized long skip(long) throws java.io.IOException
{
byte v;
long v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: long;
v = v cmp 0L;
if v > 0 goto label;
return 0L;
label:
v = virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: long getPos()>();
v = v + v;
virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void seek(long)>(v);
return v;
}
public synchronized void seek(long) throws java.io.IOException
{
byte[] v;
long v, v, v, v, v, v, v, v;
java.io.EOFException v;
byte v, v, v;
int v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: long;
v = v cmp 0L;
if v >= 0 goto label;
v = new java.io.EOFException;
specialinvoke v.<java.io.EOFException: void <init>(java.lang.String)>("Cannot seek to a negative offset");
throw v;
label:
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count>;
v = v - v;
v = v cmp v;
if v < 0 goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v cmp v;
if v >= 0 goto label;
v = v - v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = v;
return;
label:
specialinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: void resetState()>();
v = virtualinvoke v.<oadd.org.apache.hadoop.fs.FSInputChecker: long getChunkPosition(long)>(v);
v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos> = v;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: long chunkPos>;
v = v - v;
if v <= 0 goto label;
v = newarray (byte)[v];
staticinvoke <oadd.org.apache.hadoop.fs.FSInputChecker: int readFully(java.io.InputStream,byte[],int,int)>(v, v, 0, v);
label:
return;
}
protected static int readFully(java.io.InputStream, byte[], int, int) throws java.io.IOException
{
byte[] v;
int v, v, v, v, v, v, v;
java.io.InputStream v;
v := @parameter: java.io.InputStream;
v := @parameter: byte[];
v := @parameter: int;
v := @parameter: int;
v = 0;
label:
v = v + v;
v = v - v;
v = virtualinvoke v.<java.io.InputStream: int read(byte[],int,int)>(v, v, v);
if v > 0 goto label;
if v != 0 goto label;
v = v;
goto label;
label:
v = v;
label:
return v;
label:
v = v + v;
if v < v goto label;
return v;
}
protected final synchronized void set(boolean, java.util.zip.Checksum, int, int)
{
byte[] v, v, v;
java.lang.AssertionError v;
java.nio.ByteBuffer v, v;
java.util.zip.Checksum v;
int v, v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
boolean v, v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: boolean;
v := @parameter: java.util.zip.Checksum;
v := @parameter: int;
v := @parameter: int;
v = <oadd.org.apache.hadoop.fs.FSInputChecker: boolean $assertionsDisabled>;
if v != 0 goto label;
if v == 0 goto label;
if v == null goto label;
if v == 4 goto label;
v = new java.lang.AssertionError;
specialinvoke v.<java.lang.AssertionError: void <init>()>();
throw v;
label:
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int maxChunkSize> = v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: boolean verifyChecksum> = v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum> = v;
v = newarray (byte)[v];
v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] buf> = v;
v = 32 * v;
v = newarray (byte)[v];
v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] checksum> = v;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: byte[] checksum>;
v = staticinvoke <java.nio.ByteBuffer: java.nio.ByteBuffer wrap(byte[])>(v);
v = virtualinvoke v.<java.nio.ByteBuffer: java.nio.ByteBuffer asIntBuffer()>();
v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.nio.IntBuffer checksumInts> = v;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count> = 0;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = 0;
return;
}
public final boolean markSupported()
{
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
return 0;
}
public final void mark(int)
{
int v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v := @parameter: int;
return;
}
public final void reset() throws java.io.IOException
{
java.io.IOException v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v = new java.io.IOException;
specialinvoke v.<java.io.IOException: void <init>(java.lang.String)>("mark/reset not supported");
throw v;
}
private void resetState()
{
java.util.zip.Checksum v, v;
oadd.org.apache.hadoop.fs.FSInputChecker v;
v := @this: oadd.org.apache.hadoop.fs.FSInputChecker;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int count> = 0;
v.<oadd.org.apache.hadoop.fs.FSInputChecker: int pos> = 0;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum>;
if v == null goto label;
v = v.<oadd.org.apache.hadoop.fs.FSInputChecker: java.util.zip.Checksum sum>;
interfaceinvoke v.<java.util.zip.Checksum: void reset()>();
label:
return;
}
static void <clinit>()
{
org.slf4j.Logger v;
java.lang.Class v;
boolean v, v;
v = class "Loadd/org/apache/hadoop/fs/FSInputChecker;";
v = virtualinvoke v.<java.lang.Class: boolean desiredAssertionStatus()>();
if v != 0 goto label;
v = 1;
goto label;
label:
v = 0;
label:
<oadd.org.apache.hadoop.fs.FSInputChecker: boolean $assertionsDisabled> = v;
v = staticinvoke <org.slf4j.LoggerFactory: org.slf4j.Logger getLogger(java.lang.Class)>(class "Loadd/org/apache/hadoop/fs/FSInputChecker;");
<oadd.org.apache.hadoop.fs.FSInputChecker: org.slf4j.Logger LOG> = v;
return;
}
}