public class org.apache.drill.exec.store.kafka.decoders.JsonMessageReader extends java.lang.Object implements org.apache.drill.exec.store.kafka.decoders.MessageReader
{
private static final org.slf4j.Logger logger;
private final org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator stream;
private org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader;
private org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader;
private org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator negotiator;
private org.apache.drill.exec.store.kafka.ReadOptions readOptions;
private java.util.Properties kafkaConsumerProps;
public void <init>()
{
org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
specialinvoke v.<java.lang.Object: void <init>()>();
v = new org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator;
specialinvoke v.<org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator: void <init>()>();
v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator stream> = v;
return;
}
public void init(org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator, org.apache.drill.exec.store.kafka.ReadOptions, org.apache.drill.exec.store.kafka.KafkaStoragePlugin)
{
org.apache.drill.exec.store.kafka.KafkaStoragePlugin v;
java.util.Properties v;
org.apache.drill.exec.physical.resultSet.ResultSetLoader v;
org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator v;
org.apache.drill.exec.store.kafka.ReadOptions v;
org.apache.drill.exec.store.kafka.KafkaStoragePluginConfig v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v := @parameter: org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator;
v := @parameter: org.apache.drill.exec.store.kafka.ReadOptions;
v := @parameter: org.apache.drill.exec.store.kafka.KafkaStoragePlugin;
v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator negotiator> = v;
v = interfaceinvoke v.<org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator: org.apache.drill.exec.physical.resultSet.ResultSetLoader build()>();
v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader> = v;
v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions> = v;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.KafkaStoragePlugin: org.apache.drill.exec.store.kafka.KafkaStoragePluginConfig getConfig()>();
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.KafkaStoragePluginConfig: java.util.Properties getKafkaConsumerProps()>();
v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: java.util.Properties kafkaConsumerProps> = v;
return;
}
public void readMessage(org.apache.kafka.clients.consumer.ConsumerRecord)
{
java.lang.Throwable v;
java.lang.Object[] v, v;
org.apache.drill.common.exceptions.CustomErrorContext v;
org.apache.drill.common.exceptions.UserException$Builder v, v, v;
org.apache.drill.common.exceptions.UserException v;
org.apache.drill.exec.store.kafka.ReadOptions v;
java.lang.String v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
boolean v;
org.slf4j.Logger v;
org.apache.drill.exec.physical.resultSet.ResultSetLoader v;
org.apache.kafka.clients.consumer.ConsumerRecord v;
java.lang.Object v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v := @parameter: org.apache.kafka.clients.consumer.ConsumerRecord;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: java.lang.Object value()>();
label:
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: void parseAndWrite(org.apache.kafka.clients.consumer.ConsumerRecord,byte[])>(v, v);
label:
goto label;
label:
v := @caughtexception;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isSkipInvalidRecords()>();
if v != 0 goto label;
v = staticinvoke <org.apache.drill.common.exceptions.UserException: org.apache.drill.common.exceptions.UserException$Builder dataReadError(java.lang.Throwable)>(v);
v = newarray (java.lang.Object)[1];
v[0] = "store.kafka.reader.skip_invalid_records";
v = staticinvoke <java.lang.String: java.lang.String format(java.lang.String,java.lang.Object[])>("Error happened when parsing invalid record. Please set `%s` option to \'true\' to skip invalid records.", v);
v = newarray (java.lang.Object)[0];
v = virtualinvoke v.<org.apache.drill.common.exceptions.UserException$Builder: org.apache.drill.common.exceptions.UserException$Builder message(java.lang.String,java.lang.Object[])>(v, v);
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
v = interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.ResultSetLoader: org.apache.drill.common.exceptions.CustomErrorContext errorContext()>();
v = virtualinvoke v.<org.apache.drill.common.exceptions.UserException$Builder: org.apache.drill.common.exceptions.UserException$Builder addContext(org.apache.drill.common.exceptions.CustomErrorContext)>(v);
v = <org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.slf4j.Logger logger>;
v = virtualinvoke v.<org.apache.drill.common.exceptions.UserException$Builder: org.apache.drill.common.exceptions.UserException build(org.slf4j.Logger)>(v);
throw v;
label:
return;
catch org.apache.drill.exec.store.easy.json.parser.TokenIterator$RecoverableJsonException from label to label with label;
}
private void parseAndWrite(org.apache.kafka.clients.consumer.ConsumerRecord, byte[])
{
byte[] v;
org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator v, v;
org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions v;
java.lang.Integer v;
org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader$KafkaJsonLoaderBuilder v;
java.io.ByteArrayInputStream v;
org.apache.drill.exec.store.easy.json.loader.JsonLoader v;
java.lang.Long v, v;
org.apache.drill.exec.store.easy.json.parser.JsonStructureParser v;
boolean v, v, v, v, v, v, v;
org.apache.drill.exec.server.options.OptionSet v;
org.apache.kafka.clients.consumer.ConsumerRecord v;
org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator v;
org.apache.drill.common.exceptions.CustomErrorContext v;
long v, v;
java.lang.Iterable v;
org.apache.drill.exec.physical.resultSet.RowSetLoader v;
org.apache.drill.exec.store.kafka.ReadOptions v, v, v, v, v, v;
int v;
java.lang.String v, v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
org.apache.drill.exec.physical.resultSet.ResultSetLoader v, v;
org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder v, v, v, v, v;
org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader v, v;
org.apache.drill.exec.store.kafka.MetaDataField v, v, v, v, v;
java.lang.Object v, v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v := @parameter: org.apache.kafka.clients.consumer.ConsumerRecord;
v := @parameter: byte[];
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator stream>;
v = new java.io.ByteArrayInputStream;
specialinvoke v.<java.io.ByteArrayInputStream: void <init>(byte[])>(v);
virtualinvoke v.<org.apache.drill.exec.store.easy.json.loader.ClosingStreamIterator: void setValue(java.io.InputStream)>(v);
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader>;
if v != null goto label;
v = new org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions;
specialinvoke v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: void <init>()>();
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isAllTextMode()>();
v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: boolean allTextMode> = v;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isReadNumbersAsDouble()>();
v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: boolean readNumbersAsDouble> = v;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isSkipInvalidRecords()>();
v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: boolean skipMalformedRecords> = v;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isAllowNanInf()>();
v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: boolean allowNanInf> = v;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isAllowEscapeAnyChar()>();
v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: boolean enableEscapeAnyChar> = v;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.ReadOptions readOptions>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.ReadOptions: boolean isSkipInvalidRecords()>();
v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions: boolean skipMalformedDocument> = v;
v = new org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader$KafkaJsonLoaderBuilder;
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader$KafkaJsonLoaderBuilder: void <init>()>();
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader$KafkaJsonLoaderBuilder: org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder resultSetLoader(org.apache.drill.exec.physical.resultSet.ResultSetLoader)>(v);
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator negotiator>;
v = interfaceinvoke v.<org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator: org.apache.drill.exec.server.options.OptionSet queryOptions()>();
v = virtualinvoke v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder: org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder standardOptions(org.apache.drill.exec.server.options.OptionSet)>(v);
v = virtualinvoke v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder: org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder options(org.apache.drill.exec.store.easy.json.loader.JsonLoaderOptions)>(v);
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator negotiator>;
v = interfaceinvoke v.<org.apache.drill.exec.physical.impl.scan.framework.SchemaNegotiator: org.apache.drill.common.exceptions.CustomErrorContext parentErrorContext()>();
v = virtualinvoke v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder: org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder errorContext(org.apache.drill.common.exceptions.CustomErrorContext)>(v);
v = staticinvoke <org.apache.drill.exec.store.kafka.decoders.JsonMessageReader$lambda_parseAndWrite_0__11: java.lang.Iterable bootstrap$(org.apache.drill.exec.store.kafka.decoders.JsonMessageReader)>(v);
v = virtualinvoke v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder: org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder fromStream(java.lang.Iterable)>(v);
v = virtualinvoke v.<org.apache.drill.exec.store.easy.json.loader.JsonLoaderImpl$JsonLoaderBuilder: org.apache.drill.exec.store.easy.json.loader.JsonLoader build()>();
v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader> = v;
label:
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
v = interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.ResultSetLoader: org.apache.drill.exec.physical.resultSet.RowSetLoader writer()>();
interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.RowSetLoader: boolean start()>();
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader>;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader: org.apache.drill.exec.store.easy.json.parser.JsonStructureParser parser()>();
v = virtualinvoke v.<org.apache.drill.exec.store.easy.json.parser.JsonStructureParser: boolean next()>();
if v == 0 goto label;
v = <org.apache.drill.exec.store.kafka.MetaDataField: org.apache.drill.exec.store.kafka.MetaDataField KAFKA_TOPIC>;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: java.lang.String topic()>();
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: void writeValue(org.apache.drill.exec.physical.resultSet.RowSetLoader,org.apache.drill.exec.store.kafka.MetaDataField,java.lang.Object)>(v, v, v);
v = <org.apache.drill.exec.store.kafka.MetaDataField: org.apache.drill.exec.store.kafka.MetaDataField KAFKA_PARTITION_ID>;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: int partition()>();
v = staticinvoke <java.lang.Integer: java.lang.Integer valueOf(int)>(v);
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: void writeValue(org.apache.drill.exec.physical.resultSet.RowSetLoader,org.apache.drill.exec.store.kafka.MetaDataField,java.lang.Object)>(v, v, v);
v = <org.apache.drill.exec.store.kafka.MetaDataField: org.apache.drill.exec.store.kafka.MetaDataField KAFKA_OFFSET>;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: long offset()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: void writeValue(org.apache.drill.exec.physical.resultSet.RowSetLoader,org.apache.drill.exec.store.kafka.MetaDataField,java.lang.Object)>(v, v, v);
v = <org.apache.drill.exec.store.kafka.MetaDataField: org.apache.drill.exec.store.kafka.MetaDataField KAFKA_TIMESTAMP>;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: long timestamp()>();
v = staticinvoke <java.lang.Long: java.lang.Long valueOf(long)>(v);
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: void writeValue(org.apache.drill.exec.physical.resultSet.RowSetLoader,org.apache.drill.exec.store.kafka.MetaDataField,java.lang.Object)>(v, v, v);
v = <org.apache.drill.exec.store.kafka.MetaDataField: org.apache.drill.exec.store.kafka.MetaDataField KAFKA_MSG_KEY>;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: java.lang.Object key()>();
if v == null goto label;
v = virtualinvoke v.<org.apache.kafka.clients.consumer.ConsumerRecord: java.lang.Object key()>();
v = virtualinvoke v.<java.lang.Object: java.lang.String toString()>();
goto label;
label:
v = null;
label:
specialinvoke v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: void writeValue(org.apache.drill.exec.physical.resultSet.RowSetLoader,org.apache.drill.exec.store.kafka.MetaDataField,java.lang.Object)>(v, v, v);
interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.RowSetLoader: void save()>();
label:
return;
}
private void writeValue(org.apache.drill.exec.physical.resultSet.RowSetLoader, org.apache.drill.exec.store.kafka.MetaDataField, java.lang.Object)
{
org.apache.drill.exec.record.MaterializedField v;
org.apache.drill.exec.record.metadata.TupleMetadata v;
org.apache.drill.common.types.TypeProtos$MinorType v;
org.apache.drill.exec.physical.resultSet.RowSetLoader v;
org.apache.drill.common.types.TypeProtos$DataMode v;
org.apache.drill.exec.store.kafka.MetaDataField v;
org.apache.drill.exec.vector.accessor.ObjectWriter v;
java.lang.Object v;
org.apache.drill.exec.record.metadata.PrimitiveColumnMetadata v;
java.lang.String v, v, v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v := @parameter: org.apache.drill.exec.physical.resultSet.RowSetLoader;
v := @parameter: org.apache.drill.exec.store.kafka.MetaDataField;
v := @parameter: java.lang.Object;
v = interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.RowSetLoader: org.apache.drill.exec.record.metadata.TupleMetadata tupleSchema()>();
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.MetaDataField: java.lang.String getFieldName()>();
v = interfaceinvoke v.<org.apache.drill.exec.record.metadata.TupleMetadata: org.apache.drill.exec.record.MaterializedField column(java.lang.String)>(v);
if v != null goto label;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.MetaDataField: java.lang.String getFieldName()>();
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.MetaDataField: org.apache.drill.common.types.TypeProtos$MinorType getFieldType()>();
v = <org.apache.drill.common.types.TypeProtos$DataMode: org.apache.drill.common.types.TypeProtos$DataMode OPTIONAL>;
v = staticinvoke <org.apache.drill.exec.record.metadata.MetadataUtils: org.apache.drill.exec.record.metadata.PrimitiveColumnMetadata newScalar(java.lang.String,org.apache.drill.common.types.TypeProtos$MinorType,org.apache.drill.common.types.TypeProtos$DataMode)>(v, v, v);
interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.RowSetLoader: int addColumn(org.apache.drill.exec.record.metadata.ColumnMetadata)>(v);
label:
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.MetaDataField: java.lang.String getFieldName()>();
v = interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.RowSetLoader: org.apache.drill.exec.vector.accessor.ObjectWriter column(java.lang.String)>(v);
interfaceinvoke v.<org.apache.drill.exec.vector.accessor.ObjectWriter: void setObject(java.lang.Object)>(v);
return;
}
public org.apache.drill.exec.physical.resultSet.ResultSetLoader getResultSetLoader()
{
org.apache.drill.exec.physical.resultSet.ResultSetLoader v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
return v;
}
public org.apache.kafka.clients.consumer.KafkaConsumer getConsumer(org.apache.drill.exec.store.kafka.KafkaStoragePlugin)
{
org.apache.drill.exec.store.kafka.KafkaStoragePlugin v;
org.apache.kafka.common.serialization.ByteArrayDeserializer v, v;
java.util.Properties v;
org.apache.drill.exec.store.kafka.KafkaStoragePluginConfig v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
org.apache.kafka.clients.consumer.KafkaConsumer v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v := @parameter: org.apache.drill.exec.store.kafka.KafkaStoragePlugin;
v = new org.apache.kafka.clients.consumer.KafkaConsumer;
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.KafkaStoragePlugin: org.apache.drill.exec.store.kafka.KafkaStoragePluginConfig getConfig()>();
v = virtualinvoke v.<org.apache.drill.exec.store.kafka.KafkaStoragePluginConfig: java.util.Properties getKafkaConsumerProps()>();
v = new org.apache.kafka.common.serialization.ByteArrayDeserializer;
specialinvoke v.<org.apache.kafka.common.serialization.ByteArrayDeserializer: void <init>()>();
v = new org.apache.kafka.common.serialization.ByteArrayDeserializer;
specialinvoke v.<org.apache.kafka.common.serialization.ByteArrayDeserializer: void <init>()>();
specialinvoke v.<org.apache.kafka.clients.consumer.KafkaConsumer: void <init>(java.util.Properties,org.apache.kafka.common.serialization.Deserializer,org.apache.kafka.common.serialization.Deserializer)>(v, v, v);
return v;
}
public boolean endBatch()
{
org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader v;
org.apache.drill.exec.physical.resultSet.ResultSetLoader v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
boolean v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader>;
virtualinvoke v.<org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader: void endBatch()>();
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
v = interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.ResultSetLoader: boolean hasRows()>();
return v;
}
public void close()
{
org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader v;
org.slf4j.Logger v;
org.apache.drill.exec.physical.resultSet.ResultSetLoader v;
java.lang.String v;
java.lang.Exception v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
label:
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader>;
virtualinvoke v.<org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader: void close()>();
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
interfaceinvoke v.<org.apache.drill.exec.physical.resultSet.ResultSetLoader: void close()>();
label:
goto label;
label:
v := @caughtexception;
v = <org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.slf4j.Logger logger>;
v = virtualinvoke v.<java.lang.Exception: java.lang.String getMessage()>();
interfaceinvoke v.<org.slf4j.Logger: void warn(java.lang.String,java.lang.Object)>("Error while closing JsonMessageReader: {}", v);
label:
return;
catch java.lang.Exception from label to label with label;
}
public java.lang.String toString()
{
org.apache.drill.exec.physical.resultSet.ResultSetLoader v;
org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader v;
java.util.StringJoiner v, v, v;
java.lang.Class v;
java.lang.String v, v, v, v, v;
org.apache.drill.exec.store.kafka.decoders.JsonMessageReader v;
v := @this: org.apache.drill.exec.store.kafka.decoders.JsonMessageReader;
v = new java.util.StringJoiner;
v = class "Lorg/apache/drill/exec/store/kafka/decoders/JsonMessageReader;";
v = virtualinvoke v.<java.lang.Class: java.lang.String getSimpleName()>();
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (java.lang.String)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("\u[");
specialinvoke v.<java.util.StringJoiner: void <init>(java.lang.CharSequence,java.lang.CharSequence,java.lang.CharSequence)>(", ", v, "]");
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader kafkaJsonLoader>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (org.apache.drill.exec.store.kafka.decoders.KafkaJsonLoader)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("kafkaJsonLoader=\u0001");
v = virtualinvoke v.<java.util.StringJoiner: java.util.StringJoiner add(java.lang.CharSequence)>(v);
v = v.<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.apache.drill.exec.physical.resultSet.ResultSetLoader resultSetLoader>;
v = dynamicinvoke "makeConcatWithConstants" <java.lang.String (org.apache.drill.exec.physical.resultSet.ResultSetLoader)>(v) <java.lang.invoke.StringConcatFactory: java.lang.invoke.CallSite makeConcatWithConstants(java.lang.invoke.MethodHandles$Lookup,java.lang.String,java.lang.invoke.MethodType,java.lang.String,java.lang.Object[])>("resultSetLoader=\u0001");
v = virtualinvoke v.<java.util.StringJoiner: java.util.StringJoiner add(java.lang.CharSequence)>(v);
v = virtualinvoke v.<java.util.StringJoiner: java.lang.String toString()>();
return v;
}
static void <clinit>()
{
org.slf4j.Logger v;
v = staticinvoke <org.slf4j.LoggerFactory: org.slf4j.Logger getLogger(java.lang.Class)>(class "Lorg/apache/drill/exec/store/kafka/decoders/JsonMessageReader;");
<org.apache.drill.exec.store.kafka.decoders.JsonMessageReader: org.slf4j.Logger logger> = v;
return;
}
}