public class HiveScriptIOSchema extends Object implements HiveInspectors, scala.Product, scala.Serializable
HiveInspectors.typeInfoConversions| Constructor and Description | 
|---|
HiveScriptIOSchema(scala.collection.Seq<scala.Tuple2<String,String>> inputRowFormat,
                  scala.collection.Seq<scala.Tuple2<String,String>> outputRowFormat,
                  scala.Option<String> inputSerdeClass,
                  scala.Option<String> outputSerdeClass,
                  scala.collection.Seq<scala.Tuple2<String,String>> inputSerdeProps,
                  scala.collection.Seq<scala.Tuple2<String,String>> outputSerdeProps,
                  scala.Option<String> recordReaderClass,
                  scala.Option<String> recordWriterClass,
                  boolean schemaLess)  | 
| Modifier and Type | Method and Description | 
|---|---|
static HiveScriptIOSchema | 
apply(org.apache.spark.sql.catalyst.plans.logical.ScriptInputOutputSchema input)  | 
scala.Option<scala.Tuple2<org.apache.hadoop.hive.serde2.AbstractSerDe,org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector>> | 
initInputSerDe(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input)  | 
scala.Option<scala.Tuple2<org.apache.hadoop.hive.serde2.AbstractSerDe,org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector>> | 
initOutputSerDe(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output)  | 
scala.collection.Seq<scala.Tuple2<String,String>> | 
inputRowFormat()  | 
scala.collection.immutable.Map<String,String> | 
inputRowFormatMap()  | 
scala.Option<String> | 
inputSerdeClass()  | 
scala.collection.Seq<scala.Tuple2<String,String>> | 
inputSerdeProps()  | 
scala.collection.Seq<scala.Tuple2<String,String>> | 
outputRowFormat()  | 
scala.collection.immutable.Map<String,String> | 
outputRowFormatMap()  | 
scala.Option<String> | 
outputSerdeClass()  | 
scala.collection.Seq<scala.Tuple2<String,String>> | 
outputSerdeProps()  | 
scala.Option<org.apache.hadoop.hive.ql.exec.RecordReader> | 
recordReader(java.io.InputStream inputStream,
            org.apache.hadoop.conf.Configuration conf)  | 
scala.Option<String> | 
recordReaderClass()  | 
scala.Option<org.apache.hadoop.hive.ql.exec.RecordWriter> | 
recordWriter(java.io.OutputStream outputStream,
            org.apache.hadoop.conf.Configuration conf)  | 
scala.Option<String> | 
recordWriterClass()  | 
boolean | 
schemaLess()  | 
equals, getClass, hashCode, notify, notifyAll, toString, wait, wait, waitdecimalTypeInfoToCatalyst, getBinaryWritable, getBinaryWritableConstantObjectInspector, getBooleanWritable, getBooleanWritableConstantObjectInspector, getByteWritable, getByteWritableConstantObjectInspector, getDateWritable, getDateWritableConstantObjectInspector, getDecimalWritable, getDecimalWritableConstantObjectInspector, getDoubleWritable, getDoubleWritableConstantObjectInspector, getFloatWritable, getFloatWritableConstantObjectInspector, getIntWritable, getIntWritableConstantObjectInspector, getLongWritable, getLongWritableConstantObjectInspector, getPrimitiveNullWritableConstantObjectInspector, getShortWritable, getShortWritableConstantObjectInspector, getStringWritable, getStringWritableConstantObjectInspector, getTimestampWritable, getTimestampWritableConstantObjectInspector, inspectorToDataType, isSubClassOf, javaTypeToDataType, toInspector, toInspector, unwrapperFor, unwrapperFor, withNullSafe, wrap, wrap, wrap, wrapperForpublic HiveScriptIOSchema(scala.collection.Seq<scala.Tuple2<String,String>> inputRowFormat,
                          scala.collection.Seq<scala.Tuple2<String,String>> outputRowFormat,
                          scala.Option<String> inputSerdeClass,
                          scala.Option<String> outputSerdeClass,
                          scala.collection.Seq<scala.Tuple2<String,String>> inputSerdeProps,
                          scala.collection.Seq<scala.Tuple2<String,String>> outputSerdeProps,
                          scala.Option<String> recordReaderClass,
                          scala.Option<String> recordWriterClass,
                          boolean schemaLess)
public static HiveScriptIOSchema apply(org.apache.spark.sql.catalyst.plans.logical.ScriptInputOutputSchema input)
public scala.collection.Seq<scala.Tuple2<String,String>> inputRowFormat()
public scala.collection.Seq<scala.Tuple2<String,String>> outputRowFormat()
public scala.Option<String> inputSerdeClass()
public scala.Option<String> outputSerdeClass()
public scala.collection.Seq<scala.Tuple2<String,String>> inputSerdeProps()
public scala.collection.Seq<scala.Tuple2<String,String>> outputSerdeProps()
public scala.Option<String> recordReaderClass()
public scala.Option<String> recordWriterClass()
public boolean schemaLess()
public scala.collection.immutable.Map<String,String> inputRowFormatMap()
public scala.collection.immutable.Map<String,String> outputRowFormatMap()
public scala.Option<scala.Tuple2<org.apache.hadoop.hive.serde2.AbstractSerDe,org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector>> initInputSerDe(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input)
public scala.Option<scala.Tuple2<org.apache.hadoop.hive.serde2.AbstractSerDe,org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector>> initOutputSerDe(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output)
public scala.Option<org.apache.hadoop.hive.ql.exec.RecordReader> recordReader(java.io.InputStream inputStream,
                                                                              org.apache.hadoop.conf.Configuration conf)
public scala.Option<org.apache.hadoop.hive.ql.exec.RecordWriter> recordWriter(java.io.OutputStream outputStream,
                                                                              org.apache.hadoop.conf.Configuration conf)