Class DataTypeErrors

Object
org.apache.spark.sql.errors.DataTypeErrors

public class DataTypeErrors extends Object
Object for grouping error messages from (most) exceptions thrown during query execution. This does not include exceptions thrown during the eager execution of commands, which are grouped into QueryCompilationErrors.
  • Constructor Details

    • DataTypeErrors

      public DataTypeErrors()
  • Method Details

    • unsupportedOperationExceptionError

      public static org.apache.spark.SparkUnsupportedOperationException unsupportedOperationExceptionError()
    • decimalPrecisionExceedsMaxPrecisionError

      public static org.apache.spark.SparkArithmeticException decimalPrecisionExceedsMaxPrecisionError(int precision, int maxPrecision)
    • unsupportedRoundingMode

      public static SparkException unsupportedRoundingMode(scala.Enumeration.Value roundMode)
    • outOfDecimalTypeRangeError

      public static org.apache.spark.SparkArithmeticException outOfDecimalTypeRangeError(org.apache.spark.unsafe.types.UTF8String str)
    • unsupportedJavaTypeError

      public static org.apache.spark.SparkRuntimeException unsupportedJavaTypeError(Class<?> clazz)
    • nullLiteralsCannotBeCastedError

      public static org.apache.spark.SparkUnsupportedOperationException nullLiteralsCannotBeCastedError(String name)
    • notUserDefinedTypeError

      public static Throwable notUserDefinedTypeError(String name, String userClass)
    • cannotLoadUserDefinedTypeError

      public static Throwable cannotLoadUserDefinedTypeError(String name, String userClass)
    • unsupportedArrayTypeError

      public static org.apache.spark.SparkRuntimeException unsupportedArrayTypeError(Class<?> clazz)
    • schemaFailToParseError

      public static Throwable schemaFailToParseError(String schema, Throwable e)
    • invalidDayTimeIntervalType

      public static Throwable invalidDayTimeIntervalType(String startFieldName, String endFieldName)
    • invalidDayTimeField

      public static Throwable invalidDayTimeField(byte field, scala.collection.Seq<String> supportedIds)
    • invalidYearMonthField

      public static Throwable invalidYearMonthField(byte field, scala.collection.Seq<String> supportedIds)
    • decimalCannotGreaterThanPrecisionError

      public static Throwable decimalCannotGreaterThanPrecisionError(int scale, int precision)
    • negativeScaleNotAllowedError

      public static Throwable negativeScaleNotAllowedError(int scale)
    • attributeNameSyntaxError

      public static Throwable attributeNameSyntaxError(String name)
    • cannotMergeIncompatibleDataTypesError

      public static Throwable cannotMergeIncompatibleDataTypesError(DataType left, DataType right)
    • cannotMergeDecimalTypesWithIncompatibleScaleError

      public static Throwable cannotMergeDecimalTypesWithIncompatibleScaleError(int leftScale, int rightScale)
    • dataTypeUnsupportedError

      public static Throwable dataTypeUnsupportedError(String dataType, String failure)
    • invalidFieldName

      public static Throwable invalidFieldName(scala.collection.Seq<String> fieldName, scala.collection.Seq<String> path, org.apache.spark.sql.catalyst.trees.Origin context)
    • unscaledValueTooLargeForPrecisionError

      public static ArithmeticException unscaledValueTooLargeForPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
    • cannotChangeDecimalPrecisionError

      public static ArithmeticException cannotChangeDecimalPrecisionError(Decimal value, int decimalPrecision, int decimalScale, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
    • invalidInputInCastToNumberError

      public static org.apache.spark.SparkNumberFormatException invalidInputInCastToNumberError(DataType to, org.apache.spark.unsafe.types.UTF8String s, org.apache.spark.sql.catalyst.trees.SQLQueryContext context)
    • ambiguousColumnOrFieldError

      public static Throwable ambiguousColumnOrFieldError(scala.collection.Seq<String> name, int numMatches, org.apache.spark.sql.catalyst.trees.Origin context)
    • castingCauseOverflowError

      public static ArithmeticException castingCauseOverflowError(String t, DataType from, DataType to)
    • failedParsingStructTypeError

      public static org.apache.spark.SparkRuntimeException failedParsingStructTypeError(String raw)
    • fieldIndexOnRowWithoutSchemaError

      public static org.apache.spark.SparkUnsupportedOperationException fieldIndexOnRowWithoutSchemaError()
    • valueIsNullError

      public static Throwable valueIsNullError(int index)
    • charOrVarcharTypeAsStringUnsupportedError

      public static Throwable charOrVarcharTypeAsStringUnsupportedError()
    • userSpecifiedSchemaUnsupportedError

      public static Throwable userSpecifiedSchemaUnsupportedError(String operation)
    • toSQLId

      public static String toSQLId(String parts)
    • toSQLId

      public static String toSQLId(scala.collection.Seq<String> parts)
    • toSQLStmt

      public static String toSQLStmt(String text)
    • toSQLConf

      public static String toSQLConf(String conf)
    • toSQLType

      public static String toSQLType(String text)
    • toSQLType

      public static String toSQLType(org.apache.spark.sql.types.AbstractDataType t)
    • toSQLValue

      public static String toSQLValue(String value)
    • toSQLValue

      public static String toSQLValue(org.apache.spark.unsafe.types.UTF8String value)
    • toSQLValue

      public static String toSQLValue(short value)
    • toSQLValue

      public static String toSQLValue(int value)
    • toSQLValue

      public static String toSQLValue(long value)
    • toSQLValue

      public static String toSQLValue(float value)
    • toSQLValue

      public static String toSQLValue(double value)
    • getSummary

      public static String getSummary(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)
    • getQueryContext

      public static QueryContext[] getQueryContext(org.apache.spark.sql.catalyst.trees.SQLQueryContext sqlContext)