public interface SparkErrorUtils
extends org.apache.spark.internal.Logging
| Modifier and Type | Method and Description | 
|---|---|
| String | stackTraceToString(Throwable t) | 
| <T> T | tryOrIOException(scala.Function0<T> block)Execute a block of code that returns a value, re-throwing any non-fatal uncaught
 exceptions as IOException. | 
| <R extends java.io.Closeable,T> | tryWithResource(scala.Function0<R> createResource,
               scala.Function1<R,T> f) | 
| <T> T | tryWithSafeFinally(scala.Function0<T> block,
                  scala.Function0<scala.runtime.BoxedUnit> finallyBlock)Execute a block of code, then a finally block, but if exceptions happen in
 the finally block, do not suppress the original exception. | 
$init$, initializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, initLock, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, org$apache$spark$internal$Logging$$log__$eq, org$apache$spark$internal$Logging$$log_, uninitialize<T> T tryOrIOException(scala.Function0<T> block)
block - (undocumented)<R extends java.io.Closeable,T> T tryWithResource(scala.Function0<R> createResource,
                                                  scala.Function1<R,T> f)
<T> T tryWithSafeFinally(scala.Function0<T> block,
                         scala.Function0<scala.runtime.BoxedUnit> finallyBlock)
 This is primarily an issue with finally { out.close() } blocks, where
 close needs to be called to clean up out, but if an exception happened
 in out.write, it's likely out may be corrupted and out.close will
 fail as well. This would then suppress the original/likely more meaningful
 exception from the original out.write call.
block - (undocumented)finallyBlock - (undocumented)String stackTraceToString(Throwable t)