Throwables.getRootCause(sparkJobException).getMessage()); setException(he); setException(monitorError);
Throwable error = sparkJobStatus.getError(); if (error != null) { setException(error);
/** * Close will move the temp files into the right place for the fetch * task. If the job has failed it will clean up the files. */ private int close(int rc) { try { List<BaseWork> ws = work.getAllWork(); for (BaseWork w: ws) { for (Operator<?> op: w.getAllOperators()) { op.jobClose(conf, rc == 0); } } } catch (Exception e) { // jobClose needs to execute successfully otherwise fail task if (rc == 0) { rc = 3; String mesg = "Job Commit failed with exception '" + Utilities.getNameMessage(e) + "'"; console.printError(mesg, "\n" + StringUtils.stringifyException(e)); setException(e); } } return rc; }
/** * Close will move the temp files into the right place for the fetch * task. If the job has failed it will clean up the files. */ private int close(int rc) { try { List<BaseWork> ws = work.getAllWork(); for (BaseWork w: ws) { for (Operator<?> op: w.getAllOperators()) { op.jobClose(conf, rc == 0); } } } catch (Exception e) { // jobClose needs to execute successfully otherwise fail task if (rc == 0) { rc = 3; String mesg = "Job Commit failed with exception '" + Utilities.getNameMessage(e) + "'"; console.printError(mesg, "\n" + StringUtils.stringifyException(e)); setException(e); } } return rc; }
} catch (Exception e) { LOG.error("Failed to execute Spark task \"" + getId() + "\"", e); setException(e); if (e instanceof HiveException) { HiveException he = (HiveException) e;
setException(e); rc = 1; } finally {