@Override public void close() throws IOException { try { // Only publish stats if this operator's flag was set to gather stats if (!exception) { publishStats(); } } catch (HiveException e) { this.exception = true; throw new RuntimeException(e); } finally { MapredContext.close(); } }
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
MapredContext.close(); Utilities.clearWorkMap(jc);
@Override void close() { if (cache != null && cacheKey != null) { cache.release(cacheKey); } // check if there are IOExceptions if (!isAborted()) { setAborted(execContext.getIoCxt().getIOExceptions()); } // detecting failed executions by exceptions thrown by the operator tree try { if (mergeOp == null || mfWork == null) { return; } boolean abort = isAborted(); mergeOp.close(abort); ExecMapper.ReportStats rps = new ExecMapper.ReportStats(reporter, jconf); mergeOp.preorderMap(rps); } catch (Exception e) { if (!isAborted()) { // signal new failure to map-reduce l4j.error("Hit error while closing operators - failing tree"); throw new RuntimeException("Hive Runtime Error while closing operators", e); } } finally { Utilities.clearWorkMap(jconf); MapredContext.close(); } }
@Override void close() { if (cache != null && cacheKey != null) { cache.release(cacheKey); } // check if there are IOExceptions if (!isAborted()) { setAborted(execContext.getIoCxt().getIOExceptions()); } // detecting failed executions by exceptions thrown by the operator tree try { if (mergeOp == null || mfWork == null) { return; } boolean abort = isAborted(); mergeOp.close(abort); ExecMapper.ReportStats rps = new ExecMapper.ReportStats(reporter, jconf); mergeOp.preorderMap(rps); } catch (Exception e) { if (!isAborted()) { // signal new failure to map-reduce l4j.error("Hit error while closing operators - failing tree"); throw new RuntimeException("Hive Runtime Error while closing operators", e); } } finally { Utilities.clearWorkMap(jconf); MapredContext.close(); } }
MapredContext.close();
MapredContext.close();
MapredContext.close();
MapredContext.close();
@Override public void close() throws IOException { try { // Only publish stats if this operator's flag was set to gather stats if (!exception) { publishStats(); } } catch (HiveException e) { this.exception = true; throw new RuntimeException(e); } finally { MapredContext.close(); } }
MapredContext.close(); Utilities.clearWorkMap();
@Override void close() { if (cache != null && cacheKey != null) { cache.release(cacheKey); } // check if there are IOExceptions if (!abort) { abort = execContext.getIoCxt().getIOExceptions(); } // detecting failed executions by exceptions thrown by the operator tree try { if (mergeOp == null || mfWork == null) { return; } mergeOp.close(abort); ExecMapper.ReportStats rps = new ExecMapper.ReportStats(reporter, jconf); mergeOp.preorderMap(rps); } catch (Exception e) { if (!abort) { // signal new failure to map-reduce l4j.error("Hit error while closing operators - failing tree"); throw new RuntimeException("Hive Runtime Error while closing operators", e); } } finally { Utilities.clearWorkMap(); MapredContext.close(); } }
MapredContext.close(); Utilities.clearWorkMap();
MapredContext.close(); Utilities.clearWorkMap();