HADOOP-17622. Avoid usage of deprecated IOUtils#cleanup API. (#2862)
Signed-off-by: Takanobu Asanuma <tasanuma@apache.org>
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java
index 1384bb6..e7b72a9 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java
@@ -312,7 +312,7 @@
.new LocalFSFileInputStream(path), 1024);
assertNotNull(bis.getFileDescriptor());
} finally {
- IOUtils.cleanup(null, bis);
+ IOUtils.cleanupWithLogger(null, bis);
}
}
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java
index b20653e..6788940 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java
@@ -115,7 +115,7 @@
IOUtils.copyBytes(is, baos, 1024, true);
outFileContent = new String(baos.toByteArray(), "UTF-8");
} finally {
- IOUtils.cleanup(null, baos, is);
+ IOUtils.cleanupWithLogger(null, baos, is);
}
// Check the out file content. Should be something like the following:
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java
index 466c83e..61cbd85 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java
@@ -759,6 +759,6 @@
readerThread.join();
Assert.assertFalse(failed.get());
Assert.assertEquals(3, bytesRead.get());
- IOUtils.cleanup(null, socks);
+ IOUtils.cleanupWithLogger(null, socks);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java
index 3a54361..d3212b6 100755
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java
@@ -618,8 +618,8 @@
private synchronized void cleanUp() {
shouldServiceRun = false;
- IOUtils.cleanup(null, bpNamenode);
- IOUtils.cleanup(null, lifelineSender);
+ IOUtils.cleanupWithLogger(null, bpNamenode);
+ IOUtils.cleanupWithLogger(null, lifelineSender);
bpos.shutdownActor(this);
}
@@ -992,7 +992,7 @@
} catch (InterruptedException e) {
Thread.currentThread().interrupt();
}
- IOUtils.cleanup(null, lifelineNamenode);
+ IOUtils.cleanupWithLogger(null, lifelineNamenode);
}
@Override
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
index 3f1773c..52f25d7 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
@@ -369,7 +369,7 @@
streams.close();
}
if (replicaHandler != null) {
- IOUtils.cleanup(null, replicaHandler);
+ IOUtils.cleanupWithLogger(null, replicaHandler);
replicaHandler = null;
}
if (measuredFlushTime) {
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java
index 485cf00..6dcfad4 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java
@@ -252,7 +252,7 @@
if (!success) {
// If we didn't create a new VolumeScanner object, we don't
// need this reference to the volume.
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java
index 278ee09..c43fc3d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java
@@ -432,7 +432,7 @@
blk.getBlockId(), dnR.getDatanodeUuid(), success));
}
if (fis != null) {
- IOUtils.cleanup(null, fis);
+ IOUtils.cleanupWithLogger(null, fis);
}
}
}
@@ -555,7 +555,7 @@
LOG.warn("Failed to shut down socket in error handler", e);
}
}
- IOUtils.cleanup(null, shmInfo);
+ IOUtils.cleanupWithLogger(null, shmInfo);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java
index 6bc25eb..0367b4a 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java
@@ -451,7 +451,7 @@
} catch (IOException e) {
resultHandler.handle(block, e);
} finally {
- IOUtils.cleanup(null, blockSender);
+ IOUtils.cleanupWithLogger(null, blockSender);
}
metrics.incrBlockVerificationFailures();
return -1;
@@ -674,13 +674,13 @@
// Save the current position of all block iterators and close them.
for (BlockIterator iter : blockIters) {
saveBlockIterator(iter);
- IOUtils.cleanup(null, iter);
+ IOUtils.cleanupWithLogger(null, iter);
}
} finally {
VolumeScannerCBInjector.get().terminationCallBack(this);
// When the VolumeScanner exits, release the reference we were holding
// on the volume. This will allow the volume to be removed later.
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
}
}
@@ -767,7 +767,7 @@
if (iter.getBlockPoolId().equals(bpid)) {
LOG.trace("{}: disabling scanning on block pool {}", this, bpid);
i.remove();
- IOUtils.cleanup(null, iter);
+ IOUtils.cleanupWithLogger(null, iter);
if (curBlockIter == iter) {
curBlockIter = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
index d077d21..997a6d9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
@@ -234,7 +234,7 @@
}
}), MoreExecutors.directExecutor());
} else {
- IOUtils.cleanup(null, reference);
+ IOUtils.cleanupWithLogger(null, reference);
if (numVolumes.decrementAndGet() == 0) {
latch.countDown();
}
@@ -311,7 +311,7 @@
);
return true;
} else {
- IOUtils.cleanup(null, volumeReference);
+ IOUtils.cleanupWithLogger(null, volumeReference);
}
return false;
}
@@ -404,7 +404,7 @@
}
private void cleanup() {
- IOUtils.cleanup(null, reference);
+ IOUtils.cleanupWithLogger(null, reference);
invokeCallback();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java
index b1ad03f..54302e3 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java
@@ -158,7 +158,7 @@
return peer;
} finally {
if (!success) {
- IOUtils.cleanup(null, peer);
+ IOUtils.cleanupWithLogger(null, peer);
IOUtils.closeSocket(sock);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
index f40315a..f8bd8c0 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
@@ -130,7 +130,7 @@
dataInFd = null;
}
if (volumeRef != null) {
- IOUtils.cleanup(null, volumeRef);
+ IOUtils.cleanupWithLogger(null, volumeRef);
volumeRef = null;
}
// throw IOException if there is any
@@ -146,7 +146,7 @@
dataInFd = null;
IOUtils.closeStream(checksumIn);
checksumIn = null;
- IOUtils.cleanup(null, volumeRef);
+ IOUtils.cleanupWithLogger(null, volumeRef);
volumeRef = null;
}
}
\ No newline at end of file
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
index 81213a0..2a89a80d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
@@ -335,7 +335,7 @@
+ block.getLocalBlock() + " URI " + replicaToDelete.getBlockURI());
}
updateDeletedBlockId(block);
- IOUtils.cleanup(null, volumeRef);
+ IOUtils.cleanupWithLogger(null, volumeRef);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
index 41791bb..d06d3cf 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
@@ -949,11 +949,11 @@
return new ReplicaInputStreams(
blockInStream, metaInStream, ref, datanode.getFileIoProvider());
} catch (IOException e) {
- IOUtils.cleanup(null, blockInStream);
+ IOUtils.cleanupWithLogger(null, blockInStream);
throw e;
}
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
}
@@ -1421,7 +1421,7 @@
replica = append(b.getBlockPoolId(), replicaInfo, newGS,
b.getNumBytes());
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
return new ReplicaHandler(replica, ref);
@@ -1553,7 +1553,7 @@
replica = (ReplicaInPipeline) replicaInfo;
}
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
return new ReplicaHandler(replica, ref);
@@ -1648,7 +1648,7 @@
+ " for block " + b.getBlockId());
}
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
@@ -1752,7 +1752,7 @@
// bump the replica's generation stamp to newGS
rbw.getReplicaInfo().bumpReplicaGS(newGS);
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
return new ReplicaHandler(rbw, ref);
@@ -1905,7 +1905,7 @@
try {
newReplicaInfo = v.createTemporary(b);
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java
index 812a7bf..621c273 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java
@@ -117,7 +117,7 @@
}
return raf.getFD();
} catch(IOException ioe) {
- IOUtils.cleanup(null, raf);
+ IOUtils.cleanupWithLogger(null, raf);
throw ioe;
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java
index a87f13c..9400c7c 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java
@@ -357,7 +357,7 @@
} else {
// If the volume is not put into a volume scanner, it does not need to
// hold the reference.
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
}
// If the volume is used to replace a failed volume, it needs to reset the
// volume failure info for this volume.
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java
index affba02..ccc233e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java
@@ -76,7 +76,7 @@
+ " bytes still to be flushed and cannot be closed.");
}
- IOUtils.cleanup(null, bufCurrent, bufReady);
+ IOUtils.cleanupWithLogger(null, bufCurrent, bufReady);
bufCurrent = bufReady = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java
index 1784ea2..f5967e1 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java
@@ -202,7 +202,7 @@
blockFile);
return 0;
} finally {
- IOUtils.cleanup(null, metaStream, dataStream, checksumStream);
+ IOUtils.cleanupWithLogger(null, metaStream, dataStream, checksumStream);
}
}
}
@@ -287,7 +287,7 @@
+ " saved metadata to meta file " + outFile);
return 0;
} finally {
- IOUtils.cleanup(null, metaOut);
+ IOUtils.cleanupWithLogger(null, metaOut);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java
index cd4047d..ccab7b0 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java
@@ -319,10 +319,10 @@
@Override
public void close() throws IOException {
if (batch != null) {
- IOUtils.cleanup(null, batch);
+ IOUtils.cleanupWithLogger(null, batch);
batch = null;
}
- IOUtils.cleanup(null, db);
+ IOUtils.cleanupWithLogger(null, db);
db = null;
}
@@ -388,13 +388,13 @@
dirMap = new LevelDBStore(new File(dbDir, "dirMap"));
} catch (IOException e) {
LOG.error("Failed to open LevelDBs", e);
- IOUtils.cleanup(null, this);
+ IOUtils.cleanupWithLogger(null, this);
}
}
@Override
public void close() throws IOException {
- IOUtils.cleanup(null, dirChildMap, dirMap);
+ IOUtils.cleanupWithLogger(null, dirChildMap, dirMap);
dirChildMap = null;
dirMap = null;
}
@@ -515,7 +515,7 @@
@Override
public void close() throws IOException {
out.flush();
- IOUtils.cleanup(null, metadataMap);
+ IOUtils.cleanupWithLogger(null, metadataMap);
}
void append(StringBuffer buffer, int field) {
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java
index f4c264c..99b1ddbb 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java
@@ -841,7 +841,7 @@
if (buf2 != null) {
fsIn2.releaseBuffer(buf2);
}
- IOUtils.cleanup(null, fsIn, fsIn2);
+ IOUtils.cleanupWithLogger(null, fsIn, fsIn2);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java
index ef4c04d..feb77f8 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java
@@ -87,7 +87,7 @@
Assert.assertEquals(b, b2);
} finally {
if (stream != null) {
- IOUtils.cleanup(null, stream);
+ IOUtils.cleanupWithLogger(null, stream);
}
if (cluster != null) {
cluster.shutdown();
@@ -122,7 +122,7 @@
}
} finally {
for (FSDataInputStream stream : streams) {
- IOUtils.cleanup(null, stream);
+ IOUtils.cleanupWithLogger(null, stream);
}
if (cluster != null) {
cluster.shutdown();
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java
index a6409fd..96d15e5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java
@@ -96,7 +96,7 @@
@AfterClass
public static void shutdown() throws Exception {
- IOUtils.cleanup(null, hdfs, hdfsAsUser1, hdfsAsUser2);
+ IOUtils.cleanupWithLogger(null, hdfs, hdfsAsUser1, hdfsAsUser2);
if (cluster != null) {
cluster.shutdown();
}
@@ -121,7 +121,7 @@
h.close();
h = null;
} finally {
- IOUtils.cleanup(null, h);
+ IOUtils.cleanupWithLogger(null, h);
}
}
@@ -500,7 +500,7 @@
o.close();
o = null;
} finally {
- IOUtils.cleanup(null, o);
+ IOUtils.cleanupWithLogger(null, o);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java
index 0bf21ee..9881f92 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java
@@ -227,7 +227,7 @@
IOUtils.copyBytes(stm, new IOUtils.NullOutputStream(), 1024);
}
} finally {
- IOUtils.cleanup(null, stms);
+ IOUtils.cleanupWithLogger(null, stms);
}
assertEquals(5, peerCache.size());
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java
index d700765..711291c 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java
@@ -177,7 +177,7 @@
blocks = fileSystem.dfs.getLocatedBlocks(path.toString(), 0);
assertEquals(3, blocks.getLocatedBlocks().size());
} finally {
- IOUtils.cleanup(null, stm, fileSystem);
+ IOUtils.cleanupWithLogger(null, stm, fileSystem);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java
index ac0994d..ac3c122 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java
@@ -363,8 +363,8 @@
assertTrue(false);
} finally {
Mockito.reset(injector);
- IOUtils.cleanup(null, input);
- IOUtils.cleanup(null, output);
+ IOUtils.cleanupWithLogger(null, input);
+ IOUtils.cleanupWithLogger(null, output);
fileSys.close();
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java
index 46fab47..a716335 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java
@@ -741,7 +741,7 @@
// do checkpoint in SNN again
snn.doCheckpoint();
} finally {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (snn != null) {
snn.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java
index db97c02..fcf1333 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java
@@ -273,7 +273,7 @@
@After
public void shutdown() throws IOException {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java
index cce1454..4728884 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java
@@ -229,7 +229,7 @@
@AfterClass
public static void destroy() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java
index 5503238..52abbe9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java
@@ -54,7 +54,7 @@
fs.close();
fs = null;
} finally {
- IOUtils.cleanup(null, os, fs);
+ IOUtils.cleanupWithLogger(null, os, fs);
cluster.shutdown();
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java
index d6612c1..3dd0b7e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java
@@ -256,7 +256,7 @@
} catch (SocketTimeoutException e) {
GenericTestUtils.assertExceptionContains("Read timed out", e);
} finally {
- IOUtils.cleanup(null, socket, serverSocket);
+ IOUtils.cleanupWithLogger(null, socket, serverSocket);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java
index 539f216..9168057 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java
@@ -173,7 +173,7 @@
lastRecoveredTxn = elis.getLastTxId();
}
} finally {
- IOUtils.cleanup(null, streams.toArray(new Closeable[0]));
+ IOUtils.cleanupWithLogger(null, streams.toArray(new Closeable[0]));
}
return lastRecoveredTxn;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java
index 35bae49..eb64cf5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java
@@ -163,7 +163,7 @@
@After
public void shutdown() throws IOException {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
@@ -213,7 +213,7 @@
* @throws IOException if there is an I/O error
*/
private void restartNameNode() throws IOException {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
cluster.restartNameNode();
fs = cluster.getFileSystem();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java
index 67b41f8..3f99f1b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java
@@ -479,7 +479,7 @@
assertEquals("The block should be only on 1 datanode ", 1,
locatedBlocks1.get(0).getLocations().length);
} finally {
- IOUtils.cleanup(null, client);
+ IOUtils.cleanupWithLogger(null, client);
cluster.shutdown();
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java
index 459ef88..011df46 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java
@@ -332,7 +332,7 @@
stats = tracker.getStats(fadvisedFileName);
stats.assertNotDroppedInRange(0, TEST_PATH_LEN - WRITE_PACKET_SIZE);
} finally {
- IOUtils.cleanup(null, fis);
+ IOUtils.cleanupWithLogger(null, fis);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java
index fec04be..d95e76f 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java
@@ -230,7 +230,7 @@
}
void shutdownCluster() throws Exception {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java
index 8a2c0e2..3a9ad25 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java
@@ -108,7 +108,8 @@
@After
public void destroyFileSystems() {
- IOUtils.cleanup(null, fs, fsAsBruce, fsAsDiana, fsAsSupergroupMember);
+ IOUtils.cleanupWithLogger(null, fs, fsAsBruce, fsAsDiana,
+ fsAsSupergroupMember);
fs = fsAsBruce = fsAsDiana = fsAsSupergroupMember = fsAsBob = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java
index 019531c..44e86ae 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java
@@ -127,7 +127,7 @@
@After
public void destroyFileSystems() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
fs = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java
index f45728f..8e01f73 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java
@@ -55,7 +55,7 @@
@After
public void shutdown() throws Exception {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java
index d17a36f..ef91fc3 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java
@@ -206,7 +206,7 @@
assertTrue(blocks[0].isStriped());
checkStripedBlockUC((BlockInfoStriped) fileNode.getLastBlock(), false);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
@@ -271,7 +271,7 @@
assertArrayEquals(indices, blockIndices);
assertArrayEquals(expectedDNs, datanodes);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
@@ -327,7 +327,7 @@
assertTrue(storageIDs.contains(newstorage.getStorageID()));
}
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
// 3. restart the namenode. mimic the full block reports and check the
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java
index 77a3268..f7b1ea5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java
@@ -95,7 +95,7 @@
dfs.getClient().getClientName(), previous, fileNode.getId());
Assert.assertTrue("should complete successfully", complete);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
index dbf65fa..ae27730 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
@@ -1548,7 +1548,8 @@
LOG.error("edit log failover didn't work", e);
fail("Edit log failover didn't work");
} finally {
- IOUtils.cleanup(null, streams.toArray(new EditLogInputStream[0]));
+ IOUtils.cleanupWithLogger(null,
+ streams.toArray(new EditLogInputStream[0]));
}
}
@@ -1598,7 +1599,8 @@
LOG.error("edit log failover didn't work", e);
fail("Edit log failover didn't work");
} finally {
- IOUtils.cleanup(null, streams.toArray(new EditLogInputStream[0]));
+ IOUtils.cleanupWithLogger(null,
+ streams.toArray(new EditLogInputStream[0]));
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java
index 67b48b1..9ea4548 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java
@@ -154,7 +154,7 @@
editLogStream.abort();
editLogStream.abort();
} finally {
- IOUtils.cleanup(null, editLogStream);
+ IOUtils.cleanupWithLogger(null, editLogStream);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java
index 2c9075e..b27a00d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java
@@ -568,7 +568,7 @@
}
}
} finally {
- IOUtils.cleanup(null, in);
+ IOUtils.cleanupWithLogger(null, in);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java
index 0c9b499..c88570b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java
@@ -212,7 +212,7 @@
line = rdr.readLine();
}
} finally {
- IOUtils.cleanup(null, rdr, isr, fis);
+ IOUtils.cleanupWithLogger(null, rdr, isr, fis);
}
}
@@ -271,7 +271,7 @@
line = rdr.readLine();
}
} finally {
- IOUtils.cleanup(null, rdr, isr, fis);
+ IOUtils.cleanupWithLogger(null, rdr, isr, fis);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java
index 38b98a4..e89ed8d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java
@@ -140,7 +140,7 @@
Assert.assertEquals(cellSize * groupSize,
actualDiskUsed);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java
index 5064a03..c527837 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java
@@ -48,7 +48,7 @@
@After
public void shutdown() throws Exception {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java
index 9074247..e3e934b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java
@@ -1315,7 +1315,7 @@
} catch (Exception e) {
LOG.info("Got Exception while calling " + op.name, e);
} finally {
- IOUtils.cleanup(null, op.client);
+ IOUtils.cleanupWithLogger(null, op.client);
}
}
}.start();
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java
index ea9c5e2..511607e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java
@@ -86,7 +86,7 @@
@AfterClass
public static void shutdown() throws Exception {
- IOUtils.cleanup(null, hdfs, fsAsBruce, fsAsDiana);
+ IOUtils.cleanupWithLogger(null, hdfs, fsAsBruce, fsAsDiana);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java
index 4b957bf9..2c93e12 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java
@@ -79,7 +79,7 @@
@AfterClass
public static void shutdown() throws Exception {
- IOUtils.cleanup(null, hdfs);
+ IOUtils.cleanupWithLogger(null, hdfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java
index bce92ef..144f990 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java
@@ -142,7 +142,7 @@
fos = null;
}
} finally {
- IOUtils.cleanup(null, fos);
+ IOUtils.cleanupWithLogger(null, fos);
FileUtil.setWritable(TEST_DIR, true);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java
index 2212939..1fc54d6 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java
@@ -95,7 +95,7 @@
final Path dir2 = new Path("/test2");
Assert.assertTrue(fs.mkdirs(dir2));
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -130,7 +130,7 @@
verify(fs).renewDelegationToken(token);
verify(fs).cancelDelegationToken(token);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -199,7 +199,7 @@
Exception unwrapped = re.unwrapRemoteException(StandbyException.class);
Assert.assertTrue(unwrapped instanceof StandbyException);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -238,7 +238,7 @@
IOUtils.readFully(in, buf, 0, buf.length);
Assert.assertArrayEquals(data, buf);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -264,7 +264,7 @@
fs = (WebHdfsFileSystem)FileSystem.get(WEBHDFS_URI, conf);
Assert.assertEquals(2, fs.getResolvedNNAddr().length);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -304,7 +304,7 @@
} catch (IOException e) {
result = false;
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
}
synchronized (TestWebHDFSForHA.this) {
resultMap.put("mkdirs", result);
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java
index d6bbc25..c80a79b 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java
@@ -83,7 +83,7 @@
@After
public void tearDown() throws Exception {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java
index cca1c53..50f5823 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java
@@ -153,7 +153,7 @@
@After
public void tearDown() throws Exception {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java
index 49613ba..38b7933 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java
@@ -96,7 +96,7 @@
@AfterClass
public static void shutdown() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java
index e0e103b..b16f6b2 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java
@@ -81,7 +81,7 @@
@AfterClass
public static void shutdown() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java
index 77d428b..36c6e6a 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java
@@ -114,7 +114,7 @@
@AfterClass
public static void shutdown() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java
index 1c92a9c5..389fe36 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java
@@ -72,7 +72,7 @@
recordInExpectedValues(path);
}
finally {
- IOUtils.cleanup(null, fileSystem);
+ IOUtils.cleanupWithLogger(null, fileSystem);
}
}
@@ -85,7 +85,7 @@
recordInExpectedValues(path);
}
finally {
- IOUtils.cleanup(null, fileSystem, outputStream);
+ IOUtils.cleanupWithLogger(null, fileSystem, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java
index bf31657..780d82d 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java
@@ -261,7 +261,7 @@
System.out.println(fileStatus.getReplication());
}
finally {
- IOUtils.cleanup(null, outputStream);
+ IOUtils.cleanupWithLogger(null, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java
index 5315137..622e391 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java
@@ -91,7 +91,7 @@
return size;
}
finally {
- IOUtils.cleanup(null, fileSystem, outputStream);
+ IOUtils.cleanupWithLogger(null, fileSystem, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java
index 61a1745..6a310ca 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java
@@ -100,7 +100,7 @@
new Path(path))[0].getPath().toString());
}
finally {
- IOUtils.cleanup(null, fileSystem, outputStream);
+ IOUtils.cleanupWithLogger(null, fileSystem, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java
index 2cb806e..f95c4b3 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java
@@ -79,7 +79,7 @@
@Override
public void close() throws IOException {
- IOUtils.cleanup(null, src);
+ IOUtils.cleanupWithLogger(null, src);
factory.close();
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java
index bdbfc3b..fe3b5d3 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java
@@ -112,7 +112,7 @@
} catch (InterruptedException e) {
// exit thread; ignore any jobs remaining in the trace
} finally {
- IOUtils.cleanup(null, jobProducer);
+ IOUtils.cleanupWithLogger(null, jobProducer);
}
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java
index e8c7d61..cb05ab6 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java
@@ -143,7 +143,7 @@
} catch (InterruptedException e) {
return;
} finally {
- IOUtils.cleanup(null, jobProducer);
+ IOUtils.cleanupWithLogger(null, jobProducer);
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java
index bd15c29..4e7fc9c 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java
@@ -247,7 +247,7 @@
LOG.error("[STRESS] Interrupted in the main block!", e);
return;
} finally {
- IOUtils.cleanup(null, jobProducer);
+ IOUtils.cleanupWithLogger(null, jobProducer);
}
}
}
diff --git a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java
index 424405a..683bcc8 100644
--- a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java
+++ b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java
@@ -470,7 +470,7 @@
next = heap.poll();
}
} finally {
- IOUtils.cleanup(null, reader);
+ IOUtils.cleanupWithLogger(null, reader);
if (outGen != null) {
outGen.close();
}