| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.hadoop.hdfs.tools; |
| |
| import java.io.BufferedReader; |
| import java.io.IOException; |
| import java.io.InputStream; |
| import java.io.InputStreamReader; |
| import java.io.PrintStream; |
| import java.net.InetSocketAddress; |
| import java.net.URL; |
| import java.net.URLConnection; |
| import java.net.URLEncoder; |
| import java.security.PrivilegedExceptionAction; |
| |
| import org.apache.hadoop.classification.InterfaceAudience; |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.conf.Configured; |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.hdfs.server.namenode.NameNode; |
| import org.apache.hadoop.hdfs.server.namenode.NamenodeFsck; |
| import org.apache.hadoop.hdfs.DFSUtil; |
| import org.apache.hadoop.hdfs.DistributedFileSystem; |
| import org.apache.hadoop.hdfs.HdfsConfiguration; |
| import org.apache.hadoop.security.Krb5AndCertsSslSocketConnector; |
| import org.apache.hadoop.security.SecurityUtil; |
| import org.apache.hadoop.security.UserGroupInformation; |
| import org.apache.hadoop.util.StringUtils; |
| import org.apache.hadoop.util.Tool; |
| import org.apache.hadoop.util.ToolRunner; |
| |
| /** |
| * This class provides rudimentary checking of DFS volumes for errors and |
| * sub-optimal conditions. |
| * <p>The tool scans all files and directories, starting from an indicated |
| * root path. The following abnormal conditions are detected and handled:</p> |
| * <ul> |
| * <li>files with blocks that are completely missing from all datanodes.<br/> |
| * In this case the tool can perform one of the following actions: |
| * <ul> |
| * <li>none ({@link org.apache.hadoop.hdfs.server.namenode.NamenodeFsck#FIXING_NONE})</li> |
| * <li>move corrupted files to /lost+found directory on DFS |
| * ({@link org.apache.hadoop.hdfs.server.namenode.NamenodeFsck#FIXING_MOVE}). Remaining data blocks are saved as a |
| * block chains, representing longest consecutive series of valid blocks.</li> |
| * <li>delete corrupted files ({@link org.apache.hadoop.hdfs.server.namenode.NamenodeFsck#FIXING_DELETE})</li> |
| * </ul> |
| * </li> |
| * <li>detect files with under-replicated or over-replicated blocks</li> |
| * </ul> |
| * Additionally, the tool collects a detailed overall DFS statistics, and |
| * optionally can print detailed statistics on block locations and replication |
| * factors of each file. |
| * The tool also provides and option to filter open files during the scan. |
| * |
| */ |
| @InterfaceAudience.Private |
| public class DFSck extends Configured implements Tool { |
| static{ |
| HdfsConfiguration.init(); |
| } |
| |
| private final UserGroupInformation ugi; |
| private final PrintStream out; |
| |
| /** |
| * Filesystem checker. |
| * @param conf current Configuration |
| */ |
| public DFSck(Configuration conf) throws IOException { |
| this(conf, System.out); |
| } |
| |
| public DFSck(Configuration conf, PrintStream out) throws IOException { |
| super(conf); |
| this.ugi = UserGroupInformation.getCurrentUser(); |
| this.out = out; |
| } |
| |
| /** |
| * Print fsck usage information |
| */ |
| static void printUsage() { |
| System.err.println("Usage: DFSck <path> [-list-corruptfileblocks | " + |
| "[-move | -delete | -openforwrite] " + |
| "[-files [-blocks [-locations | -racks]]]]"); |
| System.err.println("\t<path>\tstart checking from this path"); |
| System.err.println("\t-move\tmove corrupted files to /lost+found"); |
| System.err.println("\t-delete\tdelete corrupted files"); |
| System.err.println("\t-files\tprint out files being checked"); |
| System.err.println("\t-openforwrite\tprint out files opened for write"); |
| System.err.println("\t-list-corruptfileblocks\tprint out list of missing " |
| + "blocks and files they belong to"); |
| System.err.println("\t-blocks\tprint out block report"); |
| System.err.println("\t-locations\tprint out locations for every block"); |
| System.err.println("\t-racks\tprint out network topology for data-node locations"); |
| System.err.println("\t\tBy default fsck ignores files opened for write, " + |
| "use -openforwrite to report such files. They are usually " + |
| " tagged CORRUPT or HEALTHY depending on their block " + |
| "allocation status"); |
| ToolRunner.printGenericCommandUsage(System.err); |
| } |
| /** |
| * @param args |
| */ |
| public int run(final String[] args) throws IOException { |
| if (args.length == 0) { |
| printUsage(); |
| return -1; |
| } |
| |
| try { |
| return UserGroupInformation.getCurrentUser().doAs( |
| new PrivilegedExceptionAction<Integer>() { |
| @Override |
| public Integer run() throws Exception { |
| return doWork(args); |
| } |
| }); |
| } catch (InterruptedException e) { |
| throw new IOException(e); |
| } |
| } |
| |
| /* |
| * To get the list, we need to call iteratively until the server says |
| * there is no more left. |
| */ |
| private Integer listCorruptFileBlocks(String dir, String baseUrl) |
| throws IOException { |
| int errCode = -1; |
| int numCorrupt = 0; |
| String lastBlock = null; |
| final String noCorruptLine = "has no CORRUPT files"; |
| final String noMoreCorruptLine = "has no more CORRUPT files"; |
| boolean allDone = false; |
| while (!allDone) { |
| final StringBuffer url = new StringBuffer(baseUrl); |
| if (lastBlock != null) { |
| url.append("&startblockafter=").append(lastBlock); |
| } |
| URL path = new URL(url.toString()); |
| SecurityUtil.fetchServiceTicket(path); |
| URLConnection connection = path.openConnection(); |
| InputStream stream = connection.getInputStream(); |
| BufferedReader input = new BufferedReader(new InputStreamReader( |
| stream, "UTF-8")); |
| try { |
| String line = null; |
| while ((line = input.readLine()) != null) { |
| if ((line.endsWith(noCorruptLine)) || |
| (line.endsWith(noMoreCorruptLine)) || |
| (line.endsWith(NamenodeFsck.NONEXISTENT_STATUS))) { |
| allDone = true; |
| break; |
| } |
| if ((line.isEmpty()) |
| || (line.startsWith("FSCK started by")) |
| || (line.startsWith("The filesystem under path"))) |
| continue; |
| numCorrupt++; |
| if (numCorrupt == 1) { |
| out.println("The list of corrupt files under path '" |
| + dir + "' are:"); |
| } |
| out.println(line); |
| try { |
| // Get the block # that we need to send in next call |
| lastBlock = line.split("\t")[0]; |
| } catch (Exception e) { |
| allDone = true; |
| break; |
| } |
| } |
| } finally { |
| input.close(); |
| } |
| } |
| out.println("The filesystem under path '" + dir + "' has " |
| + numCorrupt + " CORRUPT files"); |
| if (numCorrupt == 0) |
| errCode = 0; |
| return errCode; |
| } |
| |
| /** |
| * Derive the namenode http address from the current file system, |
| * either default or as set by "-fs" in the generic options. |
| * @return Returns http address or null if failure. |
| */ |
| private String getCurrentNamenodeAddress() { |
| //String nnAddress = null; |
| Configuration conf = getConf(); |
| |
| //get the filesystem object to verify it is an HDFS system |
| FileSystem fs; |
| try { |
| fs = FileSystem.get(conf); |
| } catch (IOException ioe) { |
| System.err.println("FileSystem is inaccessible due to:\n" |
| + StringUtils.stringifyException(ioe)); |
| return null; |
| } |
| if (!(fs instanceof DistributedFileSystem)) { |
| System.err.println("FileSystem is " + fs.getUri()); |
| return null; |
| } |
| DistributedFileSystem dfs = (DistributedFileSystem) fs; |
| |
| // Derive the nameservice ID from the filesystem URI. |
| // The URI may have been provided by a human, and the server name may be |
| // aliased, so compare InetSocketAddresses instead of URI strings, and |
| // test against both possible variants of RPC address. |
| InetSocketAddress namenode = |
| NameNode.getAddress(dfs.getUri().getAuthority()); |
| |
| return DFSUtil.getInfoServer(namenode, conf, true); |
| } |
| |
| private int doWork(final String[] args) throws IOException { |
| String proto = "http://"; |
| if (UserGroupInformation.isSecurityEnabled()) { |
| System.setProperty("https.cipherSuites", |
| Krb5AndCertsSslSocketConnector.KRB5_CIPHER_SUITES.get(0)); |
| proto = "https://"; |
| } |
| final StringBuilder url = new StringBuilder(proto); |
| |
| String namenodeAddress = getCurrentNamenodeAddress(); |
| if (namenodeAddress == null) { |
| //Error message already output in {@link #getCurrentNamenodeAddress()} |
| System.err.println("DFSck exiting."); |
| return 0; |
| } |
| url.append(namenodeAddress); |
| System.err.println("Connecting to namenode via " + url.toString()); |
| |
| url.append("/fsck?ugi=").append(ugi.getShortUserName()); |
| String dir = null; |
| boolean doListCorruptFileBlocks = false; |
| for (int idx = 0; idx < args.length; idx++) { |
| if (args[idx].equals("-move")) { url.append("&move=1"); } |
| else if (args[idx].equals("-delete")) { url.append("&delete=1"); } |
| else if (args[idx].equals("-files")) { url.append("&files=1"); } |
| else if (args[idx].equals("-openforwrite")) { url.append("&openforwrite=1"); } |
| else if (args[idx].equals("-blocks")) { url.append("&blocks=1"); } |
| else if (args[idx].equals("-locations")) { url.append("&locations=1"); } |
| else if (args[idx].equals("-racks")) { url.append("&racks=1"); } |
| else if (args[idx].equals("-list-corruptfileblocks")) { |
| url.append("&listcorruptfileblocks=1"); |
| doListCorruptFileBlocks = true; |
| } else if (!args[idx].startsWith("-")) { |
| if (null == dir) { |
| dir = args[idx]; |
| } else { |
| System.err.println("fsck: can only operate on one path at a time '" |
| + args[idx] + "'"); |
| printUsage(); |
| return -1; |
| } |
| } else { |
| System.err.println("fsck: Illegal option '" + args[idx] + "'"); |
| printUsage(); |
| return -1; |
| } |
| } |
| if (null == dir) { |
| dir = "/"; |
| } |
| url.append("&path=").append(URLEncoder.encode(dir, "UTF-8")); |
| if (doListCorruptFileBlocks) { |
| return listCorruptFileBlocks(dir, url.toString()); |
| } |
| URL path = new URL(url.toString()); |
| SecurityUtil.fetchServiceTicket(path); |
| URLConnection connection = path.openConnection(); |
| InputStream stream = connection.getInputStream(); |
| BufferedReader input = new BufferedReader(new InputStreamReader( |
| stream, "UTF-8")); |
| String line = null; |
| String lastLine = null; |
| int errCode = -1; |
| try { |
| while ((line = input.readLine()) != null) { |
| out.println(line); |
| lastLine = line; |
| } |
| } finally { |
| input.close(); |
| } |
| if (lastLine.endsWith(NamenodeFsck.HEALTHY_STATUS)) { |
| errCode = 0; |
| } else if (lastLine.endsWith(NamenodeFsck.CORRUPT_STATUS)) { |
| errCode = 1; |
| } else if (lastLine.endsWith(NamenodeFsck.NONEXISTENT_STATUS)) { |
| errCode = 0; |
| } |
| return errCode; |
| } |
| |
| public static void main(String[] args) throws Exception { |
| // -files option is also used by GenericOptionsParser |
| // Make sure that is not the first argument for fsck |
| int res = -1; |
| if ((args.length == 0 ) || ("-files".equals(args[0]))) |
| printUsage(); |
| else |
| res = ToolRunner.run(new DFSck(new HdfsConfiguration()), args); |
| System.exit(res); |
| } |
| } |