| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| package org.apache.hdt.hadoop2.release; |
| |
| import java.io.FileNotFoundException; |
| import java.io.IOException; |
| import java.io.InputStream; |
| import java.io.OutputStream; |
| import java.net.URI; |
| import java.util.ArrayList; |
| import java.util.List; |
| |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.fs.FSDataInputStream; |
| import org.apache.hadoop.fs.FSDataOutputStream; |
| import org.apache.hadoop.fs.FileStatus; |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.fs.permission.FsAction; |
| import org.apache.hadoop.fs.permission.FsPermission; |
| import org.apache.hadoop.security.UserGroupInformation; |
| import org.apache.hdt.core.hdfs.ResourceInformation; |
| import org.apache.log4j.Logger; |
| |
| /** |
| * HDFS Client for HDFS version 1.1.2.21. |
| * |
| * @author Srimanth Gunturi |
| */ |
| public class HDFSClientRelease extends org.apache.hdt.core.hdfs.HDFSClient { |
| |
| private static Logger logger = Logger.getLogger(HDFSClientRelease.class); |
| private Configuration config; |
| |
| public HDFSClientRelease() { |
| config = new Configuration(); |
| } |
| |
| private ResourceInformation getResourceInformation(FileStatus fileStatus) { |
| ResourceInformation fi = new ResourceInformation(); |
| fi.setFolder(fileStatus.isDir()); |
| fi.setGroup(fileStatus.getGroup()); |
| fi.setLastAccessedTime(fileStatus.getAccessTime()); |
| fi.setLastModifiedTime(fileStatus.getAccessTime()); |
| fi.setName(fileStatus.getPath().getName()); |
| fi.setOwner(fileStatus.getOwner()); |
| fi.setPath(fileStatus.getPath().getParent() == null ? "/" : fileStatus.getPath().getParent().toString()); |
| fi.setReplicationFactor(fileStatus.getReplication()); |
| fi.setSize(fileStatus.getLen()); |
| FsPermission fsPermission = fileStatus.getPermission(); |
| updatePermissions(fi.getUserPermissions(), fsPermission.getUserAction()); |
| updatePermissions(fi.getGroupPermissions(), fsPermission.getGroupAction()); |
| updatePermissions(fi.getOtherPermissions(), fsPermission.getOtherAction()); |
| return fi; |
| } |
| |
| private void updatePermissions(ResourceInformation.Permissions permissions, FsAction action) { |
| permissions.read = action.implies(FsAction.READ); |
| permissions.write = action.implies(FsAction.WRITE); |
| permissions.execute = action.implies(FsAction.EXECUTE); |
| } |
| |
| protected FileSystem createFS(URI uri, String user) throws IOException, InterruptedException{ |
| if(user==null) |
| return FileSystem.get(uri, config); |
| return FileSystem.get(uri, config, user); |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see org.apache.hdt.core.hdfs.HDFSClient#getResource(java.net.URI) |
| */ |
| @Override |
| public ResourceInformation getResourceInformation(URI uri, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| FileStatus fileStatus = null; |
| ResourceInformation fi = null; |
| try { |
| fileStatus = fs.getFileStatus(path); |
| fi = getResourceInformation(fileStatus); |
| } catch (FileNotFoundException fne) { |
| logger.info(fne.getMessage()); |
| logger.debug(fne.getMessage(), fne); |
| } |
| return fi; |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see org.apache.hdt.core.hdfs.HDFSClient#setResource(java.net.URI, |
| * org.apache.hdt.core.hdfs.ResourceInformation) |
| */ |
| @Override |
| public void setResourceInformation(URI uri, ResourceInformation information, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| if (!information.isFolder()) { |
| fs.setTimes(path, information.getLastModifiedTime(), information.getLastAccessedTime()); |
| } |
| if (information.getOwner() != null || information.getGroup() != null) |
| fs.setOwner(path, information.getOwner(), information.getGroup()); |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see |
| * org.apache.hdt.core.hdfs.HDFSClient#listResources(java.net.URI) |
| */ |
| @Override |
| public List<ResourceInformation> listResources(URI uri, String user) throws IOException, InterruptedException { |
| List<ResourceInformation> ris = null; |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| FileStatus[] listStatus = fs.listStatus(path); |
| if (listStatus != null) { |
| ris = new ArrayList<ResourceInformation>(); |
| for (FileStatus ls : listStatus) { |
| ris.add(getResourceInformation(ls)); |
| } |
| } |
| return ris; |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see |
| * org.apache.hdt.core.hdfs.HDFSClient#openInputStream(java.net.URI, |
| * org.eclipse.core.runtime.IProgressMonitor) |
| */ |
| @Override |
| public InputStream openInputStream(URI uri, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| FSDataInputStream open = fs.open(path); |
| return open; |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see |
| * org.apache.hdt.core.hdfs.HDFSClient#openInputStream(java.net.URI, |
| * org.eclipse.core.runtime.IProgressMonitor) |
| */ |
| @Override |
| public OutputStream createOutputStream(URI uri, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| FSDataOutputStream outputStream = fs.create(path); |
| return outputStream; |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see |
| * org.apache.hdt.core.hdfs.HDFSClient#openInputStream(java.net.URI, |
| * org.eclipse.core.runtime.IProgressMonitor) |
| */ |
| @Override |
| public OutputStream openOutputStream(URI uri, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| // TODO. Temporary fix till Issue#3 is fixed. |
| FSDataOutputStream outputStream = fs.create(path); |
| return outputStream; |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see org.apache.hdt.core.hdfs.HDFSClient#mkdirs(java.net.URI, |
| * org.eclipse.core.runtime.IProgressMonitor) |
| */ |
| @Override |
| public boolean mkdirs(URI uri, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| return fs.mkdirs(path); |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see org.apache.hdt.core.hdfs.HDFSClient#delete(java.net.URI, |
| * org.eclipse.core.runtime.IProgressMonitor) |
| */ |
| @Override |
| public void delete(URI uri, String user) throws IOException, InterruptedException { |
| FileSystem fs = createFS(uri, user); |
| Path path = new Path(uri.getPath()); |
| fs.delete(path, true); |
| } |
| |
| /* |
| * (non-Javadoc) |
| * |
| * @see |
| * org.apache.hdt.core.hdfs.HDFSClient#getDefaultUserAndGroupIds() |
| */ |
| @Override |
| public List<String> getDefaultUserAndGroupIds() throws IOException { |
| List<String> idList = new ArrayList<String>(); |
| UserGroupInformation currentUser = UserGroupInformation.getCurrentUser(); |
| idList.add(currentUser.getShortUserName()); |
| String[] groupIds = currentUser.getGroupNames(); |
| if (groupIds != null) { |
| for (String groupId : groupIds) { |
| idList.add(groupId); |
| } |
| } |
| return idList; |
| } |
| |
| } |