| /******************************************************************** |
| * 2014 - |
| * open source under Apache License Version 2.0 |
| ********************************************************************/ |
| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| #include "ClientDatanodeProtocol.pb.h" |
| #include "Datanode.h" |
| #include "Exception.h" |
| #include "ExceptionInternal.h" |
| #include "RpcHelper.h" |
| |
| #define DATANODE_VERSION 1 |
| #define DATANODE_PROTOCOL "org.apache.hadoop.hdfs.protocol.ClientDatanodeProtocol" |
| #define BLOCK_TOKEN_KIND "HDFS_BLOCK_TOKEN" |
| |
| using namespace google::protobuf; |
| |
| namespace Hdfs { |
| namespace Internal { |
| |
| DatanodeImpl::DatanodeImpl(const std::string & host, uint32_t port, |
| const SessionConfig & c, const RpcAuth & a) : |
| auth(a), client(RpcClient::getClient()), conf(c), protocol( |
| DATANODE_VERSION, DATANODE_PROTOCOL, BLOCK_TOKEN_KIND), server(host, port) { |
| server.setTokenService(""); |
| } |
| |
| void DatanodeImpl::invoke(const RpcCall & call, bool reuse) { |
| RpcChannel & channel = client.getChannel(auth, protocol, server, conf); |
| |
| try { |
| channel.invoke(call); |
| } catch (const HdfsFailoverException & e) { |
| //Datanode do not have HA configuration. |
| channel.close(true); |
| Hdfs::rethrow_if_nested(e); |
| assert(false && "HdfsFailoverException should be always a wrapper of other exception"); |
| } catch (...) { |
| channel.close(true); |
| throw; |
| } |
| |
| channel.close(!reuse); |
| } |
| |
| int64_t DatanodeImpl::getReplicaVisibleLength(const ExtendedBlock & b) { |
| try { |
| GetReplicaVisibleLengthRequestProto request; |
| GetReplicaVisibleLengthResponseProto response; |
| Build(b, request.mutable_block()); |
| invoke(RpcCall(true, "getReplicaVisibleLength", &request, &response), false); |
| return response.length(); |
| } catch (const HdfsRpcServerException & e) { |
| UnWrapper<ReplicaNotFoundException, HdfsIOException> unwraper(e); |
| unwraper.unwrap(__FILE__, __LINE__); |
| } |
| } |
| |
| void DatanodeImpl::getBlockLocalPathInfo(const ExtendedBlock & block, |
| const Token & token, BlockLocalPathInfo & info) { |
| try { |
| ExtendedBlock eb; |
| GetBlockLocalPathInfoRequestProto request; |
| GetBlockLocalPathInfoResponseProto response; |
| Build(block, request.mutable_block()); |
| Build(token, request.mutable_token()); |
| invoke(RpcCall(true, "getBlockLocalPathInfo", &request, &response), true); |
| Convert(eb, response.block()); |
| info.setBlock(eb); |
| info.setLocalBlockPath(response.localpath().c_str()); |
| info.setLocalMetaPath(response.localmetapath().c_str()); |
| } catch (const HdfsRpcServerException & e) { |
| UnWrapper<ReplicaNotFoundException, HdfsIOException> unwraper(e); |
| unwraper.unwrap(__FILE__, __LINE__); |
| } |
| } |
| |
| } |
| } |