| /** |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with this |
| * work for additional information regarding copyright ownership. The ASF |
| * licenses this file to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * <p> |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * <p> |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT |
| * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the |
| * License for the specific language governing permissions and limitations under |
| * the License. |
| */ |
| package org.apache.hadoop.hdds.scm.protocolPB; |
| |
| import com.google.common.base.Preconditions; |
| import com.google.protobuf.RpcController; |
| import com.google.protobuf.ServiceException; |
| import org.apache.commons.lang3.tuple.Pair; |
| import org.apache.hadoop.hdds.annotation.InterfaceAudience; |
| import org.apache.hadoop.hdds.protocol.proto.HddsProtos; |
| import org.apache.hadoop.hdds.protocol.proto.HddsProtos.GetScmInfoResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ActivatePipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ClosePipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ContainerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ContainerResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.DatanodeUsageInfoRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.DatanodeUsageInfoResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.DeactivatePipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.DecommissionNodesRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ForceExitSafeModeRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ForceExitSafeModeResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetContainerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetContainerWithPipelineBatchRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetContainerWithPipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetPipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetPipelineResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetSafeModeRuleStatusesRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.GetSafeModeRuleStatusesResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.InSafeModeRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ListPipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ListPipelineResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.NodeQueryRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.NodeQueryResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.PipelineRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.PipelineResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.RecommissionNodesRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ReplicationManagerStatusRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ReplicationManagerStatusResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.SCMCloseContainerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.SCMDeleteContainerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.SCMListContainerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.SCMListContainerResponseProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.SafeModeRuleStatusProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ScmContainerLocationRequest; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ScmContainerLocationRequest.Builder; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.ScmContainerLocationResponse; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.StartMaintenanceNodesRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.StartReplicationManagerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.StopReplicationManagerRequestProto; |
| import org.apache.hadoop.hdds.protocol.proto.StorageContainerLocationProtocolProtos.Type; |
| import org.apache.hadoop.hdds.scm.ScmInfo; |
| import org.apache.hadoop.hdds.scm.container.ContainerInfo; |
| import org.apache.hadoop.hdds.scm.container.common.helpers.ContainerWithPipeline; |
| import org.apache.hadoop.hdds.scm.pipeline.Pipeline; |
| import org.apache.hadoop.hdds.scm.protocol.StorageContainerLocationProtocol; |
| import org.apache.hadoop.hdds.tracing.TracingUtil; |
| import org.apache.hadoop.ipc.ProtobufHelper; |
| import org.apache.hadoop.ipc.ProtocolTranslator; |
| import org.apache.hadoop.ipc.RPC; |
| |
| import java.io.Closeable; |
| import java.io.IOException; |
| import java.util.ArrayList; |
| import java.util.HashMap; |
| import java.util.List; |
| import java.util.Map; |
| import java.util.function.Consumer; |
| |
| import static org.apache.hadoop.ozone.ClientVersions.CURRENT_VERSION; |
| |
| /** |
| * This class is the client-side translator to translate the requests made on |
| * the {@link StorageContainerLocationProtocol} interface to the RPC server |
| * implementing {@link StorageContainerLocationProtocolPB}. |
| */ |
| @InterfaceAudience.Private |
| public final class StorageContainerLocationProtocolClientSideTranslatorPB |
| implements StorageContainerLocationProtocol, ProtocolTranslator, Closeable { |
| |
| /** |
| * RpcController is not used and hence is set to null. |
| */ |
| private static final RpcController NULL_RPC_CONTROLLER = null; |
| |
| private final StorageContainerLocationProtocolPB rpcProxy; |
| |
| /** |
| * Creates a new StorageContainerLocationProtocolClientSideTranslatorPB. |
| * |
| * @param rpcProxy {@link StorageContainerLocationProtocolPB} RPC proxy |
| */ |
| public StorageContainerLocationProtocolClientSideTranslatorPB( |
| StorageContainerLocationProtocolPB rpcProxy) { |
| this.rpcProxy = rpcProxy; |
| } |
| |
| /** |
| * Helper method to wrap the request and send the message. |
| */ |
| private ScmContainerLocationResponse submitRequest( |
| StorageContainerLocationProtocolProtos.Type type, |
| Consumer<Builder> builderConsumer) throws IOException { |
| final ScmContainerLocationResponse response; |
| try { |
| |
| Builder builder = ScmContainerLocationRequest.newBuilder() |
| .setCmdType(type) |
| .setVersion(CURRENT_VERSION) |
| .setTraceID(TracingUtil.exportCurrentSpan()); |
| builderConsumer.accept(builder); |
| ScmContainerLocationRequest wrapper = builder.build(); |
| |
| response = submitRpcRequest(wrapper); |
| } catch (ServiceException ex) { |
| throw ProtobufHelper.getRemoteException(ex); |
| } |
| return response; |
| } |
| |
| private ScmContainerLocationResponse submitRpcRequest( |
| ScmContainerLocationRequest wrapper) throws ServiceException { |
| return rpcProxy.submitRequest(NULL_RPC_CONTROLLER, wrapper); |
| } |
| |
| /** |
| * Asks SCM where a container should be allocated. SCM responds with the set |
| * of datanodes that should be used creating this container. Ozone/SCM only |
| * supports replication factor of either 1 or 3. |
| * |
| * @param type - Replication Type |
| * @param factor - Replication Count |
| */ |
| @Override |
| public ContainerWithPipeline allocateContainer( |
| HddsProtos.ReplicationType type, HddsProtos.ReplicationFactor factor, |
| String owner) throws IOException { |
| |
| ContainerRequestProto request = ContainerRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setReplicationFactor(factor) |
| .setReplicationType(type) |
| .setOwner(owner) |
| .build(); |
| |
| ContainerResponseProto response = |
| submitRequest(Type.AllocateContainer, |
| builder -> builder.setContainerRequest(request)) |
| .getContainerResponse(); |
| //TODO should be migrated to use the top level status structure. |
| if (response.getErrorCode() != ContainerResponseProto.Error.success) { |
| throw new IOException(response.hasErrorMessage() ? |
| response.getErrorMessage() : "Allocate container failed."); |
| } |
| return ContainerWithPipeline.fromProtobuf( |
| response.getContainerWithPipeline()); |
| } |
| |
| @Override |
| public ContainerInfo getContainer(long containerID) throws IOException { |
| Preconditions.checkState(containerID >= 0, |
| "Container ID cannot be negative"); |
| GetContainerRequestProto request = GetContainerRequestProto |
| .newBuilder() |
| .setContainerID(containerID) |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .build(); |
| ScmContainerLocationResponse response = |
| submitRequest(Type.GetContainer, |
| (builder) -> builder.setGetContainerRequest(request)); |
| return ContainerInfo |
| .fromProtobuf(response.getGetContainerResponse().getContainerInfo()); |
| |
| } |
| |
| /** |
| * {@inheritDoc} |
| */ |
| @Override |
| public ContainerWithPipeline getContainerWithPipeline(long containerID) |
| throws IOException { |
| Preconditions.checkState(containerID >= 0, |
| "Container ID cannot be negative"); |
| GetContainerWithPipelineRequestProto request = |
| GetContainerWithPipelineRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setContainerID(containerID).build(); |
| |
| ScmContainerLocationResponse response = |
| submitRequest(Type.GetContainerWithPipeline, |
| (builder) -> builder.setGetContainerWithPipelineRequest(request)); |
| |
| return ContainerWithPipeline.fromProtobuf( |
| response.getGetContainerWithPipelineResponse() |
| .getContainerWithPipeline()); |
| |
| } |
| |
| /** |
| * {@inheritDoc} |
| */ |
| @Override |
| public List<ContainerWithPipeline> getContainerWithPipelineBatch( |
| List<Long> containerIDs) throws IOException { |
| for (Long containerID: containerIDs) { |
| Preconditions.checkState(containerID >= 0, |
| "Container ID cannot be negative"); |
| } |
| |
| GetContainerWithPipelineBatchRequestProto request = |
| GetContainerWithPipelineBatchRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .addAllContainerIDs(containerIDs) |
| .build(); |
| |
| ScmContainerLocationResponse response = |
| submitRequest(Type.GetContainerWithPipelineBatch, |
| (builder) -> builder |
| .setGetContainerWithPipelineBatchRequest(request)); |
| |
| List<HddsProtos.ContainerWithPipeline> protoCps = response |
| .getGetContainerWithPipelineBatchResponse() |
| .getContainerWithPipelinesList(); |
| |
| List<ContainerWithPipeline> cps = new ArrayList<>(); |
| |
| for (HddsProtos.ContainerWithPipeline cp : protoCps) { |
| cps.add(ContainerWithPipeline.fromProtobuf(cp)); |
| } |
| |
| return cps; |
| } |
| |
| /** |
| * {@inheritDoc} |
| */ |
| @Override |
| public List<ContainerInfo> listContainer(long startContainerID, int count) |
| throws IOException { |
| return listContainer(startContainerID, count, null); |
| } |
| |
| @Override |
| public List<ContainerInfo> listContainer(long startContainerID, int count, |
| HddsProtos.LifeCycleState state) throws IOException { |
| Preconditions.checkState(startContainerID >= 0, |
| "Container ID cannot be negative."); |
| Preconditions.checkState(count > 0, |
| "Container count must be greater than 0."); |
| SCMListContainerRequestProto.Builder builder = SCMListContainerRequestProto |
| .newBuilder(); |
| builder.setStartContainerID(startContainerID); |
| builder.setCount(count); |
| builder.setTraceID(TracingUtil.exportCurrentSpan()); |
| if (state != null) { |
| builder.setState(state); |
| } |
| |
| SCMListContainerRequestProto request = builder.build(); |
| |
| SCMListContainerResponseProto response = |
| submitRequest(Type.ListContainer, |
| builder1 -> builder1.setScmListContainerRequest(request)) |
| .getScmListContainerResponse(); |
| List<ContainerInfo> containerList = new ArrayList<>(); |
| for (HddsProtos.ContainerInfoProto containerInfoProto : response |
| .getContainersList()) { |
| containerList.add(ContainerInfo.fromProtobuf(containerInfoProto)); |
| } |
| return containerList; |
| } |
| |
| /** |
| * Ask SCM to delete a container by name. SCM will remove |
| * the container mapping in its database. |
| */ |
| @Override |
| public void deleteContainer(long containerID) |
| throws IOException { |
| Preconditions.checkState(containerID >= 0, |
| "Container ID cannot be negative"); |
| SCMDeleteContainerRequestProto request = SCMDeleteContainerRequestProto |
| .newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setContainerID(containerID) |
| .build(); |
| submitRequest(Type.DeleteContainer, |
| builder -> builder.setScmDeleteContainerRequest(request)); |
| |
| } |
| |
| /** |
| * Queries a list of Nodes based on their operational state or health state. |
| * Passing a null for either value acts as a wildcard for that state. |
| * |
| * @param opState The operation state of the node |
| * @param nodeState The health of the node |
| * @param clientVersion |
| * @return List of Datanodes. |
| */ |
| @Override |
| public List<HddsProtos.Node> queryNode( |
| HddsProtos.NodeOperationalState opState, HddsProtos.NodeState |
| nodeState, HddsProtos.QueryScope queryScope, String poolName, |
| int clientVersion) throws IOException { |
| // TODO : We support only cluster wide query right now. So ignoring checking |
| // queryScope and poolName |
| NodeQueryRequestProto.Builder builder = NodeQueryRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setScope(queryScope).setPoolName(poolName); |
| if (opState != null) { |
| builder.setOpState(opState); |
| } |
| if (nodeState != null) { |
| builder.setState(nodeState); |
| } |
| NodeQueryRequestProto request = builder.build(); |
| NodeQueryResponseProto response = submitRequest(Type.QueryNode, |
| builder1 -> builder1.setNodeQueryRequest(request)) |
| .getNodeQueryResponse(); |
| return response.getDatanodesList(); |
| } |
| |
| /** |
| * Attempts to decommission the list of nodes. |
| * @param nodes The list of hostnames or hostname:ports to decommission |
| * @throws IOException |
| */ |
| @Override |
| public void decommissionNodes(List<String> nodes) throws IOException { |
| Preconditions.checkNotNull(nodes); |
| DecommissionNodesRequestProto request = |
| DecommissionNodesRequestProto.newBuilder() |
| .addAllHosts(nodes) |
| .build(); |
| submitRequest(Type.DecommissionNodes, |
| builder -> builder.setDecommissionNodesRequest(request)); |
| } |
| |
| /** |
| * Attempts to recommission the list of nodes. |
| * @param nodes The list of hostnames or hostname:ports to recommission |
| * @throws IOException |
| */ |
| @Override |
| public void recommissionNodes(List<String> nodes) throws IOException { |
| Preconditions.checkNotNull(nodes); |
| RecommissionNodesRequestProto request = |
| RecommissionNodesRequestProto.newBuilder() |
| .addAllHosts(nodes) |
| .build(); |
| submitRequest(Type.RecommissionNodes, |
| builder -> builder.setRecommissionNodesRequest(request)); |
| } |
| |
| /** |
| * Attempts to put the list of nodes into maintenance mode. |
| * |
| * @param nodes The list of hostnames or hostname:ports to put into |
| * maintenance |
| * @param endInHours A number of hours from now where the nodes will be taken |
| * out of maintenance automatically. Passing zero will |
| * allow the nodes to stay in maintenance indefinitely |
| * @throws IOException |
| */ |
| @Override |
| public void startMaintenanceNodes(List<String> nodes, int endInHours) |
| throws IOException { |
| Preconditions.checkNotNull(nodes); |
| StartMaintenanceNodesRequestProto request = |
| StartMaintenanceNodesRequestProto.newBuilder() |
| .addAllHosts(nodes) |
| .setEndInHours(endInHours) |
| .build(); |
| submitRequest(Type.StartMaintenanceNodes, |
| builder -> builder.setStartMaintenanceNodesRequest(request)); |
| } |
| |
| /** |
| * Close a container. |
| * |
| * @param containerID ID of the container to close |
| * @throws IOException in case of any Exception |
| */ |
| @Override |
| public void closeContainer(long containerID) throws IOException { |
| Preconditions.checkState(containerID >= 0, |
| "Container ID cannot be negative"); |
| SCMCloseContainerRequestProto request = SCMCloseContainerRequestProto |
| .newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setContainerID(containerID) |
| .build(); |
| submitRequest(Type.CloseContainer, |
| builder -> builder.setScmCloseContainerRequest(request)); |
| } |
| |
| /** |
| * Creates a replication pipeline of a specified type. |
| * |
| * @param replicationType - replication type |
| * @param factor - factor 1 or 3 |
| * @param nodePool - optional machine list to build a pipeline. |
| */ |
| @Override |
| public Pipeline createReplicationPipeline(HddsProtos.ReplicationType |
| replicationType, HddsProtos.ReplicationFactor factor, HddsProtos |
| .NodePool nodePool) throws IOException { |
| PipelineRequestProto request = PipelineRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setNodePool(nodePool) |
| .setReplicationFactor(factor) |
| .setReplicationType(replicationType) |
| .build(); |
| |
| PipelineResponseProto response = |
| submitRequest(Type.AllocatePipeline, |
| builder -> builder.setPipelineRequest(request)) |
| .getPipelineResponse(); |
| if (response.getErrorCode() == |
| PipelineResponseProto.Error.success) { |
| Preconditions.checkState(response.hasPipeline(), "With success, " + |
| "must come a pipeline"); |
| return Pipeline.getFromProtobuf(response.getPipeline()); |
| } else { |
| String errorMessage = String.format("create replication pipeline " + |
| "failed. code : %s Message: %s", response.getErrorCode(), |
| response.hasErrorMessage() ? response.getErrorMessage() : ""); |
| throw new IOException(errorMessage); |
| } |
| |
| } |
| |
| @Override |
| public List<Pipeline> listPipelines() throws IOException { |
| ListPipelineRequestProto request = ListPipelineRequestProto |
| .newBuilder().setTraceID(TracingUtil.exportCurrentSpan()) |
| .build(); |
| |
| ListPipelineResponseProto response = submitRequest(Type.ListPipelines, |
| builder -> builder.setListPipelineRequest(request)) |
| .getListPipelineResponse(); |
| |
| List<Pipeline> list = new ArrayList<>(); |
| for (HddsProtos.Pipeline pipeline : response.getPipelinesList()) { |
| Pipeline fromProtobuf = Pipeline.getFromProtobuf(pipeline); |
| list.add(fromProtobuf); |
| } |
| return list; |
| |
| } |
| |
| @Override |
| public Pipeline getPipeline(HddsProtos.PipelineID pipelineID) |
| throws IOException { |
| GetPipelineRequestProto request = GetPipelineRequestProto.newBuilder() |
| .setPipelineID(pipelineID) |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .build(); |
| GetPipelineResponseProto response = submitRequest(Type.GetPipeline, |
| builder -> builder.setGetPipelineRequest(request)) |
| .getGetPipelineResponse(); |
| |
| return Pipeline.getFromProtobuf(response.getPipeline()); |
| } |
| |
| @Override |
| public void activatePipeline(HddsProtos.PipelineID pipelineID) |
| throws IOException { |
| ActivatePipelineRequestProto request = |
| ActivatePipelineRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setPipelineID(pipelineID) |
| .build(); |
| submitRequest(Type.ActivatePipeline, |
| builder -> builder.setActivatePipelineRequest(request)); |
| |
| } |
| |
| @Override |
| public void deactivatePipeline(HddsProtos.PipelineID pipelineID) |
| throws IOException { |
| |
| DeactivatePipelineRequestProto request = |
| DeactivatePipelineRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setPipelineID(pipelineID) |
| .build(); |
| submitRequest(Type.DeactivatePipeline, |
| builder -> builder.setDeactivatePipelineRequest(request)); |
| } |
| |
| @Override |
| public void closePipeline(HddsProtos.PipelineID pipelineID) |
| throws IOException { |
| |
| ClosePipelineRequestProto request = |
| ClosePipelineRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .setPipelineID(pipelineID) |
| .build(); |
| submitRequest(Type.ClosePipeline, |
| builder -> builder.setClosePipelineRequest(request)); |
| |
| } |
| |
| @Override |
| public ScmInfo getScmInfo() throws IOException { |
| HddsProtos.GetScmInfoRequestProto request = |
| HddsProtos.GetScmInfoRequestProto.newBuilder() |
| .setTraceID(TracingUtil.exportCurrentSpan()) |
| .build(); |
| |
| GetScmInfoResponseProto resp = submitRequest(Type.GetScmInfo, |
| builder -> builder.setGetScmInfoRequest(request)) |
| .getGetScmInfoResponse(); |
| ScmInfo.Builder builder = new ScmInfo.Builder() |
| .setClusterId(resp.getClusterId()) |
| .setScmId(resp.getScmId()); |
| return builder.build(); |
| |
| } |
| |
| /** |
| * Check if SCM is in safe mode. |
| * |
| * @return Returns true if SCM is in safe mode else returns false. |
| */ |
| @Override |
| public boolean inSafeMode() throws IOException { |
| InSafeModeRequestProto request = |
| InSafeModeRequestProto.getDefaultInstance(); |
| |
| return submitRequest(Type.InSafeMode, |
| builder -> builder.setInSafeModeRequest(request)) |
| .getInSafeModeResponse().getInSafeMode(); |
| |
| } |
| |
| @Override |
| public Map<String, Pair<Boolean, String>> getSafeModeRuleStatuses() |
| throws IOException { |
| GetSafeModeRuleStatusesRequestProto request = |
| GetSafeModeRuleStatusesRequestProto.getDefaultInstance(); |
| GetSafeModeRuleStatusesResponseProto response = |
| submitRequest(Type.GetSafeModeRuleStatuses, |
| builder -> builder.setGetSafeModeRuleStatusesRequest(request)) |
| .getGetSafeModeRuleStatusesResponse(); |
| Map<String, Pair<Boolean, String>> map = new HashMap(); |
| for (SafeModeRuleStatusProto statusProto : |
| response.getSafeModeRuleStatusesProtoList()) { |
| map.put(statusProto.getRuleName(), |
| Pair.of(statusProto.getValidate(), statusProto.getStatusText())); |
| } |
| return map; |
| } |
| |
| /** |
| * Force SCM out of Safe mode. |
| * |
| * @return returns true if operation is successful. |
| */ |
| @Override |
| public boolean forceExitSafeMode() throws IOException { |
| ForceExitSafeModeRequestProto request = |
| ForceExitSafeModeRequestProto.getDefaultInstance(); |
| ForceExitSafeModeResponseProto resp = |
| submitRequest(Type.ForceExitSafeMode, |
| builder -> builder.setForceExitSafeModeRequest(request)) |
| .getForceExitSafeModeResponse(); |
| |
| return resp.getExitedSafeMode(); |
| |
| } |
| |
| @Override |
| public void startReplicationManager() throws IOException { |
| |
| StartReplicationManagerRequestProto request = |
| StartReplicationManagerRequestProto.getDefaultInstance(); |
| submitRequest(Type.StartReplicationManager, |
| builder -> builder.setStartReplicationManagerRequest(request)); |
| |
| } |
| |
| @Override |
| public void stopReplicationManager() throws IOException { |
| |
| StopReplicationManagerRequestProto request = |
| StopReplicationManagerRequestProto.getDefaultInstance(); |
| submitRequest(Type.StopReplicationManager, |
| builder -> builder.setStopReplicationManagerRequest(request)); |
| |
| } |
| |
| @Override |
| public boolean getReplicationManagerStatus() throws IOException { |
| |
| ReplicationManagerStatusRequestProto request = |
| ReplicationManagerStatusRequestProto.getDefaultInstance(); |
| ReplicationManagerStatusResponseProto response = |
| submitRequest(Type.GetReplicationManagerStatus, |
| builder -> builder.setSeplicationManagerStatusRequest(request)) |
| .getReplicationManagerStatusResponse(); |
| return response.getIsRunning(); |
| |
| } |
| |
| /** |
| * Builds request for datanode usage information and receives response. |
| * |
| * @param ipaddress Address String |
| * @param uuid UUID String |
| * @return List of DatanodeUsageInfoProto. Each element contains info such as |
| * capacity, SCMUsed, and remaining space. |
| * @throws IOException |
| */ |
| @Override |
| public List<HddsProtos.DatanodeUsageInfoProto> getDatanodeUsageInfo( |
| String ipaddress, String uuid) throws IOException { |
| |
| DatanodeUsageInfoRequestProto request = |
| DatanodeUsageInfoRequestProto.newBuilder() |
| .setIpaddress(ipaddress) |
| .setUuid(uuid) |
| .build(); |
| |
| DatanodeUsageInfoResponseProto response = |
| submitRequest(Type.DatanodeUsageInfo, |
| builder -> builder.setDatanodeUsageInfoRequest(request)) |
| .getDatanodeUsageInfoResponse(); |
| return response.getInfoList(); |
| } |
| |
| /** |
| * Get usage information of most or least used datanodes. |
| * |
| * @param mostUsed true if most used, false if least used |
| * @param count Integer number of nodes to get info for |
| * @return List of DatanodeUsageInfoProto. Each element contains info such as |
| * capacity, SCMUsed, and remaining space. |
| * @throws IOException |
| */ |
| @Override |
| public List<HddsProtos.DatanodeUsageInfoProto> getDatanodeUsageInfo( |
| boolean mostUsed, int count) throws IOException { |
| DatanodeUsageInfoRequestProto request = |
| DatanodeUsageInfoRequestProto.newBuilder() |
| .setMostUsed(mostUsed) |
| .setCount(count) |
| .build(); |
| |
| DatanodeUsageInfoResponseProto response = |
| submitRequest(Type.DatanodeUsageInfo, |
| builder -> builder.setDatanodeUsageInfoRequest(request)) |
| .getDatanodeUsageInfoResponse(); |
| |
| return response.getInfoList(); |
| } |
| |
| @Override |
| public Object getUnderlyingProxyObject() { |
| return rpcProxy; |
| } |
| |
| @Override |
| public void close() { |
| RPC.stopProxy(rpcProxy); |
| } |
| } |