| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| package org.apache.iotdb.confignode.procedure.impl.schema; |
| |
| import org.apache.iotdb.common.rpc.thrift.TConsensusGroupId; |
| import org.apache.iotdb.common.rpc.thrift.TDataNodeLocation; |
| import org.apache.iotdb.common.rpc.thrift.TRegionReplicaSet; |
| import org.apache.iotdb.common.rpc.thrift.TSStatus; |
| import org.apache.iotdb.commons.exception.MetadataException; |
| import org.apache.iotdb.commons.path.PartialPath; |
| import org.apache.iotdb.commons.path.PathPatternTree; |
| import org.apache.iotdb.confignode.client.DataNodeRequestType; |
| import org.apache.iotdb.confignode.client.async.AsyncDataNodeClientPool; |
| import org.apache.iotdb.confignode.client.async.handlers.AsyncClientHandler; |
| import org.apache.iotdb.confignode.consensus.request.write.pipe.payload.PipeDeleteTimeSeriesPlan; |
| import org.apache.iotdb.confignode.consensus.request.write.pipe.payload.PipeEnrichedPlan; |
| import org.apache.iotdb.confignode.procedure.env.ConfigNodeProcedureEnv; |
| import org.apache.iotdb.confignode.procedure.exception.ProcedureException; |
| import org.apache.iotdb.confignode.procedure.exception.ProcedureSuspendedException; |
| import org.apache.iotdb.confignode.procedure.exception.ProcedureYieldException; |
| import org.apache.iotdb.confignode.procedure.impl.StateMachineProcedure; |
| import org.apache.iotdb.confignode.procedure.state.schema.DeleteTimeSeriesState; |
| import org.apache.iotdb.confignode.procedure.store.ProcedureType; |
| import org.apache.iotdb.consensus.exception.ConsensusException; |
| import org.apache.iotdb.db.exception.metadata.PathNotExistException; |
| import org.apache.iotdb.mpp.rpc.thrift.TConstructSchemaBlackListReq; |
| import org.apache.iotdb.mpp.rpc.thrift.TDeleteDataForDeleteSchemaReq; |
| import org.apache.iotdb.mpp.rpc.thrift.TDeleteTimeSeriesReq; |
| import org.apache.iotdb.mpp.rpc.thrift.TInvalidateMatchedSchemaCacheReq; |
| import org.apache.iotdb.mpp.rpc.thrift.TRollbackSchemaBlackListReq; |
| import org.apache.iotdb.pipe.api.exception.PipeException; |
| import org.apache.iotdb.rpc.TSStatusCode; |
| import org.apache.iotdb.tsfile.utils.ReadWriteIOUtils; |
| |
| import org.slf4j.Logger; |
| import org.slf4j.LoggerFactory; |
| |
| import java.io.ByteArrayOutputStream; |
| import java.io.DataOutputStream; |
| import java.io.IOException; |
| import java.nio.ByteBuffer; |
| import java.util.ArrayList; |
| import java.util.List; |
| import java.util.Map; |
| import java.util.Objects; |
| import java.util.Set; |
| import java.util.function.BiFunction; |
| import java.util.stream.Collectors; |
| |
| public class DeleteTimeSeriesProcedure |
| extends StateMachineProcedure<ConfigNodeProcedureEnv, DeleteTimeSeriesState> { |
| |
| private static final Logger LOGGER = LoggerFactory.getLogger(DeleteTimeSeriesProcedure.class); |
| |
| private String queryId; |
| |
| private PathPatternTree patternTree; |
| private transient ByteBuffer patternTreeBytes; |
| |
| private transient String requestMessage; |
| |
| private static final String CONSENSUS_WRITE_ERROR = |
| "Failed in the write API executing the consensus layer due to: "; |
| |
| public DeleteTimeSeriesProcedure(boolean isGeneratedByPipe) { |
| super(isGeneratedByPipe); |
| } |
| |
| public DeleteTimeSeriesProcedure( |
| String queryId, PathPatternTree patternTree, boolean isGeneratedByPipe) { |
| super(isGeneratedByPipe); |
| this.queryId = queryId; |
| setPatternTree(patternTree); |
| } |
| |
| @Override |
| protected Flow executeFromState(ConfigNodeProcedureEnv env, DeleteTimeSeriesState state) |
| throws ProcedureSuspendedException, ProcedureYieldException, InterruptedException { |
| long startTime = System.currentTimeMillis(); |
| try { |
| switch (state) { |
| case CONSTRUCT_BLACK_LIST: |
| LOGGER.info("Construct schemaEngine black list of timeSeries {}", requestMessage); |
| if (constructBlackList(env) > 0) { |
| setNextState(DeleteTimeSeriesState.CLEAN_DATANODE_SCHEMA_CACHE); |
| break; |
| } else { |
| setFailure( |
| new ProcedureException( |
| new PathNotExistException( |
| patternTree.getAllPathPatterns().stream() |
| .map(PartialPath::getFullPath) |
| .collect(Collectors.toList()), |
| false))); |
| return Flow.NO_MORE_STATE; |
| } |
| case CLEAN_DATANODE_SCHEMA_CACHE: |
| LOGGER.info("Invalidate cache of timeSeries {}", requestMessage); |
| invalidateCache(env); |
| break; |
| case DELETE_DATA: |
| LOGGER.info("Delete data of timeSeries {}", requestMessage); |
| deleteData(env); |
| break; |
| case DELETE_TIMESERIES_SCHEMA: |
| LOGGER.info("Delete timeSeries schemaEngine of {}", requestMessage); |
| deleteTimeSeriesSchema(env); |
| collectPayload4Pipe(env); |
| return Flow.NO_MORE_STATE; |
| default: |
| setFailure(new ProcedureException("Unrecognized state " + state)); |
| return Flow.NO_MORE_STATE; |
| } |
| return Flow.HAS_MORE_STATE; |
| } finally { |
| LOGGER.info( |
| "DeleteTimeSeries-[{}] costs {}ms", state, (System.currentTimeMillis() - startTime)); |
| } |
| } |
| |
| // return the total num of timeSeries in schemaEngine black list |
| private long constructBlackList(ConfigNodeProcedureEnv env) { |
| Map<TConsensusGroupId, TRegionReplicaSet> targetSchemaRegionGroup = |
| env.getConfigManager().getRelatedSchemaRegionGroup(patternTree); |
| if (targetSchemaRegionGroup.isEmpty()) { |
| return 0; |
| } |
| List<TSStatus> successResult = new ArrayList<>(); |
| DeleteTimeSeriesRegionTaskExecutor<TConstructSchemaBlackListReq> constructBlackListTask = |
| new DeleteTimeSeriesRegionTaskExecutor<TConstructSchemaBlackListReq>( |
| "construct schemaengine black list", |
| env, |
| targetSchemaRegionGroup, |
| DataNodeRequestType.CONSTRUCT_SCHEMA_BLACK_LIST, |
| ((dataNodeLocation, consensusGroupIdList) -> |
| new TConstructSchemaBlackListReq(consensusGroupIdList, patternTreeBytes))) { |
| @Override |
| protected List<TConsensusGroupId> processResponseOfOneDataNode( |
| TDataNodeLocation dataNodeLocation, |
| List<TConsensusGroupId> consensusGroupIdList, |
| TSStatus response) { |
| List<TConsensusGroupId> failedRegionList = new ArrayList<>(); |
| if (response.getCode() == TSStatusCode.SUCCESS_STATUS.getStatusCode()) { |
| successResult.add(response); |
| } else if (response.getCode() == TSStatusCode.MULTIPLE_ERROR.getStatusCode()) { |
| List<TSStatus> subStatusList = response.getSubStatus(); |
| for (int i = 0; i < subStatusList.size(); i++) { |
| if (subStatusList.get(i).getCode() == TSStatusCode.SUCCESS_STATUS.getStatusCode()) { |
| successResult.add(subStatusList.get(i)); |
| } else { |
| failedRegionList.add(consensusGroupIdList.get(i)); |
| } |
| } |
| } else { |
| failedRegionList.addAll(consensusGroupIdList); |
| } |
| return failedRegionList; |
| } |
| }; |
| constructBlackListTask.execute(); |
| |
| if (isFailed()) { |
| return 0; |
| } |
| |
| long preDeletedNum = 0; |
| for (TSStatus resp : successResult) { |
| preDeletedNum += Long.parseLong(resp.getMessage()); |
| } |
| return preDeletedNum; |
| } |
| |
| private void invalidateCache(ConfigNodeProcedureEnv env) { |
| Map<Integer, TDataNodeLocation> dataNodeLocationMap = |
| env.getConfigManager().getNodeManager().getRegisteredDataNodeLocations(); |
| AsyncClientHandler<TInvalidateMatchedSchemaCacheReq, TSStatus> clientHandler = |
| new AsyncClientHandler<>( |
| DataNodeRequestType.INVALIDATE_MATCHED_SCHEMA_CACHE, |
| new TInvalidateMatchedSchemaCacheReq(patternTreeBytes), |
| dataNodeLocationMap); |
| AsyncDataNodeClientPool.getInstance().sendAsyncRequestToDataNodeWithRetry(clientHandler); |
| Map<Integer, TSStatus> statusMap = clientHandler.getResponseMap(); |
| for (TSStatus status : statusMap.values()) { |
| // all dataNodes must clear the related schemaEngine cache |
| if (status.getCode() != TSStatusCode.SUCCESS_STATUS.getStatusCode()) { |
| LOGGER.error("Failed to invalidate schemaEngine cache of timeSeries {}", requestMessage); |
| setFailure( |
| new ProcedureException(new MetadataException("Invalidate schemaEngine cache failed"))); |
| return; |
| } |
| } |
| |
| setNextState(DeleteTimeSeriesState.DELETE_DATA); |
| } |
| |
| private void deleteData(ConfigNodeProcedureEnv env) { |
| deleteDataWithRawPathPattern(env); |
| } |
| |
| private void deleteDataWithRawPathPattern(ConfigNodeProcedureEnv env) { |
| executeDeleteData(env, patternTree); |
| if (isFailed()) { |
| return; |
| } |
| setNextState(DeleteTimeSeriesState.DELETE_TIMESERIES_SCHEMA); |
| } |
| |
| private void executeDeleteData(ConfigNodeProcedureEnv env, PathPatternTree patternTree) { |
| Map<TConsensusGroupId, TRegionReplicaSet> relatedDataRegionGroup = |
| env.getConfigManager().getRelatedDataRegionGroup(patternTree); |
| |
| // target timeSeries has no data |
| if (relatedDataRegionGroup.isEmpty()) { |
| return; |
| } |
| |
| DeleteTimeSeriesRegionTaskExecutor<TDeleteDataForDeleteSchemaReq> deleteDataTask = |
| new DeleteTimeSeriesRegionTaskExecutor<>( |
| "delete data", |
| env, |
| relatedDataRegionGroup, |
| true, |
| DataNodeRequestType.DELETE_DATA_FOR_DELETE_SCHEMA, |
| ((dataNodeLocation, consensusGroupIdList) -> |
| new TDeleteDataForDeleteSchemaReq( |
| new ArrayList<>(consensusGroupIdList), |
| preparePatternTreeBytesData(patternTree)) |
| .setIsGeneratedByPipe(isGeneratedByPipe))); |
| deleteDataTask.execute(); |
| } |
| |
| private void deleteTimeSeriesSchema(ConfigNodeProcedureEnv env) { |
| DeleteTimeSeriesRegionTaskExecutor<TDeleteTimeSeriesReq> deleteTimeSeriesTask = |
| new DeleteTimeSeriesRegionTaskExecutor<>( |
| "delete timeseries schemaengine", |
| env, |
| env.getConfigManager().getRelatedSchemaRegionGroup(patternTree), |
| DataNodeRequestType.DELETE_TIMESERIES, |
| ((dataNodeLocation, consensusGroupIdList) -> |
| new TDeleteTimeSeriesReq(consensusGroupIdList, patternTreeBytes) |
| .setIsGeneratedByPipe(isGeneratedByPipe))); |
| deleteTimeSeriesTask.execute(); |
| } |
| |
| private void collectPayload4Pipe(ConfigNodeProcedureEnv env) { |
| TSStatus result; |
| try { |
| result = |
| env.getConfigManager() |
| .getConsensusManager() |
| .write( |
| isGeneratedByPipe |
| ? new PipeEnrichedPlan(new PipeDeleteTimeSeriesPlan(patternTreeBytes)) |
| : new PipeDeleteTimeSeriesPlan(patternTreeBytes)); |
| } catch (ConsensusException e) { |
| LOGGER.warn(CONSENSUS_WRITE_ERROR, e); |
| result = new TSStatus(TSStatusCode.EXECUTE_STATEMENT_ERROR.getStatusCode()); |
| result.setMessage(e.getMessage()); |
| } |
| if (result.getCode() != TSStatusCode.SUCCESS_STATUS.getStatusCode()) { |
| throw new PipeException(result.getMessage()); |
| } |
| } |
| |
| @Override |
| protected void rollbackState( |
| ConfigNodeProcedureEnv env, DeleteTimeSeriesState deleteTimeSeriesState) |
| throws IOException, InterruptedException, ProcedureException { |
| if (deleteTimeSeriesState == DeleteTimeSeriesState.CONSTRUCT_BLACK_LIST) { |
| DeleteTimeSeriesRegionTaskExecutor<TRollbackSchemaBlackListReq> rollbackStateTask = |
| new DeleteTimeSeriesRegionTaskExecutor<>( |
| "roll back schemaengine black list", |
| env, |
| env.getConfigManager().getRelatedSchemaRegionGroup(patternTree), |
| DataNodeRequestType.ROLLBACK_SCHEMA_BLACK_LIST, |
| (dataNodeLocation, consensusGroupIdList) -> |
| new TRollbackSchemaBlackListReq(consensusGroupIdList, patternTreeBytes)); |
| rollbackStateTask.execute(); |
| } |
| } |
| |
| @Override |
| protected boolean isRollbackSupported(DeleteTimeSeriesState deleteTimeSeriesState) { |
| return true; |
| } |
| |
| @Override |
| protected DeleteTimeSeriesState getState(int stateId) { |
| return DeleteTimeSeriesState.values()[stateId]; |
| } |
| |
| @Override |
| protected int getStateId(DeleteTimeSeriesState deleteTimeSeriesState) { |
| return deleteTimeSeriesState.ordinal(); |
| } |
| |
| @Override |
| protected DeleteTimeSeriesState getInitialState() { |
| return DeleteTimeSeriesState.CONSTRUCT_BLACK_LIST; |
| } |
| |
| public String getQueryId() { |
| return queryId; |
| } |
| |
| public PathPatternTree getPatternTree() { |
| return patternTree; |
| } |
| |
| public void setPatternTree(PathPatternTree patternTree) { |
| this.patternTree = patternTree; |
| requestMessage = patternTree.getAllPathPatterns().toString(); |
| patternTreeBytes = preparePatternTreeBytesData(patternTree); |
| } |
| |
| private ByteBuffer preparePatternTreeBytesData(PathPatternTree patternTree) { |
| ByteArrayOutputStream byteArrayOutputStream = new ByteArrayOutputStream(); |
| DataOutputStream dataOutputStream = new DataOutputStream(byteArrayOutputStream); |
| try { |
| patternTree.serialize(dataOutputStream); |
| } catch (IOException ignored) { |
| |
| } |
| return ByteBuffer.wrap(byteArrayOutputStream.toByteArray()); |
| } |
| |
| @Override |
| public void serialize(DataOutputStream stream) throws IOException { |
| stream.writeShort( |
| isGeneratedByPipe |
| ? ProcedureType.PIPE_ENRICHED_DELETE_TIMESERIES_PROCEDURE.getTypeCode() |
| : ProcedureType.DELETE_TIMESERIES_PROCEDURE.getTypeCode()); |
| super.serialize(stream); |
| ReadWriteIOUtils.write(queryId, stream); |
| patternTree.serialize(stream); |
| } |
| |
| @Override |
| public void deserialize(ByteBuffer byteBuffer) { |
| super.deserialize(byteBuffer); |
| queryId = ReadWriteIOUtils.readString(byteBuffer); |
| setPatternTree(PathPatternTree.deserialize(byteBuffer)); |
| } |
| |
| @Override |
| public boolean equals(Object o) { |
| if (this == o) return true; |
| if (o == null || getClass() != o.getClass()) return false; |
| DeleteTimeSeriesProcedure that = (DeleteTimeSeriesProcedure) o; |
| return this.getProcId() == that.getProcId() |
| && this.getCurrentState().equals(that.getCurrentState()) |
| && this.getCycles() == getCycles() |
| && this.isGeneratedByPipe == that.isGeneratedByPipe |
| && this.patternTree.equals(that.patternTree); |
| } |
| |
| @Override |
| public int hashCode() { |
| return Objects.hash( |
| getProcId(), getCurrentState(), getCycles(), isGeneratedByPipe, patternTree); |
| } |
| |
| private class DeleteTimeSeriesRegionTaskExecutor<Q> |
| extends DataNodeRegionTaskExecutor<Q, TSStatus> { |
| |
| private final String taskName; |
| |
| DeleteTimeSeriesRegionTaskExecutor( |
| String taskName, |
| ConfigNodeProcedureEnv env, |
| Map<TConsensusGroupId, TRegionReplicaSet> targetSchemaRegionGroup, |
| DataNodeRequestType dataNodeRequestType, |
| BiFunction<TDataNodeLocation, List<TConsensusGroupId>, Q> dataNodeRequestGenerator) { |
| super(env, targetSchemaRegionGroup, false, dataNodeRequestType, dataNodeRequestGenerator); |
| this.taskName = taskName; |
| } |
| |
| DeleteTimeSeriesRegionTaskExecutor( |
| String taskName, |
| ConfigNodeProcedureEnv env, |
| Map<TConsensusGroupId, TRegionReplicaSet> targetSchemaRegionGroup, |
| boolean executeOnAllReplicaset, |
| DataNodeRequestType dataNodeRequestType, |
| BiFunction<TDataNodeLocation, List<TConsensusGroupId>, Q> dataNodeRequestGenerator) { |
| super( |
| env, |
| targetSchemaRegionGroup, |
| executeOnAllReplicaset, |
| dataNodeRequestType, |
| dataNodeRequestGenerator); |
| this.taskName = taskName; |
| } |
| |
| @Override |
| protected List<TConsensusGroupId> processResponseOfOneDataNode( |
| TDataNodeLocation dataNodeLocation, |
| List<TConsensusGroupId> consensusGroupIdList, |
| TSStatus response) { |
| List<TConsensusGroupId> failedRegionList = new ArrayList<>(); |
| if (response.getCode() == TSStatusCode.SUCCESS_STATUS.getStatusCode()) { |
| return failedRegionList; |
| } |
| |
| if (response.getCode() == TSStatusCode.MULTIPLE_ERROR.getStatusCode()) { |
| List<TSStatus> subStatus = response.getSubStatus(); |
| for (int i = 0; i < subStatus.size(); i++) { |
| if (subStatus.get(i).getCode() != TSStatusCode.SUCCESS_STATUS.getStatusCode()) { |
| failedRegionList.add(consensusGroupIdList.get(i)); |
| } |
| } |
| } else { |
| failedRegionList.addAll(consensusGroupIdList); |
| } |
| return failedRegionList; |
| } |
| |
| @Override |
| protected void onAllReplicasetFailure( |
| TConsensusGroupId consensusGroupId, Set<TDataNodeLocation> dataNodeLocationSet) { |
| setFailure( |
| new ProcedureException( |
| new MetadataException( |
| String.format( |
| "Delete timeseries %s failed when [%s] because all replicaset of schemaRegion %s failed. %s", |
| requestMessage, taskName, consensusGroupId.id, dataNodeLocationSet)))); |
| interruptTask(); |
| } |
| } |
| } |