blob: b3a30658e24cda5298c69bf3ceb554af3f47147b [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.hyracks.control.cc.work;
import java.util.HashSet;
import java.util.Map;
import java.util.Set;
import java.util.logging.Level;
import java.util.logging.Logger;
import org.apache.hyracks.api.job.JobId;
import org.apache.hyracks.control.cc.ClusterControllerService;
import org.apache.hyracks.control.cc.NodeControllerState;
import org.apache.hyracks.control.cc.job.JobRun;
import org.apache.hyracks.control.common.work.AbstractWork;
public class RemoveDeadNodesWork extends AbstractWork {
private static Logger LOGGER = Logger.getLogger(RemoveDeadNodesWork.class.getName());
private final ClusterControllerService ccs;
public RemoveDeadNodesWork(ClusterControllerService ccs) {
this.ccs = ccs;
}
@Override
public void run() {
Set<String> deadNodes = new HashSet<String>();
Map<String, NodeControllerState> nodeMap = ccs.getNodeMap();
for (Map.Entry<String, NodeControllerState> e : nodeMap.entrySet()) {
NodeControllerState state = e.getValue();
if (state.incrementLastHeartbeatDuration() >= ccs.getCCConfig().maxHeartbeatLapsePeriods) {
deadNodes.add(e.getKey());
LOGGER.info(e.getKey() + " considered dead");
}
}
Set<JobId> affectedJobIds = new HashSet<JobId>();
for (String deadNode : deadNodes) {
NodeControllerState state = nodeMap.remove(deadNode);
// Deal with dead tasks.
affectedJobIds.addAll(state.getActiveJobIds());
}
int size = affectedJobIds.size();
if (size > 0) {
if (LOGGER.isLoggable(Level.INFO)) {
LOGGER.info("Number of affected jobs: " + size);
}
for (JobId jobId : affectedJobIds) {
JobRun run = ccs.getActiveRunMap().get(jobId);
if (run != null) {
run.getScheduler().notifyNodeFailures(deadNodes);
}
}
}
if (deadNodes != null && deadNodes.size() > 0) {
ccs.getApplicationContext().notifyNodeFailure(deadNodes);
}
}
@Override
public Level logLevel() {
return Level.FINE;
}
}