blob: 62ae5b18e4a2fd14a6a0fd19c13b5ae8df978b7f [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package com.epam.datalab.backendapi.service.impl;
import com.epam.datalab.auth.UserInfo;
import com.epam.datalab.backendapi.core.Directories;
import com.epam.datalab.backendapi.core.FileHandlerCallback;
import com.epam.datalab.backendapi.core.commands.DockerAction;
import com.epam.datalab.backendapi.core.commands.DockerCommands;
import com.epam.datalab.backendapi.core.commands.RunDockerCommand;
import com.epam.datalab.backendapi.core.response.handlers.ComputationalCallbackHandler;
import com.epam.datalab.backendapi.core.response.handlers.ComputationalConfigure;
import com.epam.datalab.cloud.CloudProvider;
import com.epam.datalab.dto.base.DataEngineType;
import com.epam.datalab.dto.base.computational.ComputationalBase;
import com.epam.datalab.dto.computational.ComputationalClusterConfigDTO;
import com.epam.datalab.dto.computational.ComputationalStartDTO;
import com.epam.datalab.dto.computational.ComputationalStopDTO;
import com.epam.datalab.dto.computational.ComputationalTerminateDTO;
import com.epam.datalab.exceptions.DatalabException;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.google.inject.Inject;
import com.google.inject.Singleton;
import java.util.Objects;
import static com.epam.datalab.backendapi.core.commands.DockerAction.CREATE;
import static com.epam.datalab.backendapi.core.commands.DockerAction.RECONFIGURE_SPARK;
import static com.epam.datalab.backendapi.core.commands.DockerAction.START;
import static com.epam.datalab.backendapi.core.commands.DockerAction.STOP;
import static com.epam.datalab.backendapi.core.commands.DockerAction.TERMINATE;
@Singleton
public class SparkClusterService extends DockerService implements DockerCommands {
private static final DataEngineType SPARK_ENGINE = DataEngineType.SPARK_STANDALONE;
@Inject
private ComputationalConfigure computationalConfigure;
public String create(UserInfo ui, ComputationalBase<?> dto) {
return action(ui, dto, CREATE);
}
public String terminate(UserInfo ui, ComputationalTerminateDTO dto) {
return action(ui, dto, TERMINATE);
}
public String stop(UserInfo ui, ComputationalStopDTO dto) {
return action(ui, dto, STOP);
}
public String start(UserInfo ui, ComputationalStartDTO dto) {
return action(ui, dto, START);
}
public String updateConfig(UserInfo ui, ComputationalClusterConfigDTO clusterConfigDTO) {
String uuid = DockerCommands.generateUUID();
folderListenerExecutor.start(configuration.getImagesDirectory(),
configuration.getResourceStatusPollTimeout(),
getFileHandlerCallback(RECONFIGURE_SPARK, uuid, clusterConfigDTO));
runReconfigureSparkDockerCommand(ui, clusterConfigDTO, uuid);
return uuid;
}
private void runReconfigureSparkDockerCommand(UserInfo ui, ComputationalClusterConfigDTO clusterConfigDTO,
String uuid) {
try {
final RunDockerCommand runDockerCommand = new RunDockerCommand()
.withInteractive()
.withName(nameContainer(clusterConfigDTO.getEdgeUserName(), RECONFIGURE_SPARK,
clusterConfigDTO.getExploratoryName(),
clusterConfigDTO.getComputationalName()))
.withVolumeForRootKeys(configuration.getKeyDirectory())
.withVolumeForResponse(configuration.getImagesDirectory())
.withVolumeForLog(configuration.getDockerLogDirectory(), SPARK_ENGINE.getName())
.withResource(SPARK_ENGINE.getName())
.withRequestId(uuid)
.withConfKeyName(configuration.getAdminKey())
.withImage(DataEngineType.getDockerImageName(SPARK_ENGINE))
.withAction(RECONFIGURE_SPARK);
if (configuration.getCloudProvider() == CloudProvider.AZURE &&
Objects.nonNull(configuration.getCloudConfiguration().getAzureAuthFile()) &&
!configuration.getCloudConfiguration().getAzureAuthFile().isEmpty()) {
runDockerCommand.withVolumeFoAzureAuthFile(configuration.getCloudConfiguration().getAzureAuthFile());
}
commandExecutor.executeAsync(ui.getName(), uuid, commandBuilder.buildCommand(runDockerCommand,
clusterConfigDTO));
} catch (JsonProcessingException e) {
throw new DatalabException("Could not" + RECONFIGURE_SPARK.toString() + "computational resources cluster", e);
}
}
private String action(UserInfo ui, ComputationalBase<?> dto, DockerAction action) {
String uuid = DockerCommands.generateUUID();
folderListenerExecutor.start(configuration.getImagesDirectory(),
configuration.getResourceStatusPollTimeout(),
getFileHandlerCallback(action, uuid, dto));
try {
final RunDockerCommand runDockerCommand = new RunDockerCommand()
.withInteractive()
.withName(nameContainer(dto.getEdgeUserName(), action, dto.getExploratoryName(),
dto.getComputationalName()))
.withVolumeForRootKeys(configuration.getKeyDirectory())
.withVolumeForResponse(configuration.getImagesDirectory())
.withVolumeForLog(configuration.getDockerLogDirectory(), SPARK_ENGINE.getName())
.withResource(SPARK_ENGINE.getName())
.withRequestId(uuid)
.withConfKeyName(configuration.getAdminKey())
.withImage(DataEngineType.getDockerImageName(SPARK_ENGINE))
.withAction(action);
if (configuration.getCloudProvider() == CloudProvider.AZURE &&
Objects.nonNull(configuration.getCloudConfiguration().getAzureAuthFile()) &&
!configuration.getCloudConfiguration().getAzureAuthFile().isEmpty()) {
runDockerCommand.withVolumeFoAzureAuthFile(configuration.getCloudConfiguration().getAzureAuthFile());
}
commandExecutor.executeAsync(ui.getName(), uuid, commandBuilder.buildCommand(runDockerCommand, dto));
} catch (JsonProcessingException e) {
throw new DatalabException("Could not" + action.toString() + "computational resources cluster", e);
}
return uuid;
}
private FileHandlerCallback getFileHandlerCallback(DockerAction action, String uuid, ComputationalBase<?> dto) {
return new ComputationalCallbackHandler(computationalConfigure, selfService, action, uuid, dto);
}
private String nameContainer(String user, DockerAction action, String exploratoryName, String name) {
return nameContainer(user, action.toString(), "computational", exploratoryName, name);
}
@Override
public String getResourceType() {
return Directories.DATA_ENGINE_LOG_DIRECTORY;
}
}