| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.hudi.client; |
| |
| import org.apache.hudi.client.common.EngineProperty; |
| import org.apache.hudi.client.common.TaskContextSupplier; |
| import org.apache.hudi.common.util.Option; |
| import org.apache.hudi.exception.HoodieException; |
| |
| import org.apache.spark.SparkEnv; |
| import org.apache.spark.TaskContext; |
| import org.apache.spark.util.Utils; |
| |
| import java.io.Serializable; |
| import java.util.function.Supplier; |
| |
| /** |
| * Spark task context supplier. |
| */ |
| public class SparkTaskContextSupplier extends TaskContextSupplier implements Serializable { |
| |
| @Override |
| public Supplier<Integer> getPartitionIdSupplier() { |
| return TaskContext::getPartitionId; |
| } |
| |
| @Override |
| public Supplier<Integer> getStageIdSupplier() { |
| return () -> TaskContext.get().stageId(); |
| } |
| |
| @Override |
| public Supplier<Long> getAttemptIdSupplier() { |
| return () -> TaskContext.get().taskAttemptId(); |
| } |
| |
| @Override |
| public Option<String> getProperty(EngineProperty prop) { |
| if (prop == EngineProperty.TOTAL_MEMORY_AVAILABLE) { |
| // This is hard-coded in spark code {@link |
| // https://github.com/apache/spark/blob/576c43fb4226e4efa12189b41c3bc862019862c6/core/src/main/scala/org/apache/ |
| // spark/SparkContext.scala#L471} so have to re-define this here |
| final String DEFAULT_SPARK_EXECUTOR_MEMORY_MB = "1024"; // in MB |
| final String SPARK_EXECUTOR_MEMORY_PROP = "spark.executor.memory"; |
| if (SparkEnv.get() != null) { |
| // 1 GB is the default conf used by Spark, look at SparkContext.scala |
| return Option.ofNullable(String.valueOf(Utils.memoryStringToMb(SparkEnv.get().conf() |
| .get(SPARK_EXECUTOR_MEMORY_PROP, DEFAULT_SPARK_EXECUTOR_MEMORY_MB)) * 1024 * 1024L)); |
| } |
| return Option.empty(); |
| } else if (prop == EngineProperty.MEMORY_FRACTION_IN_USE) { |
| // This is hard-coded in spark code {@link |
| // https://github.com/apache/spark/blob/576c43fb4226e4efa12189b41c3bc862019862c6/core/src/main/scala/org/apache/ |
| // spark/memory/UnifiedMemoryManager.scala#L231} so have to re-define this here |
| final String DEFAULT_SPARK_EXECUTOR_MEMORY_FRACTION = "0.6"; |
| final String SPARK_EXECUTOR_MEMORY_FRACTION_PROP = "spark.memory.fraction"; |
| if (SparkEnv.get() != null) { |
| // 0.6 is the default value used by Spark, |
| // look at {@link |
| // https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/SparkConf.scala#L507} |
| return Option.ofNullable(SparkEnv.get().conf() |
| .get(SPARK_EXECUTOR_MEMORY_FRACTION_PROP, DEFAULT_SPARK_EXECUTOR_MEMORY_FRACTION)); |
| } |
| return Option.empty(); |
| } |
| throw new HoodieException("Unknown engine property :" + prop); |
| } |
| } |