| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.hudi.table; |
| |
| import org.apache.hudi.client.WriteStatus; |
| import org.apache.hudi.client.common.HoodieEngineContext; |
| import org.apache.hudi.client.common.HoodieSparkEngineContext; |
| import org.apache.hudi.common.model.HoodieKey; |
| import org.apache.hudi.common.model.HoodieRecord; |
| import org.apache.hudi.common.model.HoodieRecordPayload; |
| import org.apache.hudi.common.table.HoodieTableMetaClient; |
| import org.apache.hudi.common.table.timeline.versioning.TimelineLayoutVersion; |
| import org.apache.hudi.common.util.Option; |
| import org.apache.hudi.config.HoodieWriteConfig; |
| import org.apache.hudi.exception.HoodieException; |
| import org.apache.hudi.index.HoodieIndex; |
| import org.apache.hudi.index.SparkHoodieIndex; |
| |
| import org.apache.spark.api.java.JavaRDD; |
| |
| public abstract class HoodieSparkTable<T extends HoodieRecordPayload> |
| extends HoodieTable<T, JavaRDD<HoodieRecord<T>>, JavaRDD<HoodieKey>, JavaRDD<WriteStatus>> { |
| |
| protected HoodieSparkTable(HoodieWriteConfig config, HoodieEngineContext context, HoodieTableMetaClient metaClient) { |
| super(config, context, metaClient); |
| } |
| |
| public static <T extends HoodieRecordPayload> HoodieSparkTable<T> create(HoodieWriteConfig config, HoodieEngineContext context) { |
| HoodieTableMetaClient metaClient = new HoodieTableMetaClient( |
| context.getHadoopConf().get(), |
| config.getBasePath(), |
| true, |
| config.getConsistencyGuardConfig(), |
| Option.of(new TimelineLayoutVersion(config.getTimelineLayoutVersion())) |
| ); |
| return HoodieSparkTable.create(config, (HoodieSparkEngineContext) context, metaClient); |
| } |
| |
| public static <T extends HoodieRecordPayload> HoodieSparkTable<T> create(HoodieWriteConfig config, |
| HoodieSparkEngineContext context, |
| HoodieTableMetaClient metaClient) { |
| switch (metaClient.getTableType()) { |
| case COPY_ON_WRITE: |
| return new HoodieSparkCopyOnWriteTable<>(config, context, metaClient); |
| case MERGE_ON_READ: |
| return new HoodieSparkMergeOnReadTable<>(config, context, metaClient); |
| default: |
| throw new HoodieException("Unsupported table type :" + metaClient.getTableType()); |
| } |
| } |
| |
| @Override |
| protected HoodieIndex<T, JavaRDD<HoodieRecord<T>>, JavaRDD<HoodieKey>, JavaRDD<WriteStatus>> getIndex(HoodieWriteConfig config) { |
| return SparkHoodieIndex.createIndex(config); |
| } |
| } |