| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| **/ |
| |
| #define __STDC_FORMAT_MACROS |
| |
| #include "storage/FileManagerHdfs.hpp" |
| |
| #include <hdfs/hdfs.h> |
| |
| #include <cerrno> |
| #include <cinttypes> |
| #include <cstddef> |
| #include <cstdint> |
| #include <cstdio> |
| #include <cstring> |
| #include <string> |
| |
| #include "storage/Flags.hpp" |
| #include "storage/StorageBlockInfo.hpp" |
| #include "storage/StorageConstants.hpp" |
| #include "storage/StorageErrors.hpp" |
| #include "utility/StringUtil.hpp" |
| |
| #include "glog/logging.h" |
| |
| using std::size_t; |
| using std::sscanf; |
| using std::strerror; |
| using std::string; |
| |
| namespace quickstep { |
| |
| FileManagerHdfs::FileManagerHdfs(const string &storage_path) |
| : FileManager(storage_path) { |
| struct hdfsBuilder *builder = hdfsNewBuilder(); |
| hdfsBuilderSetNameNode(builder, FLAGS_hdfs_namenode_host.c_str()); |
| hdfsBuilderSetNameNodePort(builder, FLAGS_hdfs_namenode_port); |
| // hdfsBuilderConnect releases builder. |
| hdfs_ = hdfsBuilderConnect(builder); |
| DCHECK(hdfs_ != nullptr); |
| } |
| |
| FileManagerHdfs::~FileManagerHdfs() { |
| CHECK_EQ(0, hdfsDisconnect(hdfs_)); |
| } |
| |
| block_id_counter FileManagerHdfs::getMaxUsedBlockCounter(const block_id_domain block_domain) const { |
| int num_files = 0; |
| hdfsFileInfo *file_infos = hdfsListDirectory(hdfs_, storage_path_.c_str(), &num_files); |
| if (file_infos == nullptr) { |
| if (errno != ENOENT) { |
| LOG(ERROR) << "Failed to list file info with error: " << strerror(errno); |
| } |
| return 0; |
| } |
| |
| string filename_pattern("/qsblk_"); |
| filename_pattern.append(ToZeroPaddedString(block_domain, kBlockIdDomainLengthInDigits)); |
| filename_pattern.append("_%"); |
| filename_pattern.append(SCNu64); |
| filename_pattern.append(".qsb"); |
| |
| block_id_counter counter_max = 0, counter; |
| for (int i = 0; i < num_files; ++i) { |
| // NOTE(zuyu): mName looks like |
| // "/user/<username>/<storage_path_>/qsblk_<block_domain>_[0-9]*.qsb". |
| const char *filename = std::strrchr(file_infos[i].mName, '/'); |
| if (filename != nullptr && |
| sscanf(filename, filename_pattern.c_str(), &counter) == 1 && |
| counter > counter_max) { |
| counter_max = counter; |
| } |
| } |
| |
| hdfsFreeFileInfo(file_infos, num_files); |
| |
| return counter_max; |
| } |
| |
| size_t FileManagerHdfs::numSlots(const block_id block) const { |
| const string filename(blockFilename(block)); |
| |
| hdfsFileInfo *file_info = hdfsGetPathInfo(hdfs_, filename.c_str()); |
| if (file_info == nullptr) { |
| if (errno != ENOENT) { |
| LOG(ERROR) << "Failed to get size of file " << filename << " with error: " << strerror(errno); |
| } |
| return 0; |
| } |
| |
| size_t file_size = file_info->mSize; |
| hdfsFreeFileInfo(file_info, 1); |
| |
| if ((file_size % kSlotSizeBytes) != 0) { |
| throw CorruptPersistentStorage(); |
| } |
| |
| return file_size / kSlotSizeBytes; |
| } |
| |
| bool FileManagerHdfs::deleteBlockOrBlob(const block_id block) { |
| const string filename(blockFilename(block)); |
| |
| if (hdfsDelete(hdfs_, filename.c_str(), 0)) { |
| switch (errno) { |
| case EINPROGRESS: |
| case ENOENT: |
| break; |
| default: |
| LOG(ERROR) << "Failed to delete file " << filename << " with error: " << strerror(errno); |
| } |
| } |
| |
| return true; |
| } |
| |
| bool FileManagerHdfs::readBlockOrBlob(const block_id block, |
| void *buffer, |
| const size_t length) { |
| DCHECK(buffer != nullptr); |
| DCHECK_EQ(0u, length % kSlotSizeBytes); |
| |
| const string filename(blockFilename(block)); |
| |
| hdfsFile file_handle = hdfsOpenFile(hdfs_, |
| filename.c_str(), |
| O_RDONLY, |
| kSlotSizeBytes, |
| FLAGS_hdfs_num_replications, |
| kSlotSizeBytes); |
| if (file_handle == nullptr) { |
| LOG(ERROR) << "Failed to open file " << filename << " with error: " << strerror(errno); |
| return false; |
| } |
| |
| size_t bytes_total = 0; |
| while (bytes_total < length) { |
| tSize bytes = hdfsRead(hdfs_, file_handle, static_cast<char*>(buffer) + bytes_total, length - bytes_total); |
| if (bytes > 0) { |
| bytes_total += bytes; |
| } else if (bytes == -1) { |
| if (errno != EINTR) { |
| LOG(ERROR) << "Failed to read file " << filename << " with error: " << strerror(errno); |
| break; |
| } |
| } else { |
| LOG(ERROR) << "Failed to read file " << filename << " since EOF was reached unexpectedly"; |
| break; |
| } |
| } |
| |
| if (hdfsCloseFile(hdfs_, file_handle) != 0) { |
| LOG(ERROR) << "Failed to close file " << filename << " with error: " << strerror(errno); |
| } |
| |
| return (bytes_total == length); |
| } |
| |
| bool FileManagerHdfs::writeBlockOrBlob(const block_id block, |
| const void *buffer, |
| const size_t length) { |
| DCHECK(buffer != nullptr); |
| DCHECK_EQ(0u, length % kSlotSizeBytes); |
| |
| const string filename(blockFilename(block)); |
| |
| hdfsFile file_handle = hdfsOpenFile(hdfs_, |
| filename.c_str(), |
| O_WRONLY, |
| kSlotSizeBytes, |
| FLAGS_hdfs_num_replications, |
| kSlotSizeBytes); |
| if (file_handle == nullptr) { |
| LOG(ERROR) << "Failed to open file " << filename << " with error: " << strerror(errno); |
| return false; |
| } |
| |
| size_t bytes_total = 0; |
| while (bytes_total < length) { |
| tSize bytes = hdfsWrite(hdfs_, file_handle, static_cast<const char*>(buffer) + bytes_total, length - bytes_total); |
| if (bytes > 0) { |
| bytes_total += bytes; |
| } else if (bytes == -1) { |
| LOG(ERROR) << "Failed to write file " << filename << " with error: " << strerror(errno); |
| break; |
| } |
| } |
| |
| if (hdfsSync(hdfs_, file_handle) != 0) { |
| LOG(ERROR) << "Failed to sync file " << filename << " with error: " << strerror(errno); |
| } |
| |
| if (hdfsCloseFile(hdfs_, file_handle) != 0) { |
| LOG(ERROR) << "Failed to close file " << filename << " with error: " << strerror(errno); |
| } |
| |
| return (bytes_total == length); |
| } |
| |
| } // namespace quickstep |