| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| package org.apache.sysds.runtime.io; |
| |
| import org.apache.hadoop.fs.FileSystem; |
| import org.apache.hadoop.fs.Path; |
| import org.apache.hadoop.io.LongWritable; |
| import org.apache.hadoop.io.Text; |
| import org.apache.hadoop.mapred.FileInputFormat; |
| import org.apache.hadoop.mapred.InputSplit; |
| import org.apache.hadoop.mapred.JobConf; |
| import org.apache.hadoop.mapred.RecordReader; |
| import org.apache.hadoop.mapred.Reporter; |
| import org.apache.hadoop.mapred.TextInputFormat; |
| import org.apache.sysds.common.Types.ValueType; |
| import org.apache.sysds.conf.ConfigurationManager; |
| import org.apache.sysds.runtime.DMLRuntimeException; |
| import org.apache.sysds.runtime.data.TensorBlock; |
| |
| import java.io.IOException; |
| import java.util.Arrays; |
| |
| public class TensorReaderTextCell extends TensorReader { |
| |
| @Override |
| public TensorBlock readTensorFromHDFS(String fname, long[] dims, |
| int blen, ValueType[] schema) throws IOException, DMLRuntimeException { |
| //prepare file access |
| JobConf job = new JobConf(ConfigurationManager.getCachedJobConf()); |
| Path path = new Path(fname); |
| FileSystem fs = IOUtilFunctions.getFileSystem(path, job); |
| |
| //check existence and non-empty file |
| checkValidInputFile(fs, path); |
| |
| //allocate output matrix block |
| return readTextCellTensorFromHDFS(path, job, dims, schema); |
| } |
| |
| protected TensorBlock readTextCellTensorFromHDFS(Path path, JobConf job, long[] dims, ValueType[] schema) throws IOException { |
| FileInputFormat.addInputPath(job, path); |
| TextInputFormat informat = new TextInputFormat(); |
| informat.configure(job); |
| InputSplit[] splits = informat.getSplits(job, 1); |
| |
| LongWritable key = new LongWritable(); |
| Text value = new Text(); |
| int[] idims = Arrays.stream(dims).mapToInt(i -> (int) i).toArray(); |
| TensorBlock ret; |
| if (schema.length == 1) |
| ret = new TensorBlock(schema[0], idims).allocateBlock(); |
| else |
| ret = new TensorBlock(schema, idims).allocateBlock(); |
| |
| try { |
| int[] ix = new int[dims.length]; |
| for (InputSplit split : splits) { |
| RecordReader<LongWritable, Text> reader = informat.getRecordReader(split, job, Reporter.NULL); |
| try { |
| while (reader.next(key, value)) { |
| String[] parts = Arrays.stream(IOUtilFunctions.splitCSV(value.toString(), " ")) |
| .filter(s -> !s.isEmpty()).toArray(String[]::new); |
| for (int i = 0; i < ix.length; i++) { |
| ix[i] = Integer.parseInt(parts[i]) - 1; |
| } |
| ret.set(ix, parts[ix.length]); |
| } |
| } |
| finally { |
| IOUtilFunctions.closeSilently(reader); |
| } |
| } |
| } |
| catch (Exception ex) { |
| throw new IOException("Unable to read tensor in text cell format.", ex); |
| } |
| return ret; |
| } |
| } |