| /** |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.hadoop.yarn.server.timelineservice.storage; |
| import static org.junit.Assert.assertEquals; |
| import static org.junit.Assert.assertTrue; |
| |
| import java.io.File; |
| import java.nio.charset.StandardCharsets; |
| import java.nio.file.Files; |
| import java.nio.file.Path; |
| import java.nio.file.Paths; |
| import java.util.HashMap; |
| import java.util.List; |
| import java.util.Map; |
| |
| import org.apache.hadoop.conf.Configuration; |
| import org.apache.hadoop.security.UserGroupInformation; |
| import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntities; |
| import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity; |
| import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric; |
| import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetricOperation; |
| import org.apache.hadoop.yarn.conf.YarnConfiguration; |
| import org.apache.hadoop.yarn.server.timelineservice.collector.TimelineCollectorContext; |
| import org.apache.hadoop.yarn.util.timeline.TimelineUtils; |
| import org.junit.Rule; |
| import org.junit.Test; |
| import org.junit.rules.TemporaryFolder; |
| |
| /** |
| * Tests filesystem implemetation for timelineservice writer. |
| */ |
| public class TestFileSystemTimelineWriterImpl { |
| @Rule |
| public TemporaryFolder tmpFolder = new TemporaryFolder(); |
| |
| /** |
| * Unit test for PoC YARN 3264. |
| * |
| * @throws Exception |
| */ |
| @Test |
| public void testWriteEntityToFile() throws Exception { |
| TimelineEntities te = new TimelineEntities(); |
| TimelineEntity entity = new TimelineEntity(); |
| String id = "hello"; |
| String type = "world"; |
| entity.setId(id); |
| entity.setType(type); |
| entity.setCreatedTime(1425016501000L); |
| te.addEntity(entity); |
| |
| TimelineMetric metric = new TimelineMetric(); |
| String metricId = "CPU"; |
| metric.setId(metricId); |
| metric.setType(TimelineMetric.Type.SINGLE_VALUE); |
| metric.setRealtimeAggregationOp(TimelineMetricOperation.SUM); |
| metric.addValue(1425016501000L, 1234567L); |
| |
| TimelineEntity entity2 = new TimelineEntity(); |
| String id2 = "metric"; |
| String type2 = "app"; |
| entity2.setId(id2); |
| entity2.setType(type2); |
| entity2.setCreatedTime(1425016503000L); |
| entity2.addMetric(metric); |
| te.addEntity(entity2); |
| |
| Map<String, TimelineMetric> aggregatedMetrics = |
| new HashMap<String, TimelineMetric>(); |
| aggregatedMetrics.put(metricId, metric); |
| |
| FileSystemTimelineWriterImpl fsi = null; |
| try { |
| fsi = new FileSystemTimelineWriterImpl(); |
| Configuration conf = new YarnConfiguration(); |
| String outputRoot = tmpFolder.newFolder().getAbsolutePath(); |
| conf.set(FileSystemTimelineWriterImpl.TIMELINE_SERVICE_STORAGE_DIR_ROOT, |
| outputRoot); |
| fsi.init(conf); |
| fsi.start(); |
| fsi.write( |
| new TimelineCollectorContext("cluster_id", "user_id", "flow_name", |
| "flow_version", 12345678L, "app_id"), |
| te, UserGroupInformation.createRemoteUser("user_id")); |
| |
| String fileName = fsi.getOutputRoot() + File.separator + "entities" + |
| File.separator + "cluster_id" + File.separator + "user_id" + |
| File.separator + "flow_name" + File.separator + "flow_version" + |
| File.separator + "12345678" + File.separator + "app_id" + |
| File.separator + type + File.separator + id + |
| FileSystemTimelineWriterImpl.TIMELINE_SERVICE_STORAGE_EXTENSION; |
| Path path = Paths.get(fileName); |
| File f = new File(fileName); |
| assertTrue(f.exists() && !f.isDirectory()); |
| List<String> data = Files.readAllLines(path, StandardCharsets.UTF_8); |
| // ensure there's only one entity + 1 new line |
| assertTrue("data size is:" + data.size(), data.size() == 2); |
| String d = data.get(0); |
| // confirm the contents same as what was written |
| assertEquals(d, TimelineUtils.dumpTimelineRecordtoJSON(entity)); |
| |
| // verify aggregated metrics |
| String fileName2 = fsi.getOutputRoot() + File.separator + "entities" + |
| File.separator + "cluster_id" + File.separator + "user_id" + |
| File.separator + "flow_name" + File.separator + "flow_version" + |
| File.separator + "12345678" + File.separator + "app_id" + |
| File.separator + type2 + File.separator + id2 + |
| FileSystemTimelineWriterImpl.TIMELINE_SERVICE_STORAGE_EXTENSION; |
| Path path2 = Paths.get(fileName2); |
| File file = new File(fileName2); |
| assertTrue(file.exists() && !file.isDirectory()); |
| List<String> data2 = Files.readAllLines(path2, StandardCharsets.UTF_8); |
| // ensure there's only one entity + 1 new line |
| assertTrue("data size is:" + data.size(), data2.size() == 2); |
| String metricToString = data2.get(0); |
| // confirm the contents same as what was written |
| assertEquals(metricToString, |
| TimelineUtils.dumpTimelineRecordtoJSON(entity2)); |
| } finally { |
| if (fsi != null) { |
| fsi.close(); |
| } |
| } |
| } |
| |
| } |