blob: 7586c70fe3d401f0d4749449ec5f893900cdf33d [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package com.datatorrent.lib.stream;
import org.junit.Assert;
import org.junit.Test;
import com.datatorrent.lib.testbench.CountTestSink;
/**
* Performance test for {@link com.datatorrent.lib.stream.StreamDuplicater}<p>
* Benchmarks: Currently does about ?? Million tuples/sec in debugging environment. Need to test on larger nodes<br>
* <br>
*/
public class StreamDuplicaterTest
{
/**
* Test oper pass through. The Object passed is not relevant
*/
@SuppressWarnings({"rawtypes", "unchecked"})
@Test
public void testNodeProcessing() throws Exception
{
StreamDuplicater oper = new StreamDuplicater();
CountTestSink mergeSink1 = new CountTestSink();
CountTestSink mergeSink2 = new CountTestSink();
oper.out1.setSink(mergeSink1);
oper.out2.setSink(mergeSink2);
oper.beginWindow(0);
int numtuples = 1000;
Integer input = 0;
// Same input object can be used as the oper is just pass through
for (int i = 0; i < numtuples; i++) {
oper.data.process(input);
}
oper.endWindow();
// One for each key
Assert.assertEquals("number emitted tuples", numtuples, mergeSink1.count);
Assert.assertEquals("number emitted tuples", numtuples, mergeSink2.count);
}
}