Skip to content

Commit

Permalink
[FLINK-29050][test-utils] Rename AbstractTestBase, JavaProgramTestBas…
Browse files Browse the repository at this point in the history
…e MultipleProgramsTestBase with JUnit4 suffix.
  • Loading branch information
RocMarshal authored and 1996fanrui committed Apr 25, 2024
1 parent 0953199 commit ffa639a
Show file tree
Hide file tree
Showing 186 changed files with 396 additions and 394 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@
import org.apache.flink.streaming.api.datastream.DataStream;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import org.apache.flink.streaming.api.functions.sink.RichSinkFunction;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.util.FlinkRuntimeException;

import org.junit.Test;
Expand All @@ -46,7 +46,7 @@
import static org.assertj.core.api.Assertions.assertThat;

/** IT case for the {@link Source} with a coordinator. */
public class CoordinatedSourceITCase extends AbstractTestBase {
public class CoordinatedSourceITCase extends AbstractTestBaseJUnit4 {

@Test
public void testEnumeratorReaderCommunication() throws Exception {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@
import org.apache.flink.table.planner.runtime.utils.BatchTestBase;
import org.apache.flink.table.planner.runtime.utils.TestingRetractSink;
import org.apache.flink.table.planner.utils.JavaScalaConversionUtil;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.types.Row;
import org.apache.flink.util.CollectionUtil;
import org.apache.flink.util.FileUtils;
Expand Down Expand Up @@ -75,7 +75,7 @@
* IT case for HiveCatalog. TODO: move to flink-connector-hive-test end-to-end test module once it's
* setup
*/
public class HiveCatalogUdfITCase extends AbstractTestBase {
public class HiveCatalogUdfITCase extends AbstractTestBaseJUnit4 {

@Rule public TemporaryFolder tempFolder = new TemporaryFolder();

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
import org.apache.flink.api.java.ExecutionEnvironment;
import org.apache.flink.api.java.hadoop.mapred.HadoopInputFormat;
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.test.util.JavaProgramTestBase;
import org.apache.flink.test.util.JavaProgramTestBaseJUnit4;
import org.apache.flink.util.OperatingSystem;

import org.apache.hadoop.fs.FileSystem;
Expand All @@ -50,7 +50,7 @@

/** Integration tests for Hadoop IO formats. */
@RunWith(Parameterized.class)
public class HadoopIOFormatsITCase extends JavaProgramTestBase {
public class HadoopIOFormatsITCase extends JavaProgramTestBaseJUnit4 {

private static final int NUM_PROGRAMS = 2;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.core.fs.FileSystem;
import org.apache.flink.hadoopcompatibility.mapred.HadoopMapFunction;
import org.apache.flink.test.util.MultipleProgramsTestBase;
import org.apache.flink.test.util.MultipleProgramsTestBaseJUnit4;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
Expand All @@ -43,7 +43,7 @@

/** IT cases for the {@link HadoopMapFunction}. */
@RunWith(Parameterized.class)
public class HadoopMapFunctionITCase extends MultipleProgramsTestBase {
public class HadoopMapFunctionITCase extends MultipleProgramsTestBaseJUnit4 {

public HadoopMapFunctionITCase(TestExecutionMode mode) {
super(mode);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@

import org.apache.flink.test.hadoopcompatibility.mapred.example.HadoopMapredCompatWordCount;
import org.apache.flink.test.testdata.WordCountData;
import org.apache.flink.test.util.JavaProgramTestBase;
import org.apache.flink.test.util.JavaProgramTestBaseJUnit4;
import org.apache.flink.util.OperatingSystem;

import org.junit.Assume;
Expand All @@ -29,7 +29,7 @@
import static org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory;

/** IT cases for mapred. */
public class HadoopMapredITCase extends JavaProgramTestBase {
public class HadoopMapredITCase extends JavaProgramTestBaseJUnit4 {

protected String textPath;
protected String resultPath;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.hadoopcompatibility.mapred.HadoopReduceCombineFunction;
import org.apache.flink.hadoopcompatibility.mapred.HadoopReduceFunction;
import org.apache.flink.test.util.MultipleProgramsTestBase;
import org.apache.flink.test.util.MultipleProgramsTestBaseJUnit4;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
Expand All @@ -46,7 +46,7 @@

/** IT case for the {@link HadoopReduceCombineFunction}. */
@RunWith(Parameterized.class)
public class HadoopReduceCombineFunctionITCase extends MultipleProgramsTestBase {
public class HadoopReduceCombineFunctionITCase extends MultipleProgramsTestBaseJUnit4 {

public HadoopReduceCombineFunctionITCase(TestExecutionMode mode) {
super(mode);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
import org.apache.flink.api.java.ExecutionEnvironment;
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.hadoopcompatibility.mapred.HadoopReduceFunction;
import org.apache.flink.test.util.MultipleProgramsTestBase;
import org.apache.flink.test.util.MultipleProgramsTestBaseJUnit4;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
Expand All @@ -44,7 +44,7 @@

/** IT cases for the {@link HadoopReduceFunction}. */
@RunWith(Parameterized.class)
public class HadoopReduceFunctionITCase extends MultipleProgramsTestBase {
public class HadoopReduceFunctionITCase extends MultipleProgramsTestBaseJUnit4 {

public HadoopReduceFunctionITCase(TestExecutionMode mode) {
super(mode);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.hadoopcompatibility.HadoopInputs;
import org.apache.flink.test.testdata.WordCountData;
import org.apache.flink.test.util.JavaProgramTestBase;
import org.apache.flink.test.util.JavaProgramTestBaseJUnit4;
import org.apache.flink.util.Collector;
import org.apache.flink.util.OperatingSystem;

Expand All @@ -42,7 +42,7 @@
import static org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory;

/** Test WordCount with Hadoop input and output "mapred" (legacy) formats. */
public class WordCountMapredITCase extends JavaProgramTestBase {
public class WordCountMapredITCase extends JavaProgramTestBaseJUnit4 {

protected String textPath;
protected String resultPath;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
import org.apache.flink.api.java.hadoop.mapreduce.HadoopOutputFormat;
import org.apache.flink.test.hadoopcompatibility.mapreduce.example.WordCount;
import org.apache.flink.test.testdata.WordCountData;
import org.apache.flink.test.util.JavaProgramTestBase;
import org.apache.flink.test.util.JavaProgramTestBaseJUnit4;
import org.apache.flink.util.OperatingSystem;

import org.junit.Assume;
Expand All @@ -31,7 +31,7 @@
import static org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory;

/** IT cases for both the {@link HadoopInputFormat} and {@link HadoopOutputFormat}. */
public class HadoopInputOutputITCase extends JavaProgramTestBase {
public class HadoopInputOutputITCase extends JavaProgramTestBaseJUnit4 {

protected String textPath;
protected String resultPath;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.hadoopcompatibility.HadoopInputs;
import org.apache.flink.test.testdata.WordCountData;
import org.apache.flink.test.util.JavaProgramTestBase;
import org.apache.flink.test.util.JavaProgramTestBaseJUnit4;
import org.apache.flink.util.Collector;
import org.apache.flink.util.OperatingSystem;

Expand All @@ -42,7 +42,7 @@
import static org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory;

/** Test WordCount with Hadoop input and output "mapreduce" (modern) formats. */
public class WordCountMapreduceITCase extends JavaProgramTestBase {
public class WordCountMapreduceITCase extends JavaProgramTestBaseJUnit4 {

protected String textPath;
protected String resultPath;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,15 +21,15 @@ import org.apache.flink.api.scala._
import org.apache.flink.api.scala.hadoop.mapred.HadoopOutputFormat
import org.apache.flink.hadoopcompatibility.scala.HadoopInputs
import org.apache.flink.test.testdata.WordCountData
import org.apache.flink.test.util.{JavaProgramTestBase, TestBaseUtils}
import org.apache.flink.test.util.{JavaProgramTestBaseJUnit4, TestBaseUtils}
import org.apache.flink.util.OperatingSystem

import org.apache.hadoop.fs.Path
import org.apache.hadoop.io.{LongWritable, Text}
import org.apache.hadoop.mapred.{FileOutputFormat, JobConf, TextInputFormat, TextOutputFormat}
import org.junit.{Assume, Before}

class WordCountMapredITCase extends JavaProgramTestBase {
class WordCountMapredITCase extends JavaProgramTestBaseJUnit4 {
protected var textPath: String = null
protected var resultPath: String = null

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ import org.apache.flink.api.scala._
import org.apache.flink.api.scala.hadoop.mapreduce.HadoopOutputFormat
import org.apache.flink.hadoopcompatibility.scala.HadoopInputs
import org.apache.flink.test.testdata.WordCountData
import org.apache.flink.test.util.{JavaProgramTestBase, TestBaseUtils}
import org.apache.flink.test.util.{JavaProgramTestBaseJUnit4, TestBaseUtils}
import org.apache.flink.util.OperatingSystem

import org.apache.hadoop.fs.Path
Expand All @@ -31,7 +31,7 @@ import org.apache.hadoop.mapreduce.lib.input.TextInputFormat
import org.apache.hadoop.mapreduce.lib.output.{FileOutputFormat, TextOutputFormat}
import org.junit.{Assume, Before}

class WordCountMapreduceITCase extends JavaProgramTestBase {
class WordCountMapreduceITCase extends JavaProgramTestBaseJUnit4 {
protected var textPath: String = null
protected var resultPath: String = null

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@
import org.apache.flink.streaming.examples.iteration.util.IterateExampleData;
import org.apache.flink.streaming.test.examples.join.WindowJoinData;
import org.apache.flink.test.testdata.WordCountData;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;

import org.apache.commons.io.FileUtils;
import org.junit.Test;
Expand All @@ -43,7 +43,7 @@
import static org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory;

/** Integration test for streaming programs in Java examples. */
public class StreamingExamplesITCase extends AbstractTestBase {
public class StreamingExamplesITCase extends AbstractTestBaseJUnit4 {

@Test
public void testIterateExample() throws Exception {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@
import org.apache.flink.configuration.ConfigConstants;
import org.apache.flink.streaming.examples.socket.SocketWindowWordCount;
import org.apache.flink.test.testdata.WordCountData;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.util.NetUtils;

import org.junit.Test;
Expand All @@ -37,7 +37,7 @@
import static org.junit.Assert.fail;

/** Tests for {@link SocketWindowWordCount}. */
public class SocketWindowWordCountITCase extends AbstractTestBase {
public class SocketWindowWordCountITCase extends AbstractTestBaseJUnit4 {

@Test
public void testJavaProgram() throws Exception {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
import org.apache.flink.formats.avro.generated.Colors;
import org.apache.flink.formats.avro.generated.Fixed2;
import org.apache.flink.formats.avro.generated.User;
import org.apache.flink.test.util.JavaProgramTestBase;
import org.apache.flink.test.util.JavaProgramTestBaseJUnit4;

import org.apache.avro.file.DataFileReader;
import org.apache.avro.io.DatumReader;
Expand All @@ -50,7 +50,7 @@

/** IT cases for the {@link AvroOutputFormat}. */
@SuppressWarnings("serial")
public class AvroOutputFormatITCase extends JavaProgramTestBase {
public class AvroOutputFormatITCase extends JavaProgramTestBaseJUnit4 {

public static String outputPath1;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink;
import org.apache.flink.streaming.api.functions.sink.filesystem.bucketassigners.UniqueBucketAssigner;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;

import org.apache.avro.Schema;
import org.apache.avro.file.DataFileReader;
Expand Down Expand Up @@ -58,7 +58,7 @@
* Simple integration test case for writing bulk encoded files with the {@link StreamingFileSink}
* with Avro.
*/
public class AvroStreamingFileSinkITCase extends AbstractTestBase {
public class AvroStreamingFileSinkITCase extends AbstractTestBaseJUnit4 {

@Rule public final Timeout timeoutPerTest = Timeout.seconds(20);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@
import org.apache.flink.table.api.Table;
import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
import org.apache.flink.table.expressions.Expression;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.test.util.TestBaseUtils;
import org.apache.flink.types.Row;
import org.apache.flink.util.CollectionUtil;
Expand All @@ -56,7 +56,7 @@
import static org.assertj.core.api.Assertions.assertThat;

/** Tests for interoperability with Avro types. */
public class AvroTypesITCase extends AbstractTestBase {
public class AvroTypesITCase extends AbstractTestBaseJUnit4 {

private static final User USER_1 =
User.newBuilder()
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
import org.apache.flink.table.planner.factories.TestValuesTableFactory;
import org.apache.flink.table.planner.runtime.utils.TestData;
import org.apache.flink.table.planner.utils.JavaScalaConversionUtil;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.types.Row;

import org.junit.After;
Expand Down Expand Up @@ -58,7 +58,7 @@
import static org.assertj.core.api.Assertions.assertThat;

/** Tests for the CSV file format. */
public class TableCsvFormatITCase extends AbstractTestBase {
public class TableCsvFormatITCase extends AbstractTestBaseJUnit4 {

@Rule public ExpectedException exception = ExpectedException.none();

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@

package org.apache.flink.formats.hadoop.bulk;

import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.util.IOUtils;

import org.apache.hadoop.conf.Configuration;
Expand Down Expand Up @@ -46,7 +46,7 @@

/** Tests the behaviors of {@link HadoopFileCommitter}. */
@RunWith(Parameterized.class)
public abstract class AbstractFileCommitterTest extends AbstractTestBase {
public abstract class AbstractFileCommitterTest extends AbstractTestBaseJUnit4 {

private static final List<String> CONTENTS =
new ArrayList<>(Arrays.asList("first line", "second line", "third line"));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
import org.apache.flink.streaming.api.functions.sink.filesystem.HadoopPathBasedBulkFormatBuilder;
import org.apache.flink.streaming.api.functions.sink.filesystem.TestStreamingFileSinkFactory;
import org.apache.flink.streaming.api.functions.sink.filesystem.bucketassigners.BasePathBucketAssigner;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
Expand All @@ -50,7 +50,7 @@
import static org.assertj.core.api.Assertions.assertThat;

/** Base class for testing writing data to the hadoop file system with different configurations. */
public class HadoopPathBasedPartFileWriterITCase extends AbstractTestBase {
public class HadoopPathBasedPartFileWriterITCase extends AbstractTestBaseJUnit4 {
@Rule public final Timeout timeoutPerTest = Timeout.seconds(2000);

@Test
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@
import org.apache.flink.streaming.api.functions.source.ContinuousFileReaderOperatorFactory;
import org.apache.flink.streaming.api.functions.source.FileProcessingMode;
import org.apache.flink.streaming.api.functions.source.TimestampedFileInputSplit;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.util.ExceptionUtils;

import org.apache.hadoop.fs.FSDataOutputStream;
Expand Down Expand Up @@ -62,7 +62,7 @@
* IT cases for the {@link ContinuousFileMonitoringFunction} and {@link
* ContinuousFileReaderOperator}.
*/
public class ContinuousFileProcessingITCase extends AbstractTestBase {
public class ContinuousFileProcessingITCase extends AbstractTestBaseJUnit4 {

private static final int NO_OF_FILES = 5;
private static final int LINES_PER_FILE = 100;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@
import org.apache.flink.streaming.api.functions.AssignerWithPunctuatedWatermarks;
import org.apache.flink.streaming.api.watermark.Watermark;
import org.apache.flink.streaming.api.windowing.time.Time;
import org.apache.flink.test.util.AbstractTestBase;
import org.apache.flink.test.util.AbstractTestBaseJUnit4;
import org.apache.flink.types.Either;
import org.apache.flink.util.CloseableIterator;
import org.apache.flink.util.Collector;
Expand All @@ -67,7 +67,7 @@
/** End to end tests of both CEP operators and {@link NFA}. */
@SuppressWarnings("serial")
@RunWith(Parameterized.class)
public class CEPITCase extends AbstractTestBase {
public class CEPITCase extends AbstractTestBaseJUnit4 {

@Parameterized.Parameter public Configuration envConfiguration;

Expand Down
Loading

0 comments on commit ffa639a

Please sign in to comment.