Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Drop Spark 2.3 Support #895

Merged
merged 10 commits into from
Apr 19, 2021
Merged
Show file tree
Hide file tree
Changes from 7 commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
52 changes: 0 additions & 52 deletions azure-pipelines.yml
Original file line number Diff line number Diff line change
Expand Up @@ -16,11 +16,6 @@ variables:
backwardCompatibleRelease: '1.0.0'
forwardCompatibleRelease: '1.0.0'

backwardCompatibleTestOptions_Windows_2_3: ""
forwardCompatibleTestOptions_Windows_2_3: ""
backwardCompatibleTestOptions_Linux_2_3: ""
forwardCompatibleTestOptions_Linux_2_3: ""

backwardCompatibleTestOptions_Windows_2_4: ""
forwardCompatibleTestOptions_Windows_2_4: ""
backwardCompatibleTestOptions_Linux_2_4: ""
Expand Down Expand Up @@ -206,53 +201,6 @@ stages:
backwardCompatibleRelease: $(backwardCompatibleRelease)
forwardCompatibleRelease: $(forwardCompatibleRelease)
tests:
- version: '2.3.0'
jobOptions:
# 'Hosted Ubuntu 1604' test is disabled due to https://github.com/dotnet/spark/issues/753
- pool: 'Hosted VS2017'
testOptions: ''
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Windows_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Windows_2_3)
- version: '2.3.1'
jobOptions:
- pool: 'Hosted VS2017'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Windows_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Windows_2_3)
- pool: 'Hosted Ubuntu 1604'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Linux_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Linux_2_3)
- version: '2.3.2'
jobOptions:
- pool: 'Hosted VS2017'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Windows_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Windows_2_3)
- pool: 'Hosted Ubuntu 1604'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Linux_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Linux_2_3)
- version: '2.3.3'
jobOptions:
- pool: 'Hosted VS2017'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Windows_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Windows_2_3)
- pool: 'Hosted Ubuntu 1604'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Linux_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Linux_2_3)
- version: '2.3.4'
jobOptions:
- pool: 'Hosted VS2017'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Windows_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Windows_2_3)
- pool: 'Hosted Ubuntu 1604'
testOptions: ""
backwardCompatibleTestOptions: $(backwardCompatibleTestOptions_Linux_2_3)
forwardCompatibleTestOptions: $(forwardCompatibleTestOptions_Linux_2_3)
- version: '2.4.0'
jobOptions:
- pool: 'Hosted VS2017'
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -145,10 +145,10 @@ private bool IsPathValid(string path)
}

Version version = SparkEnvironment.SparkVersion;
return (version.Major, version.Minor, version.Build) switch
return version.Major switch
{
(2, _, _) => false,
(3, 0, _) => true,
2 => false,
3 => true,
_ => throw new NotSupportedException($"Spark {version} not supported.")
};
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,6 @@
// See the LICENSE file in the project root for more information.

using System.IO;
using Microsoft.Spark.E2ETest.Utils;
using Microsoft.Spark.ML.Feature;
using Microsoft.Spark.Sql;
using Microsoft.Spark.Sql.Types;
Expand All @@ -23,10 +22,10 @@ public StopWordsRemoverTests(SparkFixture fixture) : base(fixture)
}

/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
string expectedUid = "theUidWithOutLocale";
string expectedInputCol = "input_col";
Expand Down Expand Up @@ -62,16 +61,9 @@ public void TestSignaturesV2_3_X()
Assert.IsType<DataFrame>(stopWordsRemover.Transform(input));

TestFeatureBase(stopWordsRemover, "inputCol", "input_col");
}

/// <summary>
/// Test signatures for APIs introduced in Spark 2.4.*.
/// </summary>
[SkipIfSparkVersionIsLessThan(Versions.V2_4_0)]
public void TestSignaturesV2_4_X()
{
string expectedLocale = "en_GB";
StopWordsRemover stopWordsRemover = new StopWordsRemover().SetLocale(expectedLocale);
stopWordsRemover.SetLocale(expectedLocale);
Assert.Equal(expectedLocale, stopWordsRemover.GetLocale());
}
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,13 +17,13 @@ namespace Microsoft.Spark.E2ETest.IpcTests
public class SparkContextTests
{
/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
/// <remarks>
/// For the RDD related tests, refer to <see cref="RDDTests"/>.
/// </remarks>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
SparkContext sc = SparkContext.GetOrCreate(new SparkConf());

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,10 +24,10 @@ public CatalogTests(SparkFixture fixture)
}

/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
WithTable(_spark, new string[] { "users", "users2", "users3", "users4", "usersp" }, () =>
{
Expand Down
30 changes: 15 additions & 15 deletions src/csharp/Microsoft.Spark.E2ETest/IpcTests/Sql/ColumnTests.cs
Original file line number Diff line number Diff line change
Expand Up @@ -14,10 +14,10 @@ namespace Microsoft.Spark.E2ETest.IpcTests
public class ColumnTests
{
/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
Column col1 = Column("col1");
Column col2 = Column("col2");
Expand All @@ -28,22 +28,22 @@ public void TestSignaturesV2_3_X()

Assert.IsType<Column>(col1 == col2);
Assert.IsType<Column>(col1.EqualTo(col2));

Assert.IsType<Column>(col1 != col2);
Assert.IsType<Column>(col1.NotEqual(col2));
Assert.IsType<Column>(col1 > col2);
Assert.IsType<Column>(col1 > "hello");
Assert.IsType<Column>(col1.Gt(col2));

Assert.IsType<Column>(col1 > col2);
Assert.IsType<Column>(col1 > "hello");
Assert.IsType<Column>(col1.Gt(col2));
Assert.IsType<Column>(col1.Gt("hello"));
Assert.IsType<Column>(col1 < col2);
Assert.IsType<Column>(col1 < "hello");
Assert.IsType<Column>(col1.Lt(col2));

Assert.IsType<Column>(col1 < col2);
Assert.IsType<Column>(col1 < "hello");
Assert.IsType<Column>(col1.Lt(col2));
Assert.IsType<Column>(col1.Lt("hello"));

Assert.IsType<Column>(col1 <= col2);
Assert.IsType<Column>(col1 <= "hello");
Assert.IsType<Column>(col1 <= "hello");
Assert.IsType<Column>(col1.Leq(col2));
Assert.IsType<Column>(col1.Leq("hello"));

Expand All @@ -59,7 +59,7 @@ public void TestSignaturesV2_3_X()

Assert.IsType<Column>(When(col1 == col2, 0).Otherwise(col2));
Assert.IsType<Column>(When(col1 == col2, 0).Otherwise("hello"));

Assert.IsType<Column>(col1.Between(col1, col2));
Assert.IsType<Column>(col1.Between(1, 3));

Expand All @@ -69,7 +69,7 @@ public void TestSignaturesV2_3_X()

Assert.IsType<Column>(col1 | col2);
Assert.IsType<Column>(col1.Or(col2));

Assert.IsType<Column>(col1 & col2);
Assert.IsType<Column>(col1.And(col2));

Expand Down Expand Up @@ -139,7 +139,7 @@ public void TestSignaturesV2_3_X()

Assert.IsType<Column>(col1.Over(PartitionBy(col1)));
Assert.IsType<Column>(col1.Over());

Assert.Equal("col1", col1.ToString());
Assert.Equal("col2", col2.ToString());
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,15 +20,15 @@ public DataFrameReaderTests(SparkFixture fixture)
}

/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
DataFrameReader dfr = _spark.Read();

Assert.IsType<DataFrameReader>(dfr.Format("json"));

Assert.IsType<DataFrameReader>(
dfr.Schema(
new StructType(new[]
Expand Down
34 changes: 12 additions & 22 deletions src/csharp/Microsoft.Spark.E2ETest/IpcTests/Sql/DataFrameTests.cs
Original file line number Diff line number Diff line change
Expand Up @@ -437,10 +437,10 @@ private static FxDataFrame CountCharacters(FxDataFrame dataFrame)
}

/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
Assert.IsType<Column>(_df["name"]);
Assert.IsType<Column>(_df["age"]);
Expand Down Expand Up @@ -569,6 +569,16 @@ public void TestSignaturesV2_3_X()

Assert.IsType<DataFrame>(df.Sum("age"));
Assert.IsType<DataFrame>(df.Sum("age", "tempAge"));

var values = new List<object> { 19, "twenty" };

Assert.IsType<RelationalGroupedDataset>(df.Pivot("age"));

Assert.IsType<RelationalGroupedDataset>(df.Pivot(Col("age")));

Assert.IsType<RelationalGroupedDataset>(df.Pivot("age", values));

Assert.IsType<RelationalGroupedDataset>(df.Pivot(Col("age"), values));
}

Assert.IsType<RelationalGroupedDataset>(_df.Rollup("age"));
Expand Down Expand Up @@ -669,32 +679,12 @@ public void TestSignaturesV2_3_X()
_df.CreateOrReplaceGlobalTempView("global_view");

Assert.IsType<string[]>(_df.InputFiles().ToArray());
}

/// <summary>
/// Test signatures for APIs introduced in Spark 2.4.*.
/// </summary>
[SkipIfSparkVersionIsLessThan(Versions.V2_4_0)]
public void TestSignaturesV2_4_X()
{
_df.IsEmpty();

_df.IntersectAll(_df);

_df.ExceptAll(_df);

{
RelationalGroupedDataset df = _df.GroupBy("name");
var values = new List<object> { 19, "twenty" };

Assert.IsType<RelationalGroupedDataset>(df.Pivot("age"));

Assert.IsType<RelationalGroupedDataset>(df.Pivot(Col("age")));

Assert.IsType<RelationalGroupedDataset>(df.Pivot("age", values));

Assert.IsType<RelationalGroupedDataset>(df.Pivot(Col("age"), values));
}
}

/// <summary>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,10 +20,10 @@ public DataFrameWriterTests(SparkFixture fixture)
}

/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
{
DataFrameWriter dfw = _spark
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,10 +15,10 @@ namespace Microsoft.Spark.E2ETest.IpcTests
public class WindowSpecTests
{
/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
Column col1 = Column("age");
Column col2 = Column("name");
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,10 +15,10 @@ namespace Microsoft.Spark.E2ETest.IpcTests
public class WindowTests
{
/// <summary>
/// Test signatures for APIs up to Spark 2.3.*.
/// Test signatures for APIs up to Spark 2.4.*.
/// </summary>
[Fact]
public void TestSignaturesV2_3_X()
public void TestSignaturesV2_4_X()
{
Column col1 = Column("age");
Column col2 = Column("name");
Expand Down
Loading