Skip to content

Commit

Permalink
oops editor
Browse files Browse the repository at this point in the history
  • Loading branch information
felixcheung committed May 1, 2017
1 parent fc4965a commit 2c5e267
Show file tree
Hide file tree
Showing 14 changed files with 14 additions and 14 deletions.
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_Serde.R
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ sparkSession <- sparkR.session(enableHiveSupport = FALSE)

test_that("SerDe of primitive types", {
skip_on_cran()

x <- callJStatic("SparkRHandler", "echo", 1L)
expect_equal(x, 1L)
expect_equal(class(x), "integer")
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_Windows.R
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ context("Windows-specific tests")

test_that("sparkJars tag in SparkContext", {
skip_on_cran()

if (.Platform$OS.type != "windows") {
skip("This test is only for Windows, skipped")
}
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_binaryFile.R
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@ test_that("saveAsObjectFile()/objectFile() following RDD transformations works",

test_that("saveAsObjectFile()/objectFile() works with multiple paths", {
skip_on_cran()

fileName1 <- tempfile(pattern = "spark-test", fileext = ".tmp")
fileName2 <- tempfile(pattern = "spark-test", fileext = ".tmp")

Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_broadcast.R
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ test_that("using broadcast variable", {

test_that("without using broadcast variable", {
skip_on_cran()

randomMat <- matrix(nrow = 10, ncol = 10, data = rnorm(100))

useBroadcast <- function(x) {
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_client.R
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@ test_that("multiple packages don't produce a warning", {

test_that("sparkJars sparkPackages as character vectors", {
skip_on_cran()

args <- generateSparkSubmitArgs("", "", c("one.jar", "two.jar", "three.jar"), "",
c("com.databricks:spark-avro_2.10:2.0.1"))
expect_match(args, "--jars one.jar,two.jar,three.jar")
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_context.R
Original file line number Diff line number Diff line change
Expand Up @@ -183,7 +183,7 @@ test_that("spark.lapply should perform simple transforms", {

test_that("add and get file to be downloaded with Spark job on every node", {
skip_on_cran()

sparkR.sparkContext()
# Test add file.
path <- tempfile(pattern = "hello", fileext = ".txt")
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_includePackage.R
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ test_that("include inside function", {

test_that("use include package", {
skip_on_cran()

# Only run the test if plyr is installed.
if ("plyr" %in% rownames(installed.packages())) {
suppressPackageStartupMessages(library(plyr))
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_mllib_clustering.R
Original file line number Diff line number Diff line change
Expand Up @@ -300,7 +300,7 @@ test_that("spark.lda with text input", {

test_that("spark.posterior and spark.perplexity", {
skip_on_cran()

text <- read.text(absoluteSparkPath("data/mllib/sample_lda_data.txt"))
model <- spark.lda(text, features = "value", k = 3)

Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_mllib_regression.R
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ sparkSession <- sparkR.session(enableHiveSupport = FALSE)

test_that("formula of spark.glm", {
skip_on_cran()

training <- suppressWarnings(createDataFrame(iris))
# directly calling the spark API
# dot minus and intercept vs native glm
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_shuffle.R
Original file line number Diff line number Diff line change
Expand Up @@ -228,7 +228,7 @@ test_that("partitionBy works with dependencies", {

test_that("test partitionBy with string keys", {
skip_on_cran()

words <- flatMap(strListRDD, function(line) { strsplit(line, " ")[[1]] })
wordCount <- lapply(words, function(word) { list(word, 1L) })

Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_sparkR.R
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ context("functions in sparkR.R")

test_that("sparkCheckInstall", {
skip_on_cran()

# "local, yarn-client, mesos-client" mode, SPARK_HOME was set correctly,
# and the SparkR job was submitted by "spark-submit"
sparkHome <- paste0(tempdir(), "/", "sparkHome")
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_sparkSQL.R
Original file line number Diff line number Diff line change
Expand Up @@ -3319,7 +3319,7 @@ compare_list <- function(list1, list2) {
# This should always be the **very last test** in this test file.
test_that("No extra files are created in SPARK_HOME by starting session and making calls", {
skip_on_cran()

# Check that it is not creating any extra file.
# Does not check the tempdir which would be cleaned up after.
filesAfter <- list.files(path = sparkRDir, all.files = TRUE)
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_streaming.R
Original file line number Diff line number Diff line change
Expand Up @@ -135,7 +135,7 @@ test_that("Unsupported operation", {

test_that("Terminated by error", {
skip_on_cran()

df <- read.stream("json", path = jsonDir, schema = schema, maxFilesPerTrigger = -1)
counts <- count(group_by(df, "name"))
# This would not fail before returning with a StreamingQuery,
Expand Down
2 changes: 1 addition & 1 deletion R/pkg/inst/tests/testthat/test_textFile.R
Original file line number Diff line number Diff line change
Expand Up @@ -158,7 +158,7 @@ test_that("textFile() on multiple paths", {

test_that("Pipelined operations on RDDs created using textFile", {
skip_on_cran()

fileName <- tempfile(pattern = "spark-test", fileext = ".tmp")
writeLines(mockFile, fileName)

Expand Down

0 comments on commit 2c5e267

Please sign in to comment.