• mongo-spark 安装排故 ./sbt check


    [error]     at com.mongodb.connection.CommandProtocol.execute(CommandProtocol.java:114)
    [error]     at com.mongodb.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:168)
    [error]     at com.mongodb.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:289)
    [error]     at com.mongodb.connection.DefaultServerConnection.command(DefaultServerConnection.java:176)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:216)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:207)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:146)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:139)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:89)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.OperationHelper.withConnectionSource(OperationHelper.java:422)
    [error]     at com.mongodb.operation.OperationHelper.withConnection(OperationHelper.java:413)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:45)
    [error]     at com.mongodb.Mongo.execute(Mongo.java:845)
    [error]     at com.mongodb.Mongo$2.execute(Mongo.java:828)
    [error]     at com.mongodb.MongoDatabaseImpl.drop(MongoDatabaseImpl.java:143)
    [error]     at com.mongodb.spark.MongoDBDefaults.dropDB(MongoDBDefaults.scala:67)
    [error]     at com.mongodb.spark.JavaRequiresMongoDB.tearDown(JavaRequiresMongoDB.java:106)
    [error]     ...
    [info] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromTheSparkConf started
    [error] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromTheSparkConf failed: com.mongodb.MongoCommandException: Command failed with error 13: 'not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }' on server localhost:27017. The full response is { "ok" : 0.0, "errmsg" : "not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }", "code" : 13, "codeName" : "Unauthorized" }, took 0.001 sec
    [error]     at com.mongodb.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:115)
    [error]     at com.mongodb.connection.CommandProtocol.execute(CommandProtocol.java:114)
    [error]     at com.mongodb.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:168)
    [error]     at com.mongodb.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:289)
    [error]     at com.mongodb.connection.DefaultServerConnection.command(DefaultServerConnection.java:176)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:216)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:207)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:146)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:139)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:89)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.OperationHelper.withConnectionSource(OperationHelper.java:422)
    [error]     at com.mongodb.operation.OperationHelper.withConnection(OperationHelper.java:413)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:45)
    [error]     at com.mongodb.Mongo.execute(Mongo.java:845)
    [error]     at com.mongodb.Mongo$2.execute(Mongo.java:828)
    [error]     at com.mongodb.MongoDatabaseImpl.drop(MongoDatabaseImpl.java:143)
    [error]     at com.mongodb.spark.MongoDBDefaults.dropDB(MongoDBDefaults.scala:67)
    [error]     at com.mongodb.spark.JavaRequiresMongoDB.setUp(JavaRequiresMongoDB.java:100)
    [error]     ...
    [error] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromTheSparkConf failed: com.mongodb.MongoCommandException: Command failed with error 13: 'not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }' on server localhost:27017. The full response is { "ok" : 0.0, "errmsg" : "not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }", "code" : 13, "codeName" : "Unauthorized" }, took 0.001 sec
    [error]     at com.mongodb.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:115)
    [error]     at com.mongodb.connection.CommandProtocol.execute(CommandProtocol.java:114)
    [error]     at com.mongodb.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:168)
    [error]     at com.mongodb.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:289)
    [error]     at com.mongodb.connection.DefaultServerConnection.command(DefaultServerConnection.java:176)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:216)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:207)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:146)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:139)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:89)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.OperationHelper.withConnectionSource(OperationHelper.java:422)
    [error]     at com.mongodb.operation.OperationHelper.withConnection(OperationHelper.java:413)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:45)
    [error]     at com.mongodb.Mongo.execute(Mongo.java:845)
    [error]     at com.mongodb.Mongo$2.execute(Mongo.java:828)
    [error]     at com.mongodb.MongoDatabaseImpl.drop(MongoDatabaseImpl.java:143)
    [error]     at com.mongodb.spark.MongoDBDefaults.dropDB(MongoDBDefaults.scala:67)
    [error]     at com.mongodb.spark.JavaRequiresMongoDB.tearDown(JavaRequiresMongoDB.java:106)
    [error]     ...
    [info] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromAJavaMapAndUseDefaults started
    [error] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromAJavaMapAndUseDefaults failed: com.mongodb.MongoCommandException: Command failed with error 13: 'not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }' on server localhost:27017. The full response is { "ok" : 0.0, "errmsg" : "not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }", "code" : 13, "codeName" : "Unauthorized" }, took 0.001 sec
    [error]     at com.mongodb.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:115)
    [error]     at com.mongodb.connection.CommandProtocol.execute(CommandProtocol.java:114)
    [error]     at com.mongodb.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:168)
    [error]     at com.mongodb.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:289)
    [error]     at com.mongodb.connection.DefaultServerConnection.command(DefaultServerConnection.java:176)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:216)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:207)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:146)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:139)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:89)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.OperationHelper.withConnectionSource(OperationHelper.java:422)
    [error]     at com.mongodb.operation.OperationHelper.withConnection(OperationHelper.java:413)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:45)
    [error]     at com.mongodb.Mongo.execute(Mongo.java:845)
    [error]     at com.mongodb.Mongo$2.execute(Mongo.java:828)
    [error]     at com.mongodb.MongoDatabaseImpl.drop(MongoDatabaseImpl.java:143)
    [error]     at com.mongodb.spark.MongoDBDefaults.dropDB(MongoDBDefaults.scala:67)
    [error]     at com.mongodb.spark.JavaRequiresMongoDB.setUp(JavaRequiresMongoDB.java:100)
    [error]     ...
    [error] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromAJavaMapAndUseDefaults failed: com.mongodb.MongoCommandException: Command failed with error 13: 'not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }' on server localhost:27017. The full response is { "ok" : 0.0, "errmsg" : "not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }", "code" : 13, "codeName" : "Unauthorized" }, took 0.001 sec
    [error]     at com.mongodb.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:115)
    [error]     at com.mongodb.connection.CommandProtocol.execute(CommandProtocol.java:114)
    [error]     at com.mongodb.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:168)
    [error]     at com.mongodb.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:289)
    [error]     at com.mongodb.connection.DefaultServerConnection.command(DefaultServerConnection.java:176)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:216)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:207)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:146)
    [error]     at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:139)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:89)
    [error]     at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.OperationHelper.withConnectionSource(OperationHelper.java:422)
    [error]     at com.mongodb.operation.OperationHelper.withConnection(OperationHelper.java:413)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:86)
    [error]     at com.mongodb.operation.DropDatabaseOperation.execute(DropDatabaseOperation.java:45)
    [error]     at com.mongodb.Mongo.execute(Mongo.java:845)
    [error]     at com.mongodb.Mongo$2.execute(Mongo.java:828)
    [error]     at com.mongodb.MongoDatabaseImpl.drop(MongoDatabaseImpl.java:143)
    [error]     at com.mongodb.spark.MongoDBDefaults.dropDB(MongoDBDefaults.scala:67)
    [error]     at com.mongodb.spark.JavaRequiresMongoDB.tearDown(JavaRequiresMongoDB.java:106)
    [error]     ...
    [info] Test run finished: 6 failed, 0 ignored, 3 total, 0.003s
    [info] HelpersSpec:
    [info] Exception encountered when attempting to run a suite with class name: UDF.HelpersSpec *** ABORTED ***
    [info]   com.mongodb.MongoCommandException: Command failed with error 13: 'not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }' on server localhost:27017. The full response is { "ok" : 0.0, "errmsg" : "not authorized on mongo-spark-connector-test to execute command { dropDatabase: 1 }", "code" : 13, "codeName" : "Unauthorized" }
    [info]   at com.mongodb.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:115)
    [info]   at com.mongodb.connection.CommandProtocol.execute(CommandProtocol.java:114)
    [info]   at com.mongodb.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:168)
    [info]   at com.mongodb.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:289)
    [info]   at com.mongodb.connection.DefaultServerConnection.command(DefaultServerConnection.java:176)
    [info]   at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:216)
    [info]   at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:207)
    [info]   at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:146)
    [info]   at com.mongodb.operation.CommandOperationHelper.executeWrappedCommandProtocol(CommandOperationHelper.java:139)
    [info]   at com.mongodb.operation.DropDatabaseOperation$1.call(DropDatabaseOperation.java:89)
    [info]   ...
    17/12/01 09:34:10 INFO AuthConnectionSpec: Ended Test: 'AuthConnectionSpec'
    [info] AuthConnectionSpec:
    [info] MongoRDD
    [info] - should be able to connect to an authenticated db *** FAILED ***
    [info]   org.apache.spark.SparkException: Only one SparkContext may be running in this JVM (see SPARK-2243). To ignore this error, set spark.driver.allowMultipleContexts = true. The currently running SparkContext was created at:
    [info] org.apache.spark.SparkContext.<init>(SparkContext.scala:76)
    [info] com.mongodb.spark.SparkConfOverrideSpec.<init>(SparkConfOverrideSpec.scala:66)
    [info] sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    [info] sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    [info] sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    [info] java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    [info] java.lang.Class.newInstance(Class.java:442)
    [info] org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:646)
    [info] sbt.TestRunner.runTest$1(TestFramework.scala:76)
    [info] sbt.TestRunner.run(TestFramework.scala:85)
    [info] sbt.TestFramework$$anon$2$$anonfun$$init$$1$$anonfun$apply$8.apply(TestFramework.scala:202)
    [info] sbt.TestFramework$$anon$2$$anonfun$$init$$1$$anonfun$apply$8.apply(TestFramework.scala:202)
    [info] sbt.TestFramework$.sbt$TestFramework$$withContextLoader(TestFramework.scala:185)
    [info] sbt.TestFramework$$anon$2$$anonfun$$init$$1.apply(TestFramework.scala:202)
    [info] sbt.TestFramework$$anon$2$$anonfun$$init$$1.apply(TestFramework.scala:202)
    [info] sbt.TestFunction.apply(TestFramework.scala:207)
    [info] sbt.Tests$.sbt$Tests$$processRunnable$1(Tests.scala:239)
    [info] sbt.Tests$$anonfun$makeSerial$1.apply(Tests.scala:245)
    [info] sbt.Tests$$anonfun$makeSerial$1.apply(Tests.scala:245)
    [info] sbt.std.Transform$$anon$3$$anonfun$apply$2.apply(System.scala:44)
    [info]   at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2472)
    [info]   at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2468)
    [info]   at scala.Option.foreach(Option.scala:257)
    [info]   at org.apache.spark.SparkContext$.assertNoOtherContextIsRunning(SparkContext.scala:2468)
    [info]   at org.apache.spark.SparkContext$.markPartiallyConstructed(SparkContext.scala:2557)
    [info]   at org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
    [info]   at com.mongodb.spark.AuthConnectionSpec$$anonfun$1.apply$mcV$sp(AuthConnectionSpec.scala:31)
    [info]   at com.mongodb.spark.AuthConnectionSpec$$anonfun$1.apply(AuthConnectionSpec.scala:30)
    [info]   at com.mongodb.spark.AuthConnectionSpec$$anonfun$1.apply(AuthConnectionSpec.scala:30)
    [info]   at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
    [info]   ...
    [info] Reading scoverage instrumentation [/usr/local/mongo-spark/target/scala-2.11/scoverage-data/scoverage.coverage.xml]
    [info] Reading scoverage measurements...
    [info] Generating scoverage reports...
    [info] Written Cobertura report [/usr/local/mongo-spark/target/scala-2.11/coverage-report/cobertura.xml]
    [info] Written XML coverage report [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/scoverage.xml]
    [info] Written HTML coverage report [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/index.html]
    [info] Coverage reports completed
    [info] All done. Coverage was [24.73%]
    [info] ScalaCheck
    [info] Passed: Total 0, Failed 0, Errors 0, Passed 0
    [info] ScalaTest
    [info] Run completed in 3 seconds, 56 milliseconds.
    [info] Total number of tests run: 23
    [info] Suites: completed 4, aborted 18
    [info] Tests: succeeded 22, failed 1, canceled 0, ignored 0, pending 0
    [info] *** 18 SUITES ABORTED ***
    [info] *** 1 TEST FAILED ***
    [error] Error: Total 79, Failed 39, Errors 18, Passed 22
    [error] Failed tests:
    [error]     com.mongodb.spark.AuthConnectionSpec
    [error]     com.mongodb.spark.sql.MongoDataFrameReaderTest
    [error]     com.mongodb.spark.config.ReadConfigTest
    [error]     com.mongodb.spark.MongoSparkTest
    [error]     com.mongodb.spark.sql.MongoDataFrameWriterTest
    [error]     com.mongodb.spark.config.WriteConfigTest
    [error]     com.mongodb.spark.sql.MongoDataFrameTest
    [error]     com.mongodb.spark.NoSparkConfTest
    [error]     com.mongodb.spark.MongoConnectorTest
    [error]     com.mongodb.spark.sql.fieldTypes.api.java.FieldTypesTest
    [error] Error during tests:
    [error]     com.mongodb.spark.rdd.partitioner.MongoSplitVectorPartitionerSpec
    [error]     com.mongodb.spark.connection.DefaultMongoClientFactorySpec
    [error]     com.mongodb.spark.rdd.partitioner.MongoShardedPartitionerSpec
    [error]     com.mongodb.spark.rdd.partitioner.PartitionerHelperSpec
    [error]     com.mongodb.spark.rdd.partitioner.MongoSamplePartitionerSpec
    [error]     com.mongodb.spark.MongoConnectorSpec
    [error]     com.mongodb.spark.sql.MongoInferSchemaSpec
    [error]     com.mongodb.spark.rdd.partitioner.MongoPaginateBySizePartitionerSpec
    [error]     com.mongodb.spark.sql.MapFunctionsSpec
    [error]     com.mongodb.spark.sql.MongoRelationHelperSpec
    [error]     UDF.HelpersSpec
    [error]     com.mongodb.spark.connection.MongoClientRefCounterSpec
    [error]     com.mongodb.spark.sql.MongoDataFrameSpec
    [error]     com.mongodb.spark.MongoRDDSpec
    [error]     com.mongodb.spark.NoSparkConfSpec
    [error]     com.mongodb.spark.sql.fieldTypes.FieldTypesSpec
    [error]     com.mongodb.spark.connection.MongoClientCacheSpec
    [error]     com.mongodb.spark.rdd.partitioner.MongoPaginateByCountPartitionerSpec
    [error]     com.mongodb.spark.SparkConfOverrideSpec
    [error] (mongo-spark-connector/test:test) sbt.TestsFailedException: Tests unsuccessful
    [error] Total time: 189 s, completed 2017-12-1 9:34:11
    [root@hadoop1 mongo-spark]# ./sbt check
    [info] Loading project definition from /usr/local/mongo-spark/project
    [info] Set current project to mongo-spark-connector (in build file:/usr/local/mongo-spark/)
    [warn] Credentials file /root/.ivy2/.spCredentials does not exist
    [success] Total time: 0 s, completed 2017-12-1 9:34:44
    [info] scalastyle using config /usr/local/mongo-spark/project/scalastyle-config.xml
    [info] Processed 71 file(s)
    [info] Found 0 errors
    [info] Found 0 warnings
    [info] Found 0 infos
    [info] Finished in 7 ms
    [success] created output: /usr/local/mongo-spark/target
    [success] Total time: 2 s, completed 2017-12-1 9:34:46
    [success] Total time: 0 s, completed 2017-12-1 9:34:46
    [warn] Credentials file /root/.ivy2/.spCredentials does not exist
    [info] Updating {file:/usr/local/mongo-spark/}mongo-spark-connector...
    [info] Resolving org.scala-lang#scala-library;2.11.8 ...
    [info] Formatting 33 Scala sources {file:/usr/local/mongo-spark/}mongo-spark-connector(test) ...
    [info] Resolving org.apache#apache;14 ...
    [error] Server access Error: 连接超时 (Connection timed out) url=https://repo1.maven.org/maven2/org/apache/spark/spark-parent_2.11/2.2.0/spark-parent_2.11-2.2.0.jar
    [info] Resolving jline#jline;2.12.1 ...
    [info] Done updating.
    [info] Formatting 71 Scala sources {file:/usr/local/mongo-spark/}mongo-spark-connector(compile) ...
    [info] Compiling 71 Scala sources and 15 Java sources to /usr/local/mongo-spark/target/scala-2.11/classes...
    [info] [info] Cleaning datadir [/usr/local/mongo-spark/target/scala-2.11/scoverage-data]
    [info] [info] Beginning coverage instrumentation
    [info] [info] Instrumentation completed [2907 statements]
    [info] [info] Wrote instrumentation file [/usr/local/mongo-spark/target/scala-2.11/scoverage-data/scoverage.coverage.xml]
    [info] [info] Will write measurement data to [/usr/local/mongo-spark/target/scala-2.11/scoverage-data]
    [info] Compiling 33 Scala sources and 14 Java sources to /usr/local/mongo-spark/target/scala-2.11/test-classes...
    17/12/01 09:37:48 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
    17/12/01 09:37:48 WARN Utils: Your hostname, hadoop1 resolves to a loopback address: 127.0.0.1; using 192.168.2.51 instead (on interface eno1)
    17/12/01 09:37:48 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
    17/12/01 09:37:49 INFO log: Logging initialized @188125ms
    17/12/01 09:37:49 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:37:49 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:37:50 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:37:51 INFO MongoRelation: requiredColumns: _id, age, name, filters: 
    [info] Test run started
    [info] Test com.mongodb.spark.NoSparkConfTest.shouldBeAbleToUseConfigsWithRDDs started
    [info] Test com.mongodb.spark.NoSparkConfTest.shouldBeAbleToUseConfigsWithDataFrames started
    [info] Test run finished: 0 failed, 0 ignored, 2 total, 3.87s
    17/12/01 09:37:52 INFO MongoPaginateByCountPartitionerSpec: Running Test: 'MongoPaginateByCountPartitionerSpec'
    17/12/01 09:37:52 INFO MongoPaginateByCountPartitionerSpec: Running Test: 'MongoPaginateByCountPartitionerSpec'
    17/12/01 09:37:52 INFO MongoPaginateByCountPartitionerSpec: Running Test: 'MongoPaginateByCountPartitionerSpec'
    17/12/01 09:37:52 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoPaginateByCountPartitionerSpec'
    17/12/01 09:37:52 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:37:52 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:53 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00001), LessThan(_id,00031)
    17/12/01 09:37:53 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00001), LessThan(_id,00031)
    17/12/01 09:37:53 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), LessThan(_id,00099)
    17/12/01 09:37:53 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), LessThan(_id,00099)
    17/12/01 09:37:53 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThan(_id,09000)
    17/12/01 09:37:53 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), GreaterThan(_id,09000)
    17/12/01 09:37:53 INFO MongoPaginateByCountPartitionerSpec: Running Test: 'MongoPaginateByCountPartitionerSpec'
    17/12/01 09:37:53 WARN MongoPaginateByCountPartitioner: Inefficient partitioning, creating a partition per document. Decrease the `numberOfPartitions` property.
    17/12/01 09:37:53 INFO MongoPaginateByCountPartitioner: Empty collection (MongoPaginateByCountPartitionerSpec), using a single partition
    17/12/01 09:37:53 INFO MongoPaginateByCountPartitioner: Empty collection (MongoPaginateByCountPartitionerSpec), using a single partition
    17/12/01 09:37:53 INFO MongoPaginateByCountPartitionerSpec: Ended Test: 'MongoPaginateByCountPartitionerSpec'
    [info] MongoPaginateByCountPartitionerSpec:
    [info] MongoPaginateByCountPartitioner
    [info] - should partition the database as expected
    [info] - should partition on an alternative shardkey as expected
    [info] - should use the users pipeline when set in a rdd / dataframe
    [info] - should handle fewer documents than partitions
    [info] - should handle no collection
    [info] WriteConfigSpec:
    [info] WriteConfig
    [info] - should have the expected defaults
    [info] - should be creatable from SparkConfig
    [info] - should round trip options
    [info] - should validate the values
    17/12/01 09:37:54 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:37:54 INFO MongoRelation: requiredColumns: _id, age, name, filters: 
    17/12/01 09:37:54 INFO SparkConfOverrideSpec: Ended Test: 'be able to able to override partial configs with options'
    [info] SparkConfOverrideSpec:
    [info] MongoRDD
    [info] - should be able to override partial configs with Read / Write Configs
    [info] DataFrame Readers and Writers
    [info] - should be able to able to override partial configs with options
    [info] Test run started
    [info] Test com.mongodb.spark.sql.MongoDataFrameWriterTest.shouldBeEasilyCreatedFromADataFrameAndSaveToMongo started
    [info] Test com.mongodb.spark.sql.MongoDataFrameWriterTest.shouldTakeACustomOptions started
    [info] Test com.mongodb.spark.sql.MongoDataFrameWriterTest.shouldBeEasilyCreatedFromMongoSpark started
    [info] Test run finished: 0 failed, 0 ignored, 3 total, 0.658s
    17/12/01 09:37:55 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:55 INFO MongoRelation: requiredColumns: _id, arrayInt, binary, boolean, code, codeWithScope, date, dbPointer, document, double, int32, int64, maxKey, minKey, null, objectId, oldBinary, regex, string, symbol, timestamp, undefined, filters: 
    [info] Test run started
    [info] Test com.mongodb.spark.sql.MongoDataFrameTest.shouldRoundTripAllBsonTypes started
    [info] Test run finished: 0 failed, 0 ignored, 1 total, 0.4s
    17/12/01 09:37:55 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:55 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:37:55 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:37:55 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:55 INFO MongoRelation: requiredColumns: name, age, filters: 
    17/12/01 09:37:56 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: a, filters: 
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: a, filters: 
    17/12/01 09:37:56 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: a, filters: 
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: a, filters: 
    17/12/01 09:37:56 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: _id, a, filters: 
    17/12/01 09:37:56 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:56 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: _id, a, filters: 
    17/12/01 09:37:56 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:56 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:56 INFO MongoRelation: requiredColumns: , filters: IsNotNull(age)
    17/12/01 09:37:57 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:57 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:57 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:57 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:57 INFO MongoRelation: requiredColumns: , filters: IsNotNull(age)
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:57 INFO MongoRelation: requiredColumns: , filters: IsNotNull(age)
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:57 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:58 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:58 INFO MongoRelation: requiredColumns: _id, arrayInt, binary, bool, code, codeWithScope, date, dbPointer, dbl, decimal, document, int32, int64, maxKey, minKey, nullValue, objectId, oldBinary, regex, string, symbol, timestamp, undefined, filters: 
    17/12/01 09:37:58 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:58 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:58 INFO MongoRelation: requiredColumns: nullValue, int32, int64, bool, date, dbl, decimal, string, minKey, maxKey, objectId, code, codeWithScope, regex, symbol, timestamp, undefined, binary, oldBinary, arrayInt, document, dbPointer, filters: 
    17/12/01 09:37:58 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:58 INFO MongoRelation: requiredColumns: name, attributes, filters: IsNotNull(name)
    17/12/01 09:37:58 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:59 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:59 INFO MongoRelation: requiredColumns: _id, count, filters: IsNotNull(_id), IsNotNull(count)
    17/12/01 09:37:59 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:37:59 INFO MongoRelation: requiredColumns: _id, count, filters: IsNotNull(_id), IsNotNull(count)
    17/12/01 09:37:59 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:59 INFO MongoRelation: requiredColumns: _id, name, age, filters: 
    17/12/01 09:37:59 INFO MongoDataFrameSpec: Running Test: 'MongoDataFrameSpec'
    17/12/01 09:37:59 INFO MongoRelation: requiredColumns: _id, name, filters: 
    17/12/01 09:37:59 INFO MongoRelation: requiredColumns: _id, name, age, filters: 
    17/12/01 09:37:59 INFO MongoDataFrameSpec: Ended Test: 'MongoDataFrameSpec'
    [info] MongoDataFrameSpec:
    [info] DataFrameReader
    [info] - should should be easily created from the SQLContext and load from Mongo
    [info] - should handle selecting out of order columns
    [info] - should handle mixed numerics with long precedence
    [info] - should handle mixed numerics with double precedence
    [info] - should handle array fields with null values
    [info] - should handle document fields with null values
    [info] - should be easily created with a provided case class
    [info] - should include any pipelines when inferring the schema
    [info] - should use any pipelines when set via the MongoRDD
    [info] - should throw an exception if pipeline is invalid
    [info] DataFrameWriter
    [info] - should be easily created from a DataFrame and save to Mongo
    [info] - should take custom writeConfig
    [info] - should support INSERT INTO SELECT statements
    [info] - should support INSERT OVERWRITE SELECT statements
    [info] DataFrames
    [info] - should round trip all bson types
    [info] - should be able to cast all types to a string value
    [info] - should be able to round trip schemas containing MapTypes
    [info] - should be able to upsert and replace data in an existing collection
    [info] - should be able to handle optional _id fields when upserting / replacing data in a collection
    [info] - should be able to set only the data in the Dataset to the collection
    17/12/01 09:37:59 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoSamplePartitionerSpec'
    17/12/01 09:38:00 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoSamplePartitionerSpec'
    17/12/01 09:38:00 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoSamplePartitionerSpec'
    17/12/01 09:38:00 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoSamplePartitionerSpec'
    17/12/01 09:38:00 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:00 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:00 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), LessThan(_id,02041)
    17/12/01 09:38:00 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), LessThan(_id,02041)
    17/12/01 09:38:01 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00000)
    17/12/01 09:38:01 INFO MongoSamplePartitioner: Could not find collection (MongoSamplePartitionerSpec), using a single partition
    17/12/01 09:38:01 INFO MongoSamplePartitionerSpec: Ended Test: 'MongoSamplePartitionerSpec'
    [info] MongoSamplePartitionerSpec:
    [info] MongoSamplePartitioner
    [info] - should partition the database as expected
    [info] - should partition on an alternative shardkey as expected
    [info] - should partition with a composite key
    [info] - should use the users pipeline when set in a rdd / dataframe
    [info] - should have a default bounds of min to max key
    [info] - should handle no collection
    [info] - should handle an empty collection
    [info] ReadConfigSpec:
    [info] ReadConfig
    [info] - should have the expected defaults
    [info] - should be creatable from SparkConfig
    [info] - should use the URI for default values
    [info] - should override URI values with named values
    [info] - should round trip options
    [info] - should be able to create a map
    [info] - should create the expected ReadPreference and ReadConcern
    [info] - should validate the values
    17/12/01 09:38:01 INFO MongoClientRefCounterSpec: Ended Test: 'MongoClientRefCounterSpec'
    [info] MongoClientRefCounterSpec:
    [info] MongoClientRefCounter
    [info] - should count references as expected
    [info] - should be able to acquire multiple times
    [info] - should throw an exception for invalid releases of a MongoClient
    17/12/01 09:38:01 INFO MongoDBDefaults: Loading sample Data: ~5MB data into 'MongoSplitVectorPartitionerSpec'
    17/12/01 09:38:01 INFO MongoSplitVectorPartitioner: No splitKeys were calculated by the splitVector command, proceeding with a single partition. If this is undesirable try lowering 'partitionSizeMB' property to produce more partitions.
    17/12/01 09:38:01 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoSplitVectorPartitionerSpec'
    17/12/01 09:38:01 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoSplitVectorPartitionerSpec'
    17/12/01 09:38:01 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:01 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00001), LessThan(_id,00031)
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00001), LessThan(_id,00031)
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), LessThan(_id,00049)
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), LessThan(_id,00049)
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00051)
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00051)
    17/12/01 09:38:02 INFO MongoSplitVectorPartitioner: No splitKeys were calculated by the splitVector command, proceeding with a single partition. If this is undesirable try lowering 'partitionSizeMB' property to produce more partitions.
    17/12/01 09:38:02 INFO MongoSplitVectorPartitioner: Could not find collection (MongoSplitVectorPartitionerSpec), using a single partition
    17/12/01 09:38:02 INFO MongoSplitVectorPartitioner: No splitKeys were calculated by the splitVector command, proceeding with a single partition. If this is undesirable try lowering 'partitionSizeMB' property to produce more partitions.
    17/12/01 09:38:02 INFO MongoSplitVectorPartitionerSpec: Ended Test: 'MongoSplitVectorPartitionerSpec'
    [info] MongoSplitVectorPartitionerSpec:
    [info] MongoSplitVectorPartitioner
    [info] - should partition the database as expected
    [info] - should use the provided pipeline for min and max keys
    [info] - should use the users pipeline when set in a rdd / dataframe
    [info] - should have a default bounds of min to max key
    [info] - should handle no collection
    [info] - should handle an empty collection
    17/12/01 09:38:02 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, binary, dbPointer, javaScript, javaScriptWithScope, maxKey, minKey, regularExpression, symbol, timestamp, undefined, filters: 
    17/12/01 09:38:02 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:02 INFO MongoRelation: requiredColumns: _id, binary, dbPointer, javaScript, javaScriptWithScope, maxKey, minKey, regularExpression, symbol, timestamp, undefined, filters: 
    [info] Test run started
    [info] Test com.mongodb.spark.sql.fieldTypes.api.java.FieldTypesTest.shouldBeAbleToCreateADatasetBasedOnAJavaBeanRepresentingComplexBsonTypes started
    [info] Test com.mongodb.spark.sql.fieldTypes.api.java.FieldTypesTest.shouldAllowTheRoundTrippingOfAJavaBeanRepresentingComplexBsonTypes started
    [info] Test run finished: 0 failed, 0 ignored, 2 total, 0.387s
    17/12/01 09:38:02 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:02 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:02 INFO MongoConnectorSpec: Ended Test: 'MongoConnectorSpec'
    [info] MongoConnectorSpec:
    [info] MongoConnector
    [info] - should create a MongoClient
    [info] - should set the correct localThreshold
    [info] - should Use the cache for MongoClients
    [info] - should create a MongoClient with a custom MongoConnectionFactory
    17/12/01 09:38:02 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoPaginateBySizePartitionerSpec'
    17/12/01 09:38:02 INFO MongoPaginateBySizePartitioner: Inefficient partitioning, creating a single partition. Decrease the `partitionsizemb` property.
    17/12/01 09:38:02 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoPaginateBySizePartitionerSpec'
    17/12/01 09:38:03 INFO MongoPaginateBySizePartitionerSpec: Running Test: 'MongoPaginateBySizePartitionerSpec'
    17/12/01 09:38:03 INFO MongoDBDefaults: Loading sample Data: ~10MB data into 'MongoPaginateBySizePartitionerSpec'
    17/12/01 09:38:03 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:03 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00001), LessThan(_id,00031)
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), GreaterThanOrEqual(_id,00001), LessThan(_id,00031)
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), LessThan(_id,00049)
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), LessThan(_id,00049)
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, filters: IsNotNull(_id), GreaterThan(_id,00050)
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, pk, s, filters: IsNotNull(_id), GreaterThan(_id,00050)
    17/12/01 09:38:03 INFO MongoPaginateBySizePartitioner: Inefficient partitioning, creating a single partition. Decrease the `partitionsizemb` property.
    17/12/01 09:38:03 INFO MongoPaginateBySizePartitioner: Could not find collection (MongoPaginateBySizePartitionerSpec), using a single partition
    17/12/01 09:38:03 INFO MongoPaginateBySizePartitioner: Empty collection (MongoPaginateBySizePartitionerSpec), using a single partition
    17/12/01 09:38:03 INFO MongoPaginateBySizePartitionerSpec: Ended Test: 'MongoPaginateBySizePartitionerSpec'
    [info] MongoPaginateBySizePartitionerSpec:
    [info] MongoPaginateBySizePartitioner
    [info] - should partition the database as expected
    [info] - should partition on an alternative shardkey as expected
    [info] - should use the users pipeline when set in a rdd / dataframe
    [info] - should have a default bounds of min to max key
    [info] - should handle no collection
    [info] - should handle an empty collection
    17/12/01 09:38:03 INFO MapFunctionsSpec: Ended Test: 'MapFunctionsSpec'
    [info] MapFunctionsSpec:
    [info] documentToRow
    [info] - should convert a Document into a Row with the given schema
    [info] - should not prune the schema when given a document with missing values
    [info] - should prune the schema when limited by passed required columns
    [info] - should ignore any extra data in the document that is not included in the schema
    [info] - should handle nested schemas
    [info] - should handle schemas containing maps
    [info] - should throw an exception when passed maps without string keys
    [info] rowToDocument
    [info] - should convert a Row into a Document
    [info] - should handle nested schemas
    [info] - should handle nested schemas within nested arrays
    [info] - should handle mixed numerics based on the schema
    [info] - should throw a MongoTypeConversionException when casting to an invalid DataType
    17/12/01 09:38:03 INFO FieldTypesSpec: Running Test: 'FieldTypesSpec'
    17/12/01 09:38:03 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:03 INFO MongoRelation: requiredColumns: _id, binary, dbPointer, javaScript, javaScriptWithScope, minKey, maxKey, regularExpression, symbol, timestamp, undefined, filters: 
    17/12/01 09:38:04 INFO FieldTypesSpec: Running Test: 'FieldTypesSpec'
    17/12/01 09:38:04 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:04 INFO MongoRelation: requiredColumns: _id, binary, dbPointer, javaScript, javaScriptWithScope, minKey, maxKey, regularExpression, symbol, timestamp, undefined, filters: 
    17/12/01 09:38:04 INFO FieldTypesSpec: Ended Test: 'FieldTypesSpec'
    [info] FieldTypesSpec:
    [info] Fields
    [info] - should allow the round tripping of a case class representing complex bson types
    [info] - should be able to create a dataset based on a case class representing complex bson types
    [info] - should be able to create a Regular Expression from a Pattern
    [info] BsonValueOrderingSpec:
    [info] The BsonValueOrdering trait
    [info] - should order all bson types correctly
    [info] - should compare numbers types correctly
    [info] - should compare numbers and longs correctly
    [info] - should compare string types correctly
    [info] - should compare timestamp types correctly
    [info] - should compare binary types correctly
    [info] - should compare array types correctly
    [info] - should compare regex types correctly
    [info] - should compare document types correctly
    [info] - should have no defined order for undefined types
    17/12/01 09:38:04 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:04 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:04 WARN MongoClientCache: Release without acquire for key: Mongo{options=MongoClientOptions{description='null', applicationName='null', readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, fsync=null, journal=null, readConcern=com.mongodb.ReadConcern@0, codecRegistry=org.bson.codecs.configuration.ProvidersCodecRegistry@8cef5306, commandListeners=[], clusterListeners=[], serverListeners=[], serverMonitorListeners=[], minConnectionsPerHost=0, maxConnectionsPerHost=100, threadsAllowedToBlockForConnectionMultiplier=5, serverSelectionTimeout=30000, maxWaitTime=120000, maxConnectionIdleTime=0, maxConnectionLifeTime=0, connectTimeout=10000, socketTimeout=0, socketKeepAlive=false, sslEnabled=false, sslInvalidHostNamesAllowed=false, alwaysUseMBeans=false, heartbeatFrequency=10000, minHeartbeatFrequency=500, heartbeatConnectTimeout=20000, heartbeatSocketTimeout=20000, localThreshold=15, requiredReplicaSetName='null', dbDecoderFactory=com.mongodb.DefaultDBDecoder$1@1999df34, dbEncoderFactory=com.mongodb.DefaultDBEncoder$1@4f04d552, socketFactory=null, cursorFinalizerEnabled=true, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitQueueSize=500, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, keepAlive=false, receiveBufferSize=0, sendBufferSize=0}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=20000, readTimeoutMS=20000, keepAlive=false, receiveBufferSize=0, sendBufferSize=0}}}
    17/12/01 09:38:05 WARN MongoClientCache: Release without acquire for key: Mongo{options=MongoClientOptions{description='null', applicationName='null', readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, fsync=null, journal=null, readConcern=com.mongodb.ReadConcern@0, codecRegistry=org.bson.codecs.configuration.ProvidersCodecRegistry@8cef5306, commandListeners=[], clusterListeners=[], serverListeners=[], serverMonitorListeners=[], minConnectionsPerHost=0, maxConnectionsPerHost=100, threadsAllowedToBlockForConnectionMultiplier=5, serverSelectionTimeout=30000, maxWaitTime=120000, maxConnectionIdleTime=0, maxConnectionLifeTime=0, connectTimeout=10000, socketTimeout=0, socketKeepAlive=false, sslEnabled=false, sslInvalidHostNamesAllowed=false, alwaysUseMBeans=false, heartbeatFrequency=10000, minHeartbeatFrequency=500, heartbeatConnectTimeout=20000, heartbeatSocketTimeout=20000, localThreshold=15, requiredReplicaSetName='null', dbDecoderFactory=com.mongodb.DefaultDBDecoder$1@1999df34, dbEncoderFactory=com.mongodb.DefaultDBEncoder$1@4f04d552, socketFactory=null, cursorFinalizerEnabled=true, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitQueueSize=500, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, keepAlive=false, receiveBufferSize=0, sendBufferSize=0}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=20000, readTimeoutMS=20000, keepAlive=false, receiveBufferSize=0, sendBufferSize=0}}}
    17/12/01 09:38:05 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:05 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:05 INFO MongoClientCacheSpec: Ended Test: 'MongoClientCacheSpec'
    [info] MongoClientCacheSpec:
    [info] MongoClientCache
    [info] - should create a client and then close the client once released
    [info] - should create a client and then close the client once released and after the timeout
    [info] - should return a different client once released
    [info] - should not throw an exception when trying to release unacquired client
    [info] - should eventually close all released clients on shutdown
    17/12/01 09:38:05 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:05 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:38:05 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:38:05 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:38:05 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:05 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRelation: requiredColumns: age, filters: IsNotNull(age), GreaterThan(age,100)
    [info] Test run started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldBeEasilyCreatedViaMongoSpark started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldIncludeAnyPipelinesWhenInferringTheSchema started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldBeEasilyCreatedViaMongoSparkAndSQLContext started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldThrowAnExceptionIfPipelineIsInvalid started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldBeEasilyCreatedWithMongoSparkAndJavaBean started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldBeEasilyCreatedWithAProvidedRDDAndJavaBean started
    [info] Test com.mongodb.spark.sql.MongoDataFrameReaderTest.shouldBeEasilyCreatedFromTheSQLContext started
    [info] Test run finished: 0 failed, 0 ignored, 7 total, 1.098s
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 WARN TestPartitioner: Could not find collection (MongoRDDSpec), using single partition
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:06 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:06 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:07 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRelation: requiredColumns: counter, filters: 
    17/12/01 09:38:07 ERROR Executor: Exception in task 0.0 in stage 24.0 (TID 24)
    com.mongodb.spark.exceptions.MongoTypeConversionException: Cannot cast STRING into a IntegerType (value: BsonString{value='a'})
        at com.mongodb.spark.sql.MapFunctions$.com$mongodb$spark$sql$MapFunctions$$convertToDataType(MapFunctions.scala:83)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:39)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:37)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
        at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
        at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186)
        at com.mongodb.spark.sql.MapFunctions$.documentToRow(MapFunctions.scala:37)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
        at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
        at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:395)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:234)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:228)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
        at org.apache.spark.scheduler.Task.run(Task.scala:108)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    17/12/01 09:38:07 WARN TaskSetManager: Lost task 0.0 in stage 24.0 (TID 24, localhost, executor driver): com.mongodb.spark.exceptions.MongoTypeConversionException: Cannot cast STRING into a IntegerType (value: BsonString{value='a'})
        at com.mongodb.spark.sql.MapFunctions$.com$mongodb$spark$sql$MapFunctions$$convertToDataType(MapFunctions.scala:83)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:39)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:37)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
        at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
        at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186)
        at com.mongodb.spark.sql.MapFunctions$.documentToRow(MapFunctions.scala:37)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
        at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
        at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:395)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:234)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:228)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
        at org.apache.spark.scheduler.Task.run(Task.scala:108)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    
    17/12/01 09:38:07 ERROR TaskSetManager: Task 0 in stage 24.0 failed 1 times; aborting job
    17/12/01 09:38:07 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRelation: requiredColumns: counter, filters: 
    17/12/01 09:38:07 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRDDSpec: Running Test: 'MongoRDDSpec'
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRDDSpec: Ended Test: 'MongoRDDSpec'
    [info] MongoRDDSpec:
    [info] MongoRDD
    [info] - should be easily created from the SparkContext
    [info] - should be able to handle non existent collections
    [info] - should be able to query via a pipeline
    [info] - should be able to handle different collection types
    [info] - should be able to create a DataFrame by inferring the schema
    [info] - should be able to create a DataFrame when provided a case class
    [info] - should be able to create a DataFrame with a set schema
    [info] - should be able to create a Dataset when provided a case class
    [info] - should not allow Nothing when trying to create a Dataset
    [info] - should throw when creating a Dataset with invalid data
    [info] - should use default values when creating a Dataset with missing data
    [info] - should be easy to use a custom partitioner
    [info] - should be easy to use a custom partitioner that is an object
    17/12/01 09:38:07 INFO MongoRelationHelperSpec: Ended Test: 'MongoRelationHelperSpec'
    [info] MongoRelationHelperSpec:
    [info] createPipeline
    [info] - should create an empty pipeline if no projection or filters
    [info] - should project the required fields
    [info] - should explicitly exclude _id from the projection if not required
    [info] - should handle spark Filters
    [info] - should and multiple spark Filters
    17/12/01 09:38:07 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    17/12/01 09:38:07 INFO MongoClientCache: Creating MongoClient: [localhost:27017]
    [info] Test run started
    [info] Test com.mongodb.spark.MongoConnectorTest.shouldCreateMongoConnectorWithCustomMongoClientFactory started
    [info] Test com.mongodb.spark.MongoConnectorTest.shouldCreateMongoConnector started
    [info] Test com.mongodb.spark.MongoConnectorTest.shouldUseTheMongoClientCache started
    [info] Test com.mongodb.spark.MongoConnectorTest.shouldCreateMongoConnectorFromJavaSparkContext started
    [info] Test run finished: 0 failed, 0 ignored, 4 total, 0.043s
    17/12/01 09:38:07 INFO MongoShardedPartitionerSpec: Ended Test: 'MongoShardedPartitionerSpec'
    [info] MongoShardedPartitionerSpec:
    [info] MongoShardedPartitioner
    [info] - should partition the database as expected !!! CANCELED !!!
    [info]   Not a Sharded MongoDB (MongoShardedPartitionerSpec.scala:30)
    [info] - should have a default bounds of min to max key !!! CANCELED !!!
    [info]   Not a Sharded MongoDB (MongoShardedPartitionerSpec.scala:43)
    [info] - should handle no  collection !!! CANCELED !!!
    [info]   Not a Sharded MongoDB (MongoShardedPartitionerSpec.scala:52)
    [info] - should handle an empty collection !!! CANCELED !!!
    [info]   Not a Sharded MongoDB (MongoShardedPartitionerSpec.scala:58)
    [info] - should calculate the expected hosts for a single node shard
    [info] - should calculate the expected hosts for a multi node shard
    [info] - should return distinct hosts
    [info] - should calculate the expected Partitions
    17/12/01 09:38:07 INFO PartitionerHelperSpec: Ended Test: 'PartitionerHelperSpec'
    [info] PartitionerHelperSpec:
    [info] PartitionerHelper
    [info] - should create the expected partitions query
    [info] - should create the correct partitions
    17/12/01 09:38:07 INFO DefaultMongoClientFactorySpec: Ended Test: 'DefaultMongoClientFactorySpec'
    [info] DefaultMongoClientFactorySpec:
    [info] DefaultMongoClientFactory
    [info] - should create a MongoClient from the connection string
    [info] - should implement equals based on the prefix less options map
    [info] - should set the localThreshold correctly
    [info] - should validate the connection string
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 WARN TestPartitioner: Could not find collection (com.mongodb.spark.MongoSparkTest), using single partition
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRelation: requiredColumns: counter, filters: 
    17/12/01 09:38:07 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:07 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:07 ERROR Executor: Exception in task 0.0 in stage 1.0 (TID 1)
    com.mongodb.spark.exceptions.MongoTypeConversionException: Cannot cast STRING into a IntegerType (value: BsonString{value='a'})
        at com.mongodb.spark.sql.MapFunctions$.com$mongodb$spark$sql$MapFunctions$$convertToDataType(MapFunctions.scala:83)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:39)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:37)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
        at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
        at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186)
        at com.mongodb.spark.sql.MapFunctions$.documentToRow(MapFunctions.scala:37)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
        at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
        at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:395)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:234)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:228)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
        at org.apache.spark.scheduler.Task.run(Task.scala:108)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    17/12/01 09:38:07 WARN TaskSetManager: Lost task 0.0 in stage 1.0 (TID 1, localhost, executor driver): com.mongodb.spark.exceptions.MongoTypeConversionException: Cannot cast STRING into a IntegerType (value: BsonString{value='a'})
        at com.mongodb.spark.sql.MapFunctions$.com$mongodb$spark$sql$MapFunctions$$convertToDataType(MapFunctions.scala:83)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:39)
        at com.mongodb.spark.sql.MapFunctions$$anonfun$3.apply(MapFunctions.scala:37)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
        at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
        at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
        at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186)
        at com.mongodb.spark.sql.MapFunctions$.documentToRow(MapFunctions.scala:37)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at com.mongodb.spark.sql.MongoRelation$$anonfun$buildScan$2.apply(MongoRelation.scala:45)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
        at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIterator.processNext(Unknown Source)
        at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43)
        at org.apache.spark.sql.execution.WholeStageCodegenExec$$anonfun$8$$anon$1.hasNext(WholeStageCodegenExec.scala:395)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:234)
        at org.apache.spark.sql.execution.SparkPlan$$anonfun$2.apply(SparkPlan.scala:228)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$25.apply(RDD.scala:827)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
        at org.apache.spark.scheduler.Task.run(Task.scala:108)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    
    17/12/01 09:38:07 ERROR TaskSetManager: Task 0 in stage 1.0 failed 1 times; aborting job
    17/12/01 09:38:07 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:07 INFO MongoRelation: requiredColumns: counter, filters: 
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 INFO MongoRelation: requiredColumns: counter, filters: 
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:08 INFO MongoRelation: requiredColumns: counter, filters: 
    [info] Test run started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToCreateADataFrameUsingJavaBean started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToHandleNoneExistentCollections started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldThrowWhenCreatingADatasetWithInvalidData started
    [info] Test com.mongodb.spark.MongoSparkTest.useDefaultValuesWhenCreatingADatasetWithMissingData started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToCreateADataFrameByInferringTheSchemaUsingSparkSession started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToCreateADataFrameByInferringTheSchema started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToQueryViaAPipeLine started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToCreateADatasetUsingJavaBeanWithSparkSession started
    [info] Test com.mongodb.spark.MongoSparkTest.useACustomPartitioner started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeCreatableFromTheSparkContextWithAlternativeReadAndWriteConfigs started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeCreatableFromTheSparkContext started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToHandleDifferentCollectionTypes started
    [info] Test com.mongodb.spark.MongoSparkTest.shouldBeAbleToCreateADatasetUsingJavaBean started
    [info] Test run finished: 0 failed, 0 ignored, 13 total, 1.207s
    17/12/01 09:38:08 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:09 INFO MongoRelation: requiredColumns: _id, age, name, filters: 
    17/12/01 09:38:09 INFO NoSparkConfSpec: Ended Test: 'be able to accept just options'
    [info] NoSparkConfSpec:
    [info] MongoRDD
    [info] - should be able to accept just Read / Write Configs
    [info] DataFrame Readers and Writers
    [info] - should be able to accept just options
    [info] Test run started
    [info] Test com.mongodb.spark.config.WriteConfigTest.shouldBeCreatableFromAJavaMap started
    [info] Test com.mongodb.spark.config.WriteConfigTest.shouldBeCreatableFromTheSparkConf started
    [info] Test com.mongodb.spark.config.WriteConfigTest.shouldBeCreatableFromAJavaMapAndUseDefaults started
    [info] Test run finished: 0 failed, 0 ignored, 3 total, 0.002s
    17/12/01 09:38:09 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:09 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:09 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:10 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:10 WARN TaskSetManager: Stage 100 contains a task of very large size (202 KB). The maximum recommended task size is 100 KB.
    17/12/01 09:38:10 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:11 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:11 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    17/12/01 09:38:12 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 INFO MongoInferSchemaSpec: Running Test: 'MongoInferSchemaSpec'
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN MongoInferSchema: Array Field 'a' contains conflicting types converting to StringType
    17/12/01 09:38:12 WARN SparkSession$Builder: Using an existing SparkSession; some configuration may not take effect.
    17/12/01 09:38:12 WARN MongoInferSchema: Array Field 'a' contains conflicting types converting to StringType
    17/12/01 09:38:12 INFO MongoInferSchemaSpec: Ended Test: 'MongoInferSchemaSpec'
    [info] MongoInferSchemaSpec:
    [info] MongoSchemaHelper
    [info] - should be able to infer the schema from simple types
    [info] - should be able to infer the schema from a flat array
    [info] - should be able to infer the schema from a flat document
    [info] - should be able to infer the schema from a nested array
    [info] - should be able to infer the schema from a multi level document
    [info] - should be able to infer the schema with custom sampleSize
    [info] - should ignore empty arrays and null values in arrays
    [info] - should use any set pipelines on the RDD
    [info] - should upscale number types based on numeric precedence
    [info] - should be able to infer the schema from arrays with mixed keys
    [info] - should be able to infer the schema from arrays with mixed numerics
    [info] - should be able to infer the schema from nested arrays with mixed keys
    [info] - should still mark incompatible schemas with a StringType
    [info] Test run started
    [info] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromAJavaMap started
    [info] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromTheSparkConf started
    [info] Test com.mongodb.spark.config.ReadConfigTest.shouldBeCreatableFromAJavaMapAndUseDefaults started
    [info] Test run finished: 0 failed, 0 ignored, 3 total, 0.003s
    17/12/01 09:38:12 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:12 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:12 INFO MongoRelation: requiredColumns: binary, filters: IsNotNull(binary)
    17/12/01 09:38:12 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:12 INFO MongoRelation: requiredColumns: oldBinary, filters: IsNotNull(oldBinary)
    17/12/01 09:38:12 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:12 INFO MongoRelation: requiredColumns: dbPointer, filters: IsNotNull(dbPointer)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: code, filters: IsNotNull(code)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: codeWithScope, filters: IsNotNull(codeWithScope)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: maxKey, filters: IsNotNull(maxKey)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: minKey, filters: IsNotNull(minKey)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: objectId, filters: IsNotNull(objectId)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: regex, filters: IsNotNull(regex)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: regexWithOptions, filters: IsNotNull(regexWithOptions)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: symbol, filters: IsNotNull(symbol)
    17/12/01 09:38:13 INFO HelpersSpec: Running Test: 'HelpersSpec'
    17/12/01 09:38:13 INFO MongoRelation: requiredColumns: timestamp, filters: IsNotNull(timestamp)
    17/12/01 09:38:13 INFO HelpersSpec: Ended Test: 'HelpersSpec'
    [info] HelpersSpec:
    [info] the user defined function helpers
    [info] - should handle Binary values
    [info] - should handle Binary values with a subtype
    [info] - should handle DbPointers
    [info] - should handle JavaScript
    [info] - should handle JavaScript with scope
    [info] - should handle maxKeys
    [info] - should handle minKeys
    [info] - should handle ObjectIds
    [info] - should handle Regular Expressions
    [info] - should handle Regular Expressions with options
    [info] - should handle Symbols
    [info] - should handle Timestamps
    17/12/01 09:38:13 WARN SparkContext: Using an existing SparkContext; some configuration may not take effect.
    17/12/01 09:38:13 WARN TestPartitioner: Could not find collection (beabletoconnecttoanauthenticateddb), using single partition
    17/12/01 09:38:13 INFO AuthConnectionSpec: Ended Test: 'AuthConnectionSpec'
    [info] AuthConnectionSpec:
    [info] MongoRDD
    [info] - should be able to connect to an authenticated db
    [info] Reading scoverage instrumentation [/usr/local/mongo-spark/target/scala-2.11/scoverage-data/scoverage.coverage.xml]
    [info] Reading scoverage measurements...
    [info] Generating scoverage reports...
    [info] Written Cobertura report [/usr/local/mongo-spark/target/scala-2.11/coverage-report/cobertura.xml]
    [info] Written XML coverage report [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/scoverage.xml]
    [info] Written HTML coverage report [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/index.html]
    [info] Coverage reports completed
    [info] All done. Coverage was [80.60%]
    [info] ScalaCheck
    [info] Passed: Total 0, Failed 0, Errors 0, Passed 0
    [info] ScalaTest
    [info] Run completed in 27 seconds, 8 milliseconds.
    [info] Total number of tests run: 151
    [info] Suites: completed 23, aborted 0
    [info] Tests: succeeded 151, failed 0, canceled 4, ignored 0, pending 0
    [info] All tests passed.
    [info] Passed: Total 189, Failed 0, Errors 0, Passed 189, Canceled 4
    [success] Total time: 209 s, completed 2017-12-1 9:38:15
    [info] Aggregating coverage from subprojects...
    [info] Found 1 subproject report files [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/scoverage.xml]
    [info] No subproject data to aggregate, skipping reports
    [success] Total time: 0 s, completed 2017-12-1 9:38:15
    [info] Waiting for measurement data to sync...
    [info] Reading scoverage instrumentation [/usr/local/mongo-spark/target/scala-2.11/scoverage-data/scoverage.coverage.xml]
    [info] Reading scoverage measurements...
    [info] Generating scoverage reports...
    [info] Written Cobertura report [/usr/local/mongo-spark/target/scala-2.11/coverage-report/cobertura.xml]
    [info] Written XML coverage report [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/scoverage.xml]
    [info] Written HTML coverage report [/usr/local/mongo-spark/target/scala-2.11/scoverage-report/index.html]
    17/12/01 09:38:17 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    [info] Coverage reports completed
    [success] Total time: 2 s, completed 2017-12-1 9:38:17
    17/12/01 09:38:17 INFO MongoClientCache: Closing MongoClient: [localhost:27017]
    [root@hadoop1 mongo-spark]# 

    mongodb/mongo-spark: The MongoDB Spark Connector https://github.com/mongodb/mongo-spark

    ./sbt check 报错后

    控制台提示没有权限

    ps -aux | grep mongo;

    杀死需要权限认证的mongodb服务(./mongodb --port 27017 --auth),调整为./mongodb --port 27017 

    之后重新./sbt check

    ok

  • 相关阅读:
    [原译]关于源代码控制的五个误区
    [原译]AVALONDOCK 2.0入门指南第一部分
    [原译]11个高效的VS调试技巧
    【原译】一个可定制的WPF任务对话框
    [原译]YAXLib XML序列化神器
    为什么SIZEOF(STR.SUBSTR(0,3).C_STR())=8?
    C#实现GIF图片反转
    转 基于jquery ajax 用户无刷新登录详解介绍
    php中判断文件空目录是否有读写权限
    防止mysql用户root密码弱口令的攻击方法
  • 原文地址:https://www.cnblogs.com/rsapaper/p/7940460.html
Copyright © 2020-2023  润新知