-
Notifications
You must be signed in to change notification settings - Fork 141
Description
Hi,
I am quite new to stream mining and I am trying to get the streamDM running. However when I finished setting up all the dependencies I got some issues reported below.
Expected behavior
I want to obtain the confusion matrix by using the example commands.
Observed behavior
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
21/03/09 18:40:50 INFO SparkContext: Running Spark version 2.1.0
21/03/09 18:40:50 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
21/03/09 18:40:50 WARN Utils: Your hostname, localhost.localdomain resolves to a loopback address: 127.0.0.1; using 192.168.122.1 instead (on interface virbr0)
21/03/09 18:40:50 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
21/03/09 18:40:50 INFO SecurityManager: Changing view acls to: shunyao
21/03/09 18:40:50 INFO SecurityManager: Changing modify acls to: shunyao
21/03/09 18:40:50 INFO SecurityManager: Changing view acls groups to:
21/03/09 18:40:50 INFO SecurityManager: Changing modify acls groups to:
21/03/09 18:40:50 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(shunyao); groups with view permissions: Set(); users with modify permissions: Set(shunyao); groups with modify permissions: Set()
21/03/09 18:40:50 INFO Utils: Successfully started service 'sparkDriver' on port 41123.
21/03/09 18:40:50 INFO SparkEnv: Registering MapOutputTracker
21/03/09 18:40:50 INFO SparkEnv: Registering BlockManagerMaster
21/03/09 18:40:50 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
21/03/09 18:40:50 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
21/03/09 18:40:50 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-8d18da96-7b2c-48c9-bbd2-b49a29b16274
21/03/09 18:40:50 INFO MemoryStore: MemoryStore started with capacity 413.9 MB
21/03/09 18:40:50 INFO SparkEnv: Registering OutputCommitCoordinator
21/03/09 18:40:51 INFO Utils: Successfully started service 'SparkUI' on port 4040.
21/03/09 18:40:51 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://192.168.122.1:4040
21/03/09 18:40:51 INFO SparkContext: Added JAR file:/home/shunyao/Downloads/streamDM-0.2/scripts/../target/scala-2.10/streamdm-spark-streaming-_2.10-0.2.jar at spark://192.168.122.1:41123/jars/streamdm-spark-streaming-_2.10-0.2.jar with timestamp 1615333251155
21/03/09 18:40:51 INFO Executor: Starting executor ID driver on host localhost
21/03/09 18:40:51 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 45261.
21/03/09 18:40:51 INFO NettyBlockTransferService: Server created on 192.168.122.1:45261
21/03/09 18:40:51 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
21/03/09 18:40:51 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, 192.168.122.1, 45261, None)
21/03/09 18:40:51 INFO BlockManagerMasterEndpoint: Registering block manager 192.168.122.1:45261 with 413.9 MB RAM, BlockManagerId(driver, 192.168.122.1, 45261, None)
21/03/09 18:40:51 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, 192.168.122.1, 45261, None)
21/03/09 18:40:51 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, 192.168.122.1, 45261, None)
Exception in thread "main" java.lang.Exception: Problem creating instance of class: EvaluatePrequential
at com.github.javacliparser.ClassOption.cliStringToObject(ClassOption.java:139)
at org.apache.spark.streamdm.streamDMJob$.main(streamDMJob.scala:43)
at org.apache.spark.streamdm.streamDMJob.main(streamDMJob.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.IllegalArgumentException: Problems with option: streamReader
at com.github.javacliparser.ClassOption.setValueViaCLIString(ClassOption.java:60)
at com.github.javacliparser.AbstractOption.resetToDefault(AbstractOption.java:90)
at com.github.javacliparser.AbstractClassOption.(AbstractClassOption.java:84)
at com.github.javacliparser.AbstractClassOption.(AbstractClassOption.java:63)
at com.github.javacliparser.ClassOption.(ClassOption.java:34)
at org.apache.spark.streamdm.tasks.EvaluatePrequential.(EvaluatePrequential.scala:47)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at java.lang.Class.newInstance(Class.java:442)
at com.github.javacliparser.ClassOption.cliStringToObject(ClassOption.java:137)
... 11 more
Caused by: java.lang.Exception: Class not found: FileReader
at com.github.javacliparser.ClassOption.cliStringToObject(ClassOption.java:132)
at com.github.javacliparser.ClassOption.setValueViaCLIString(ClassOption.java:57)
... 22 more
21/03/09 18:40:51 INFO SparkContext: Invoking stop() from shutdown hook
21/03/09 18:40:51 INFO SparkUI: Stopped Spark web UI at http://192.168.122.1:4040
21/03/09 18:40:51 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
21/03/09 18:40:51 INFO MemoryStore: MemoryStore cleared
21/03/09 18:40:51 INFO BlockManager: BlockManager stopped
21/03/09 18:40:51 INFO BlockManagerMaster: BlockManagerMaster stopped
21/03/09 18:40:51 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
21/03/09 18:40:51 INFO SparkContext: Successfully stopped SparkContext
21/03/09 18:40:51 INFO ShutdownHookManager: Shutdown hook called
21/03/09 18:40:51 INFO ShutdownHookManager: Deleting directory /tmp/spark-6f0a2908-b3f5-477c-bccc-6302be0e18f8
Steps to reproduce the issue
I tried different OS: macOS, Kali, Ubuntu and CentOS 7.
These steps are what I did and followed:
Step 1: Download & install the Java openjdk8.
Step 2: Download & install the Spark (I tried 2.3.2 and 2.1.0 with Hadoop or without Hadoop)
Step 3: Download & install Scala (from https://www.scala-lang.org/files/archive/scala-2.11.8.deb)
Step 4: Download & install SBT (from https://sbt.bintray.com/native-packages/sbt/0.13.15/sbt-0.13.15.tgz)
Step 5: Download StreamDM from Github.
Step 6: Change the SPARK_HOME to the Spark path.
Step 7: In the streamDM directory, run sbt package
Step 8: In the /scripts, run ./spark.sh or ./spark.sh "EvaluatePrequential -l (SGDLearner -l 0.01 -o LogisticLoss -r ZeroRegularizer) –s (FileReader –k 100 –d 60 –f ../data/mydata)"
Result: Errors occurred.
Command line
sbt package
./spark.sh
./spark.sh "EvaluatePrequential -l (SGDLearner -l 0.01 -o LogisticLoss -r ZeroRegularizer) –s (FileReader –k 100 –d 60 –f ../data/mydata)"
Infrastructure details (Contains the versions that I tried)
- Java Version: 8 | 11
- Scala Version: 2.10.2 | 2.11.8 | 2.11.11
- Spark version: 2.1.0 | 2.3.2
- OS version: macOS, Ubuntu 20.04, Kali, CentOS 7
- Cluster mode or local mode? Local
For the Spark, do I need to download with Hadoop or without Hadoop or the file name only contains spark-x.x.x.tgz?
Such as:
- spark-2.3.2-bin-hadoop2.7.tgz
- spark-2.3.2-bin-without-hadoop.tgz
- spark-2.3.2.tgz
Is anything wrong with my setup? I produced the same error on different OS.
Please help, thanks in advance :)