Domanda

I want to use jni to call my c++ lib in spark. When i sbt run my program, it shows that java.lang.UnsatisfiedLinkError: no hq_Image_Process in java.library.path , so obviously the program can not find my hq_Image_Process.so .

In hadoop, -files can distribute the xxx.so file to the slaves like this:

[hadoop@Master ~]$ hadoop jar JniTest3.jar -files /home/hadoop/Documents/java/jni1/bin/libFakeSegmentForJni.so FakeSegmentForJni.TestFakeSegmentForJni input output

Are there any ways to call my hq_Image_Process.so like hadoop in spark? I would appreciate any help.

È stato utile?

Soluzione

First of all, the native library must be preinstalled on all worker nodes. Path to that library must be specified in spark-env.sh:

export SPARK_LIBRARY_PATH=/path/to/native/library

SPARK_PRINT_LAUNCH_COMMAND environment variable might be used to diagnose it:

export SPARK_PRINT_LAUNCH_COMMAND=1

If everything's set correctly, you will see output like this:

Spark Command:
/path/to/java -cp <long list of jars> -Djava.library.path=/path/to/native/library <etc>
========================================

Altri suggerimenti

Solution of accepted answer was for older (<1.0) Spark versions.

You need to set the following properties (either or both) on your spark-defaults.conf.

spark.driver.extraLibraryPath   /path/to/native/library
spark.executor.extraLibraryPath /path/to/native/library

The property keys are documented at Configuration Section of Spark docs.

Autorizzato sotto: CC-BY-SA insieme a attribuzione
Non affiliato a StackOverflow
scroll top