Stanford CoreNLP Nodes

Building Stanford CoreNLP jars for Sparkflows

Stanford CoreNLP nodes do not come as part of the packaged jar files. The code is available in a github repo and can be checked out and built.

https://github.com/sparkflows/sparkflows-stanfordcorenlp

Installing Stanford CoreNLP model jars

Sparkflows needs version 3.4.1 of Stanford CoreNLP model jars : stanford-corenlp-3.4.1-models.jar

 

eg :     wget http://repo1.maven.org/maven2/edu/stanford/nlp/stanford-corenlp/3.4.1/stanford-corenlp-3.4.1-models.jar

 

  • Copy the model jars into the user-lib directory under sparkflows install directory

  • Restart the fire server.

Passing the model file when executing on the Spark Cluster

When executing on the spark cluster, the model jar has to be passed to the job using the --jars option

eg : 

  • --jars /home/centos/sparkflows-fire-1.4.0/user-lib/stanford-corenlp-3.4.1-models.jar

Passing the Stanford CoreNLP sparkflows wrapper jar file when executing on the Spark Cluster

When executing on the spark cluster, the sparkflows wrapper jar has to be passed using the --jars option

eg : 

  • --jars /home/centos/sparkflows-fire-1.4.0/user-lib/sparkflows-stanfordcorenlp-1.4.0.jar

SUPPORT

For support please email:

SOCIAL

  • facebook
  • linkedin
  • twitter
  • angellist
© 2019 Sparkflows, Inc. All rights reserved. 

Terms and Conditions