zeppelin

Zeppelin下使用anaconda和matplotlib

一世执手 提交于 2019-12-03 13:00:48
以下操作只需在部署Zeppelin的服务器下即可 1、https://www.continuum.io/downloads 下载anaconda 2、安装,我下载的是linux版本的python2.7,安装过程简单,我安装在了/usr/local/anaconda2下 3、替换系统的python版本,我的是python2.6,大家可以在/usr/bin/下看python的版本,我是python2.6,所以 mv /usr/bin/python /usr/bin/python2.6 ln -s /usr/local/anaconda2/bin/python2.7 /usr/bin/python 4、在 /conf/zeppelin-env.sh中添加 export PYSPARK_PYTHON=/home/spark-1.6.0-bin-hadoop2.6/python export PYTHONPATH=/home/spark-1.6.0-bin-hadoop2.6/python:/home/spark-1.6.0-bin-hadoop2.6/python/lib/py4j-0.9-src.zip 5、编辑/etc/profile,添加 export PYTHONPATH=/home/spark-1.6.0-bin-hadoop2.6/python://home/spark-1.6

Apache zeppelin build process failure in zeppelin-web with bower

匿名 (未验证) 提交于 2019-12-03 08:59:04
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I am trying to build zeppelin locally with windows and babun/cygwin. This site got me headed in the right direction, but I run into the following error when the build gets to Web Application: [ERROR] Failed to execute goal com.github.eirslett:frontend-maven-plugin:0.0.23:bower (bower install) on project zeppelin-web: Failed to run task: 'bower --allow-root install' failed. (error code 8) -> [Help 1] I can go into the zeppelin-web directory and run bower install successfully, but I'm not sure where to go from here? If I try to do mvn install

Is it possible to integrate Zeppelin notes with git?

匿名 (未验证) 提交于 2019-12-03 08:54:24
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: Is it possible to integrate Zeppelin notes with git? One can set the repository location but how to set that to a remote git repository. This functionality is however, available on Amazon EMR 回答1: Yes. It's possible I use following way. Create a github repo and push all note book. like git clone https://github.com/rockiey/zeppelin-notebooks.git cd zeppelin-notebooks cp -rf ../zeppelin/notebook/* . git add -A git commit -m "init" git push Delete notebook directory cd zeppelin rm -rf notebook Clone github repo to notebook. cd zeppelin git

Apache Zeppelin installation grunt build error

匿名 (未验证) 提交于 2019-12-03 08:46:08
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: My configuration is as follows: Ubuntu 15.04 Java 1.7 Spark 1.4.1 Hadoop 2.7 Maven 3.3.3 I am trying to install Apache Zeppelin after successfully cloning it from github and using the following command mvn clean package -DskipTests Despite several attempts, I am getting the following error after some initial success: [ERROR] Failed to execute goal com.github.eirslett:frontend-maven-plugin:0.0.23:grunt (grunt build) on project zeppelin-web: Failed to run task: 'grunt --no-color' failed. (error code 3) -> [Help 1] Any help will be highly

Apache Zeppelin - Highcharts

匿名 (未验证) 提交于 2019-12-03 07:50:05
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I am trying Apache zeppelin . I wanted to have highcharts. So I thought of using %html interpreter. I have done this print("%html <h3> Hello World!! </h3>") It perfectly works. Know I have the code for highcharts <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> <html xmlns="http://www.w3.org/1999/xhtml"> <head> <meta http-equiv="Content-Type" content="text/html; charset=utf-8" /> <title>Untitled Document</title> <script type="text/javascript" src="http://ajax.googleapis

Reading csv files in zeppelin using spark-csv

匿名 (未验证) 提交于 2019-12-03 03:05:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I wanna read csv files in Zeppelin and would like to use databricks' spark-csv package: https://github.com/databricks/spark-csv In the spark-shell, I can use spark-csv with spark-shell --packages com.databricks:spark-csv_2.11:1.2.0 But how do I tell Zeppelin to use that package? Thanks in advance! 回答1: You need to add the Spark Packages repository to Zeppelin before you can use %dep on spark packages. %dep z.reset() z.addRepo("Spark Packages Repo").url("http://dl.bintray.com/spark-packages/maven") z.load("com.databricks:spark-csv_2.10:1.2.0"

Getting NullPointerException when running Spark Code in Zeppelin 0.7.1

匿名 (未验证) 提交于 2019-12-03 02:51:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I have installed Zeppelin 0.7.1 . When I tried to execute the Example spark program(which was available with Zeppelin Tutorial notebook), I am getting the following error java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:391) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:380) at org.apache.zeppelin.spark

Getting error while building the Apache Zeppelin

匿名 (未验证) 提交于 2019-12-03 02:38:01
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I have my hadoop already setup with cloudera. I wanted to install zeppelin to connect with hive and build the UI for my queries. While building the zeppelin command with the following command: sudo mvn clean package -Pspark-1.3 -Dspark.version=1.3.0 -Dhadoop.version=2.6.0-cdh5.4.7 -Phadoop-2.6 -Pyarn -DskipTests I get this error at the web-application module : [ERROR] npm ERR! Linux 3.19.0-71-generic [ERROR] npm ERR! argv "/home/zeppelin/incubator-zeppelin/zeppelin-web/node/node" "/home/zeppelin/incubator-zeppelin/zeppelin-web/node/node

Running zeppelin on spark cluster mode

匿名 (未验证) 提交于 2019-12-03 02:33:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I am using this tutorial spark cluster on yarn mode in docker container to launch zeppelin in spark cluster in yarn mode. However I am stuck at step 4. I can't find conf/zeppelin-env.sh in my docker container to put further configuration. I tried putting these conf folder of zeppelin but yet now successful. Apart from that zeppelin notebook is also not running on localhost:9001. I am very new to distributed system, it would be great if someone can help me start zeppelin on spark cluster in yarn mode. Here is my docker-compose file to enable

Zeppelin java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.rdd.RDDOperationScope$

匿名 (未验证) 提交于 2019-12-03 02:30:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: When I run the code val home = "/Users/adremja/Documents/Kaggle/outbrain" val documents_categories = sc.textFile(home + "/documents_categories.csv") documents_categories take(10) foreach println in spark-shell it works perfectly scala> val home = "/Users/adremja/Documents/Kaggle/outbrain" home: String = /Users/adremja/Documents/Kaggle/outbrain scala> val documents_categories = sc.textFile(home + "/documents_categories.csv") documents_categories: org.apache.spark.rdd.RDD[String] = /Users/adremja/Documents/Kaggle/outbrain/documents_categories