diff --git a/Hadoop-Benchmark.md b/Hadoop-Benchmark.md index 37e92a7..7c286f5 100644 --- a/Hadoop-Benchmark.md +++ b/Hadoop-Benchmark.md @@ -26,7 +26,7 @@ Then get the seaweedfs hadoop client jar. ``` cd share/hadoop/common/lib/ -wget https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.6.1/seaweedfs-hadoop2-client-1.6.1.jar +wget https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.6.2/seaweedfs-hadoop2-client-1.6.2.jar ``` # TestDFSIO Benchmark diff --git a/Hadoop-Compatible-File-System.md b/Hadoop-Compatible-File-System.md index 7b10836..b6e9a70 100644 --- a/Hadoop-Compatible-File-System.md +++ b/Hadoop-Compatible-File-System.md @@ -10,12 +10,12 @@ $ mvn install # build for hadoop2 $cd $GOPATH/src/github.com/chrislusf/seaweedfs/other/java/hdfs2 $ mvn package -$ ls -al target/seaweedfs-hadoop2-client-1.6.1.jar +$ ls -al target/seaweedfs-hadoop2-client-1.6.2.jar # build for hadoop3 $cd $GOPATH/src/github.com/chrislusf/seaweedfs/other/java/hdfs3 $ mvn package -$ ls -al target/seaweedfs-hadoop3-client-1.6.1.jar +$ ls -al target/seaweedfs-hadoop3-client-1.6.2.jar ``` Maven @@ -23,7 +23,7 @@ Maven com.github.chrislusf seaweedfs-hadoop3-client - 1.6.1 + 1.6.2 or @@ -31,23 +31,23 @@ or com.github.chrislusf seaweedfs-hadoop2-client - 1.6.1 + 1.6.2 ``` Or you can download the latest version from MavenCentral * https://mvnrepository.com/artifact/com.github.chrislusf/seaweedfs-hadoop2-client - * [seaweedfs-hadoop2-client-1.6.1.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.6.1/seaweedfs-hadoop2-client-1.6.1.jar) + * [seaweedfs-hadoop2-client-1.6.2.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.6.2/seaweedfs-hadoop2-client-1.6.2.jar) * https://mvnrepository.com/artifact/com.github.chrislusf/seaweedfs-hadoop3-client - * [seaweedfs-hadoop3-client-1.6.1.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop3-client/1.6.1/seaweedfs-hadoop3-client-1.6.1.jar) + * [seaweedfs-hadoop3-client-1.6.2.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop3-client/1.6.2/seaweedfs-hadoop3-client-1.6.2.jar) # Test SeaweedFS on Hadoop Suppose you are getting a new Hadoop installation. Here are the minimum steps to get SeaweedFS to run. -You would need to start a weed filer first, build the seaweedfs-hadoop2-client-1.6.1.jar -or seaweedfs-hadoop3-client-1.6.1.jar, and do the following: +You would need to start a weed filer first, build the seaweedfs-hadoop2-client-1.6.2.jar +or seaweedfs-hadoop3-client-1.6.2.jar, and do the following: ``` # optionally adjust hadoop memory allocation @@ -60,12 +60,12 @@ $ echo "" > etc/hadoop/mapred-site.xml # on hadoop2 $ bin/hdfs dfs -Dfs.defaultFS=seaweedfs://localhost:8888 \ -Dfs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem \ - -libjars ./seaweedfs-hadoop2-client-1.6.1.jar \ + -libjars ./seaweedfs-hadoop2-client-1.6.2.jar \ -ls / # or on hadoop3 $ bin/hdfs dfs -Dfs.defaultFS=seaweedfs://localhost:8888 \ -Dfs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem \ - -libjars ./seaweedfs-hadoop3-client-1.6.1.jar \ + -libjars ./seaweedfs-hadoop3-client-1.6.2.jar \ -ls / ``` @@ -112,9 +112,9 @@ $ bin/hadoop classpath # Copy SeaweedFS HDFS client jar to one of the folders $ cd ${HADOOP_HOME} # for hadoop2 -$ cp ./seaweedfs-hadoop2-client-1.6.1.jar share/hadoop/common/lib/ +$ cp ./seaweedfs-hadoop2-client-1.6.2.jar share/hadoop/common/lib/ # or for hadoop3 -$ cp ./seaweedfs-hadoop3-client-1.6.1.jar share/hadoop/common/lib/ +$ cp ./seaweedfs-hadoop3-client-1.6.2.jar share/hadoop/common/lib/ ``` Now you can do this: diff --git a/Run-Presto-on-SeaweedFS.md b/Run-Presto-on-SeaweedFS.md index 57289e5..db8cd3a 100644 --- a/Run-Presto-on-SeaweedFS.md +++ b/Run-Presto-on-SeaweedFS.md @@ -5,10 +5,10 @@ The installation steps are divided into 2 steps: * https://cwiki.apache.org/confluence/display/Hive/AdminManual+Metastore+Administration ### Configure Hive Metastore to support SeaweedFS -1. Copy the seaweedfs-hadoop2-client-1.6.1.jar to hive lib directory,for example: +1. Copy the seaweedfs-hadoop2-client-1.6.2.jar to hive lib directory,for example: ``` -cp seaweedfs-hadoop2-client-1.6.1.jar /opt/hadoop/share/hadoop/common/lib/ -cp seaweedfs-hadoop2-client-1.6.1.jar /opt/hive-metastore/lib/ +cp seaweedfs-hadoop2-client-1.6.2.jar /opt/hadoop/share/hadoop/common/lib/ +cp seaweedfs-hadoop2-client-1.6.2.jar /opt/hive-metastore/lib/ ``` 2. Modify core-site.xml modify core-site.xml to support SeaweedFS, 30888 is the filer port @@ -50,9 +50,9 @@ metastore.thrift.port is the access port exposed by the Hive Metadata service it Follow instructions for installation of Presto: * https://prestosql.io/docs/current/installation/deployment.html ### Configure Presto to support SeaweedFS -1. Copy the seaweedfs-hadoop2-client-1.6.1.jar to Presto directory,for example: +1. Copy the seaweedfs-hadoop2-client-1.6.2.jar to Presto directory,for example: ``` -cp seaweedfs-hadoop2-client-1.6.1.jar /opt/presto-server-347/plugin/hive-hadoop2/ +cp seaweedfs-hadoop2-client-1.6.2.jar /opt/presto-server-347/plugin/hive-hadoop2/ ``` 2. Modify core-site.xml diff --git a/run-HBase-on-SeaweedFS.md b/run-HBase-on-SeaweedFS.md index d458ddd..4ff02e0 100644 --- a/run-HBase-on-SeaweedFS.md +++ b/run-HBase-on-SeaweedFS.md @@ -1,7 +1,7 @@ # Installation for HBase Two steps to run HBase on SeaweedFS -1. Copy the seaweedfs-hadoop2-client-1.6.1.jar to `${HBASE_HOME}/lib` +1. Copy the seaweedfs-hadoop2-client-1.6.2.jar to `${HBASE_HOME}/lib` 1. And add the following 2 properties in `${HBASE_HOME}/conf/hbase-site.xml` ``` diff --git a/run-Spark-on-SeaweedFS.md b/run-Spark-on-SeaweedFS.md index 26d69ff..0a84082 100644 --- a/run-Spark-on-SeaweedFS.md +++ b/run-Spark-on-SeaweedFS.md @@ -11,12 +11,12 @@ To make these files visible to Spark, set HADOOP_CONF_DIR in $SPARK_HOME/conf/sp ## installation not inheriting from Hadoop cluster configuration -Copy the seaweedfs-hadoop2-client-1.6.1.jar to all executor machines. +Copy the seaweedfs-hadoop2-client-1.6.2.jar to all executor machines. Add the following to spark/conf/spark-defaults.conf on every node running Spark ``` -spark.driver.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.6.1.jar -spark.executor.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.6.1.jar +spark.driver.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.6.2.jar +spark.executor.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.6.2.jar ``` And modify the configuration at runtime: @@ -37,8 +37,8 @@ And modify the configuration at runtime: 1. change the spark-defaults.conf ``` -spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.1.jar -spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.1.jar +spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.2.jar +spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.2.jar spark.hadoop.fs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem ``` @@ -81,8 +81,8 @@ spark.history.fs.cleaner.enabled=true spark.history.fs.logDirectory=seaweedfs://localhost:8888/spark2-history/ spark.eventLog.dir=seaweedfs://localhost:8888/spark2-history/ -spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.1.jar -spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.1.jar +spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.2.jar +spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.6.2.jar spark.hadoop.fs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem spark.hadoop.fs.defaultFS=seaweedfs://localhost:8888 ```