From b8143ec6c730f3962ba022cb52cbdf36dc8789f9 Mon Sep 17 00:00:00 2001 From: Chris Lu Date: Thu, 12 Nov 2020 23:33:28 -0800 Subject: [PATCH] 1.5.4 --- Hadoop-Benchmark.md | 2 +- Hadoop-Compatible-File-System.md | 24 ++++++++++++------------ run-HBase-on-SeaweedFS.md | 2 +- run-Spark-on-SeaweedFS.md | 14 +++++++------- 4 files changed, 21 insertions(+), 21 deletions(-) diff --git a/Hadoop-Benchmark.md b/Hadoop-Benchmark.md index 8159f75..c9e6b00 100644 --- a/Hadoop-Benchmark.md +++ b/Hadoop-Benchmark.md @@ -26,7 +26,7 @@ Then get the seaweedfs hadoop client jar. ``` cd share/hadoop/common/lib/ -wget https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.5.3/seaweedfs-hadoop2-client-1.5.3.jar +wget https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.5.4/seaweedfs-hadoop2-client-1.5.4.jar ``` # TestDFSIO Benchmark diff --git a/Hadoop-Compatible-File-System.md b/Hadoop-Compatible-File-System.md index f2364d3..66d089a 100644 --- a/Hadoop-Compatible-File-System.md +++ b/Hadoop-Compatible-File-System.md @@ -10,12 +10,12 @@ $ mvn install # build for hadoop2 $cd $GOPATH/src/github.com/chrislusf/seaweedfs/other/java/hdfs2 $ mvn package -$ ls -al target/seaweedfs-hadoop2-client-1.5.3.jar +$ ls -al target/seaweedfs-hadoop2-client-1.5.4.jar # build for hadoop3 $cd $GOPATH/src/github.com/chrislusf/seaweedfs/other/java/hdfs3 $ mvn package -$ ls -al target/seaweedfs-hadoop3-client-1.5.3.jar +$ ls -al target/seaweedfs-hadoop3-client-1.5.4.jar ``` Maven @@ -23,7 +23,7 @@ Maven com.github.chrislusf seaweedfs-hadoop3-client - 1.5.3 + 1.5.4 or @@ -31,23 +31,23 @@ or com.github.chrislusf seaweedfs-hadoop2-client - 1.5.3 + 1.5.4 ``` Or you can download the latest version from MavenCentral * https://mvnrepository.com/artifact/com.github.chrislusf/seaweedfs-hadoop2-client - * [seaweedfs-hadoop2-client-1.5.3.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.5.3/seaweedfs-hadoop2-client-1.5.3.jar) + * [seaweedfs-hadoop2-client-1.5.4.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop2-client/1.5.4/seaweedfs-hadoop2-client-1.5.4.jar) * https://mvnrepository.com/artifact/com.github.chrislusf/seaweedfs-hadoop3-client - * [seaweedfs-hadoop3-client-1.5.3.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop3-client/1.5.3/seaweedfs-hadoop3-client-1.5.3.jar) + * [seaweedfs-hadoop3-client-1.5.4.jar](https://oss.sonatype.org/service/local/repositories/releases/content/com/github/chrislusf/seaweedfs-hadoop3-client/1.5.4/seaweedfs-hadoop3-client-1.5.4.jar) # Test SeaweedFS on Hadoop Suppose you are getting a new Hadoop installation. Here are the minimum steps to get SeaweedFS to run. -You would need to start a weed filer first, build the seaweedfs-hadoop2-client-1.5.3.jar -or seaweedfs-hadoop3-client-1.5.3.jar, and do the following: +You would need to start a weed filer first, build the seaweedfs-hadoop2-client-1.5.4.jar +or seaweedfs-hadoop3-client-1.5.4.jar, and do the following: ``` # optionally adjust hadoop memory allocation @@ -60,12 +60,12 @@ $ echo "" > etc/hadoop/mapred-site.xml # on hadoop2 $ bin/hdfs dfs -Dfs.defaultFS=seaweedfs://localhost:8888 \ -Dfs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem \ - -libjars ./seaweedfs-hadoop2-client-1.5.3.jar \ + -libjars ./seaweedfs-hadoop2-client-1.5.4.jar \ -ls / # or on hadoop3 $ bin/hdfs dfs -Dfs.defaultFS=seaweedfs://localhost:8888 \ -Dfs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem \ - -libjars ./seaweedfs-hadoop3-client-1.5.3.jar \ + -libjars ./seaweedfs-hadoop3-client-1.5.4.jar \ -ls / ``` @@ -106,9 +106,9 @@ $ bin/hadoop classpath # Copy SeaweedFS HDFS client jar to one of the folders $ cd ${HADOOP_HOME} # for hadoop2 -$ cp ./seaweedfs-hadoop2-client-1.5.3.jar share/hadoop/common/lib/ +$ cp ./seaweedfs-hadoop2-client-1.5.4.jar share/hadoop/common/lib/ # or for hadoop3 -$ cp ./seaweedfs-hadoop3-client-1.5.3.jar share/hadoop/common/lib/ +$ cp ./seaweedfs-hadoop3-client-1.5.4.jar share/hadoop/common/lib/ ``` Now you can do this: diff --git a/run-HBase-on-SeaweedFS.md b/run-HBase-on-SeaweedFS.md index beebeb3..10220af 100644 --- a/run-HBase-on-SeaweedFS.md +++ b/run-HBase-on-SeaweedFS.md @@ -1,7 +1,7 @@ # Installation for HBase Two steps to run HBase on SeaweedFS -1. Copy the seaweedfs-hadoop2-client-1.5.3.jar to `${HBASE_HOME}/lib` +1. Copy the seaweedfs-hadoop2-client-1.5.4.jar to `${HBASE_HOME}/lib` 1. And add the following 2 properties in `${HBASE_HOME}/conf/hbase-site.xml` ``` diff --git a/run-Spark-on-SeaweedFS.md b/run-Spark-on-SeaweedFS.md index 5467a43..527a4e4 100644 --- a/run-Spark-on-SeaweedFS.md +++ b/run-Spark-on-SeaweedFS.md @@ -11,12 +11,12 @@ To make these files visible to Spark, set HADOOP_CONF_DIR in $SPARK_HOME/conf/sp ## installation not inheriting from Hadoop cluster configuration -Copy the seaweedfs-hadoop2-client-1.5.3.jar to all executor machines. +Copy the seaweedfs-hadoop2-client-1.5.4.jar to all executor machines. Add the following to spark/conf/spark-defaults.conf on every node running Spark ``` -spark.driver.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.5.3.jar -spark.executor.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.5.3.jar +spark.driver.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.5.4.jar +spark.executor.extraClassPath=/path/to/seaweedfs-hadoop2-client-1.5.4.jar ``` And modify the configuration at runtime: @@ -37,8 +37,8 @@ And modify the configuration at runtime: 1. change the spark-defaults.conf ``` -spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.3.jar -spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.3.jar +spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.4.jar +spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.4.jar spark.hadoop.fs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem ``` @@ -81,8 +81,8 @@ spark.history.fs.cleaner.enabled=true spark.history.fs.logDirectory=seaweedfs://localhost:8888/spark2-history/ spark.eventLog.dir=seaweedfs://localhost:8888/spark2-history/ -spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.3.jar -spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.3.jar +spark.driver.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.4.jar +spark.executor.extraClassPath=/Users/chris/go/src/github.com/chrislusf/seaweedfs/other/java/hdfs2/target/seaweedfs-hadoop2-client-1.5.4.jar spark.hadoop.fs.seaweedfs.impl=seaweed.hdfs.SeaweedFileSystem spark.hadoop.fs.defaultFS=seaweedfs://localhost:8888 ```