mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2024-01-19 02:48:24 +00:00
rename weedc option name to publicUrl
git-svn-id: https://weed-fs.googlecode.com/svn/trunk@27 282b0af5-e82d-9cf1-ede4-77906d7719d0
This commit is contained in:
parent
c27f6175d7
commit
4c36190dd3
|
@ -17,8 +17,8 @@ var (
|
|||
port = flag.Int("port", 8080, "http listen port")
|
||||
chunkFolder = flag.String("dir", "/tmp", "data directory to store files")
|
||||
volumes = flag.String("volumes", "0,1-3,4", "comma-separated list of volume ids or range of ids")
|
||||
publicServer = flag.String("pserver", "localhost:8080", "public server to serve data read")
|
||||
metaServer = flag.String("mserver", "localhost:9333", "metadata server to store mappings")
|
||||
publicUrl = flag.String("publicUrl", "localhost:8080", "public url to serve data read")
|
||||
metaServer = flag.String("mserver", "localhost:9333", "master directory server to store mappings")
|
||||
IsDebug = flag.Bool("debug", false, "enable debug mode")
|
||||
pulse = flag.Int("pulseSeconds", 5, "number of seconds between heartbeats")
|
||||
|
||||
|
@ -128,7 +128,7 @@ func parseURLPath(path string) (vid, fid, ext string) {
|
|||
func main() {
|
||||
flag.Parse()
|
||||
//TODO: now default to 1G, this value should come from server?
|
||||
store = storage.NewStore(*port, *publicServer, *chunkFolder, *volumes)
|
||||
store = storage.NewStore(*port, *publicUrl, *chunkFolder, *volumes)
|
||||
defer store.Close()
|
||||
http.HandleFunc("/", storeHandler)
|
||||
http.HandleFunc("/status", statusHandler)
|
||||
|
|
|
@ -40,11 +40,11 @@ func dirAssignHandler(w http.ResponseWriter, r *http.Request) {
|
|||
}
|
||||
func dirJoinHandler(w http.ResponseWriter, r *http.Request) {
|
||||
s := r.RemoteAddr[0:strings.Index(r.RemoteAddr, ":")+1] + r.FormValue("port")
|
||||
publicServer := r.FormValue("publicServer")
|
||||
publicUrl := r.FormValue("publicUrl")
|
||||
volumes := new([]storage.VolumeInfo)
|
||||
json.Unmarshal([]byte(r.FormValue("volumes")), volumes)
|
||||
log.Println("Recieved updates from", s, "volumes", r.FormValue("volumes"))
|
||||
mapper.Add(*directory.NewMachine(s, publicServer, *volumes))
|
||||
mapper.Add(*directory.NewMachine(s, publicUrl, *volumes))
|
||||
}
|
||||
func dirStatusHandler(w http.ResponseWriter, r *http.Request) {
|
||||
writeJson(w, r, mapper)
|
||||
|
|
|
@ -37,8 +37,8 @@ type Mapper struct {
|
|||
fileIdCounter uint64
|
||||
}
|
||||
|
||||
func NewMachine(server, publicServer string, volumes []storage.VolumeInfo) *Machine {
|
||||
return &Machine{Server: MachineInfo{Url: server, PublicUrl: publicServer}, Volumes: volumes}
|
||||
func NewMachine(server, publicUrl string, volumes []storage.VolumeInfo) *Machine {
|
||||
return &Machine{Server: MachineInfo{Url: server, PublicUrl: publicUrl}, Volumes: volumes}
|
||||
}
|
||||
|
||||
func NewMapper(dirname string, filename string) (m *Mapper) {
|
||||
|
|
|
@ -14,15 +14,15 @@ type Store struct {
|
|||
volumes map[uint64]*Volume
|
||||
dir string
|
||||
Port int
|
||||
PublicServer string
|
||||
PublicUrl string
|
||||
}
|
||||
type VolumeInfo struct {
|
||||
Id uint32
|
||||
Size int64
|
||||
}
|
||||
|
||||
func NewStore(port int, publicServer, dirname string, volumeListString string) (s *Store) {
|
||||
s = &Store{Port: port, PublicServer: publicServer, dir: dirname}
|
||||
func NewStore(port int, publicUrl, dirname string, volumeListString string) (s *Store) {
|
||||
s = &Store{Port: port, PublicUrl: publicUrl, dir: dirname}
|
||||
s.volumes = make(map[uint64]*Volume)
|
||||
|
||||
for _, range_string := range strings.Split(volumeListString, ",") {
|
||||
|
@ -74,7 +74,7 @@ func (s *Store) Join(mserver string) {
|
|||
bytes, _ := json.Marshal(stats)
|
||||
values := make(url.Values)
|
||||
values.Add("port", strconv.Itoa(s.Port))
|
||||
values.Add("publicServer", s.PublicServer)
|
||||
values.Add("publicUrl", s.PublicUrl)
|
||||
values.Add("volumes", string(bytes))
|
||||
log.Println("Exiting volumes", string(bytes))
|
||||
util.Post("http://"+mserver+"/dir/join", values)
|
||||
|
|
Loading…
Reference in a new issue