|
|
|
@ -18,21 +18,27 @@ package main |
|
|
|
|
|
|
|
|
|
import ( |
|
|
|
|
"bytes" |
|
|
|
|
"context" |
|
|
|
|
"crypto/md5" |
|
|
|
|
crand "crypto/rand" |
|
|
|
|
"crypto/tls" |
|
|
|
|
"errors" |
|
|
|
|
"fmt" |
|
|
|
|
"io" |
|
|
|
|
"io/ioutil" |
|
|
|
|
"math/rand" |
|
|
|
|
"net/http" |
|
|
|
|
"net/http/httptrace" |
|
|
|
|
"os" |
|
|
|
|
"os/exec" |
|
|
|
|
"strings" |
|
|
|
|
"sync" |
|
|
|
|
"time" |
|
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/log" |
|
|
|
|
"github.com/ethereum/go-ethereum/metrics" |
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/api" |
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/api/client" |
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/spancontext" |
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/testutil" |
|
|
|
|
opentracing "github.com/opentracing/opentracing-go" |
|
|
|
|
"github.com/pborman/uuid" |
|
|
|
|
|
|
|
|
|
cli "gopkg.in/urfave/cli.v1" |
|
|
|
@ -40,11 +46,11 @@ import ( |
|
|
|
|
|
|
|
|
|
func generateEndpoints(scheme string, cluster string, app string, from int, to int) { |
|
|
|
|
if cluster == "prod" { |
|
|
|
|
for port := from; port <= to; port++ { |
|
|
|
|
for port := from; port < to; port++ { |
|
|
|
|
endpoints = append(endpoints, fmt.Sprintf("%s://%v.swarm-gateways.net", scheme, port)) |
|
|
|
|
} |
|
|
|
|
} else { |
|
|
|
|
for port := from; port <= to; port++ { |
|
|
|
|
for port := from; port < to; port++ { |
|
|
|
|
endpoints = append(endpoints, fmt.Sprintf("%s://%s-%v-%s.stg.swarm-gateways.net", scheme, app, port, cluster)) |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -58,22 +64,48 @@ func cliUploadAndSync(c *cli.Context) error { |
|
|
|
|
log.PrintOrigins(true) |
|
|
|
|
log.Root().SetHandler(log.LvlFilterHandler(log.Lvl(verbosity), log.StreamHandler(os.Stdout, log.TerminalFormat(true)))) |
|
|
|
|
|
|
|
|
|
defer func(now time.Time) { log.Info("total time", "time", time.Since(now), "kb", filesize) }(time.Now()) |
|
|
|
|
metrics.GetOrRegisterCounter("upload-and-sync", nil).Inc(1) |
|
|
|
|
|
|
|
|
|
generateEndpoints(scheme, cluster, appName, from, to) |
|
|
|
|
errc := make(chan error) |
|
|
|
|
go func() { |
|
|
|
|
errc <- uploadAndSync(c) |
|
|
|
|
}() |
|
|
|
|
|
|
|
|
|
log.Info("uploading to " + endpoints[0] + " and syncing") |
|
|
|
|
select { |
|
|
|
|
case err := <-errc: |
|
|
|
|
if err != nil { |
|
|
|
|
metrics.GetOrRegisterCounter("upload-and-sync.fail", nil).Inc(1) |
|
|
|
|
} |
|
|
|
|
return err |
|
|
|
|
case <-time.After(time.Duration(timeout) * time.Second): |
|
|
|
|
metrics.GetOrRegisterCounter("upload-and-sync.timeout", nil).Inc(1) |
|
|
|
|
return fmt.Errorf("timeout after %v sec", timeout) |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
func uploadAndSync(c *cli.Context) error { |
|
|
|
|
defer func(now time.Time) { |
|
|
|
|
totalTime := time.Since(now) |
|
|
|
|
|
|
|
|
|
log.Info("total time", "time", totalTime, "kb", filesize) |
|
|
|
|
metrics.GetOrRegisterCounter("upload-and-sync.total-time", nil).Inc(int64(totalTime)) |
|
|
|
|
}(time.Now()) |
|
|
|
|
|
|
|
|
|
generateEndpoints(scheme, cluster, appName, from, to) |
|
|
|
|
seed := int(time.Now().UnixNano() / 1e6) |
|
|
|
|
log.Info("uploading to "+endpoints[0]+" and syncing", "seed", seed) |
|
|
|
|
|
|
|
|
|
f, cleanup := generateRandomFile(filesize * 1000) |
|
|
|
|
defer cleanup() |
|
|
|
|
randomBytes := testutil.RandomBytes(seed, filesize*1000) |
|
|
|
|
|
|
|
|
|
hash, err := upload(f, endpoints[0]) |
|
|
|
|
t1 := time.Now() |
|
|
|
|
hash, err := upload(&randomBytes, endpoints[0]) |
|
|
|
|
if err != nil { |
|
|
|
|
log.Error(err.Error()) |
|
|
|
|
return err |
|
|
|
|
} |
|
|
|
|
metrics.GetOrRegisterCounter("upload-and-sync.upload-time", nil).Inc(int64(time.Since(t1))) |
|
|
|
|
|
|
|
|
|
fhash, err := digest(f) |
|
|
|
|
fhash, err := digest(bytes.NewReader(randomBytes)) |
|
|
|
|
if err != nil { |
|
|
|
|
log.Error(err.Error()) |
|
|
|
|
return err |
|
|
|
@ -81,23 +113,47 @@ func cliUploadAndSync(c *cli.Context) error { |
|
|
|
|
|
|
|
|
|
log.Info("uploaded successfully", "hash", hash, "digest", fmt.Sprintf("%x", fhash)) |
|
|
|
|
|
|
|
|
|
time.Sleep(3 * time.Second) |
|
|
|
|
time.Sleep(time.Duration(syncDelay) * time.Second) |
|
|
|
|
|
|
|
|
|
wg := sync.WaitGroup{} |
|
|
|
|
for _, endpoint := range endpoints { |
|
|
|
|
if single { |
|
|
|
|
rand.Seed(time.Now().UTC().UnixNano()) |
|
|
|
|
randIndex := 1 + rand.Intn(len(endpoints)-1) |
|
|
|
|
ruid := uuid.New()[:8] |
|
|
|
|
wg.Add(1) |
|
|
|
|
go func(endpoint string, ruid string) { |
|
|
|
|
for { |
|
|
|
|
start := time.Now() |
|
|
|
|
err := fetch(hash, endpoint, fhash, ruid) |
|
|
|
|
fetchTime := time.Since(start) |
|
|
|
|
if err != nil { |
|
|
|
|
continue |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
metrics.GetOrRegisterMeter("upload-and-sync.single.fetch-time", nil).Mark(int64(fetchTime)) |
|
|
|
|
wg.Done() |
|
|
|
|
return |
|
|
|
|
} |
|
|
|
|
}(endpoint, ruid) |
|
|
|
|
}(endpoints[randIndex], ruid) |
|
|
|
|
} else { |
|
|
|
|
for _, endpoint := range endpoints { |
|
|
|
|
ruid := uuid.New()[:8] |
|
|
|
|
wg.Add(1) |
|
|
|
|
go func(endpoint string, ruid string) { |
|
|
|
|
for { |
|
|
|
|
start := time.Now() |
|
|
|
|
err := fetch(hash, endpoint, fhash, ruid) |
|
|
|
|
fetchTime := time.Since(start) |
|
|
|
|
if err != nil { |
|
|
|
|
continue |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
metrics.GetOrRegisterMeter("upload-and-sync.each.fetch-time", nil).Mark(int64(fetchTime)) |
|
|
|
|
wg.Done() |
|
|
|
|
return |
|
|
|
|
} |
|
|
|
|
}(endpoint, ruid) |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
wg.Wait() |
|
|
|
|
log.Info("all endpoints synced random file successfully") |
|
|
|
@ -107,16 +163,33 @@ func cliUploadAndSync(c *cli.Context) error { |
|
|
|
|
|
|
|
|
|
// fetch is getting the requested `hash` from the `endpoint` and compares it with the `original` file
|
|
|
|
|
func fetch(hash string, endpoint string, original []byte, ruid string) error { |
|
|
|
|
ctx, sp := spancontext.StartSpan(context.Background(), "upload-and-sync.fetch") |
|
|
|
|
defer sp.Finish() |
|
|
|
|
|
|
|
|
|
log.Trace("sleeping", "ruid", ruid) |
|
|
|
|
time.Sleep(3 * time.Second) |
|
|
|
|
|
|
|
|
|
log.Trace("http get request", "ruid", ruid, "api", endpoint, "hash", hash) |
|
|
|
|
client := &http.Client{Transport: &http.Transport{ |
|
|
|
|
TLSClientConfig: &tls.Config{InsecureSkipVerify: true}, |
|
|
|
|
}} |
|
|
|
|
res, err := client.Get(endpoint + "/bzz:/" + hash + "/") |
|
|
|
|
|
|
|
|
|
var tn time.Time |
|
|
|
|
reqUri := endpoint + "/bzz:/" + hash + "/" |
|
|
|
|
req, _ := http.NewRequest("GET", reqUri, nil) |
|
|
|
|
|
|
|
|
|
opentracing.GlobalTracer().Inject( |
|
|
|
|
sp.Context(), |
|
|
|
|
opentracing.HTTPHeaders, |
|
|
|
|
opentracing.HTTPHeadersCarrier(req.Header)) |
|
|
|
|
|
|
|
|
|
trace := client.GetClientTrace("upload-and-sync - http get", "upload-and-sync", ruid, &tn) |
|
|
|
|
|
|
|
|
|
req = req.WithContext(httptrace.WithClientTrace(ctx, trace)) |
|
|
|
|
transport := http.DefaultTransport |
|
|
|
|
|
|
|
|
|
//transport.TLSClientConfig = &tls.Config{InsecureSkipVerify: true}
|
|
|
|
|
|
|
|
|
|
tn = time.Now() |
|
|
|
|
res, err := transport.RoundTrip(req) |
|
|
|
|
if err != nil { |
|
|
|
|
log.Warn(err.Error(), "ruid", ruid) |
|
|
|
|
log.Error(err.Error(), "ruid", ruid) |
|
|
|
|
return err |
|
|
|
|
} |
|
|
|
|
log.Trace("http get response", "ruid", ruid, "api", endpoint, "hash", hash, "code", res.StatusCode, "len", res.ContentLength) |
|
|
|
@ -147,16 +220,19 @@ func fetch(hash string, endpoint string, original []byte, ruid string) error { |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// upload is uploading a file `f` to `endpoint` via the `swarm up` cmd
|
|
|
|
|
func upload(f *os.File, endpoint string) (string, error) { |
|
|
|
|
var out bytes.Buffer |
|
|
|
|
cmd := exec.Command("swarm", "--bzzapi", endpoint, "up", f.Name()) |
|
|
|
|
cmd.Stdout = &out |
|
|
|
|
err := cmd.Run() |
|
|
|
|
if err != nil { |
|
|
|
|
return "", err |
|
|
|
|
func upload(dataBytes *[]byte, endpoint string) (string, error) { |
|
|
|
|
swarm := client.NewClient(endpoint) |
|
|
|
|
f := &client.File{ |
|
|
|
|
ReadCloser: ioutil.NopCloser(bytes.NewReader(*dataBytes)), |
|
|
|
|
ManifestEntry: api.ManifestEntry{ |
|
|
|
|
ContentType: "text/plain", |
|
|
|
|
Mode: 0660, |
|
|
|
|
Size: int64(len(*dataBytes)), |
|
|
|
|
}, |
|
|
|
|
} |
|
|
|
|
hash := strings.TrimRight(out.String(), "\r\n") |
|
|
|
|
return hash, nil |
|
|
|
|
|
|
|
|
|
// upload data to bzz:// and retrieve the content-addressed manifest hash, hex-encoded.
|
|
|
|
|
return swarm.Upload(f, "", false) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
func digest(r io.Reader) ([]byte, error) { |
|
|
|
@ -179,27 +255,3 @@ func generateRandomData(datasize int) ([]byte, error) { |
|
|
|
|
} |
|
|
|
|
return b, nil |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// generateRandomFile is creating a temporary file with the requested byte size
|
|
|
|
|
func generateRandomFile(size int) (f *os.File, teardown func()) { |
|
|
|
|
// create a tmp file
|
|
|
|
|
tmp, err := ioutil.TempFile("", "swarm-test") |
|
|
|
|
if err != nil { |
|
|
|
|
panic(err) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// callback for tmp file cleanup
|
|
|
|
|
teardown = func() { |
|
|
|
|
tmp.Close() |
|
|
|
|
os.Remove(tmp.Name()) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
buf := make([]byte, size) |
|
|
|
|
_, err = crand.Read(buf) |
|
|
|
|
if err != nil { |
|
|
|
|
panic(err) |
|
|
|
|
} |
|
|
|
|
ioutil.WriteFile(tmp.Name(), buf, 0755) |
|
|
|
|
|
|
|
|
|
return tmp, teardown |
|
|
|
|
} |
|
|
|
|