mirror of
https://github.com/httprunner/httprunner.git
synced 2026-05-11 18:11:21 +08:00
Merge branch 'master' into ios-wda
This commit is contained in:
@@ -14,6 +14,7 @@
|
||||
- fix: optimize websocket step initialization
|
||||
- fix: reuse plugin instance if already initialized
|
||||
- fix: deep copy api step to avoid data racing
|
||||
- feat: support ping/dns/traceroute for dial test
|
||||
|
||||
## v4.1.6 (2022-07-04)
|
||||
|
||||
|
||||
2
go.mod
2
go.mod
@@ -10,6 +10,7 @@ require (
|
||||
github.com/getsentry/sentry-go v0.13.0
|
||||
github.com/go-errors/errors v1.0.1
|
||||
github.com/go-openapi/spec v0.20.6
|
||||
github.com/go-ping/ping v1.1.0
|
||||
github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510
|
||||
github.com/google/uuid v1.3.0
|
||||
github.com/gorilla/websocket v1.4.1
|
||||
@@ -18,6 +19,7 @@ require (
|
||||
github.com/jmespath/go-jmespath v0.4.0
|
||||
github.com/json-iterator/go v1.1.12
|
||||
github.com/maja42/goval v1.2.1
|
||||
github.com/miekg/dns v1.1.25
|
||||
github.com/mitchellh/mapstructure v1.4.1
|
||||
github.com/olekukonko/tablewriter v0.0.5
|
||||
github.com/pkg/errors v0.9.1
|
||||
|
||||
12
go.sum
12
go.sum
@@ -149,6 +149,8 @@ github.com/go-openapi/spec v0.20.6/go.mod h1:2OpW+JddWPrpXSCIX8eOx7lZ5iyuWj3RYR6
|
||||
github.com/go-openapi/swag v0.19.5/go.mod h1:POnQmlKehdgb5mhVOsnJFsivZCEZ/vjK9gh66Z9tfKk=
|
||||
github.com/go-openapi/swag v0.19.15 h1:D2NRCBzS9/pEY3gP9Nl8aDqGUcPFrwG2p+CNFrLyrCM=
|
||||
github.com/go-openapi/swag v0.19.15/go.mod h1:QYRuS/SOXUCsnplDa677K7+DxSOj6IPNl/eQntq43wQ=
|
||||
github.com/go-ping/ping v1.1.0 h1:3MCGhVX4fyEUuhsfwPrsEdQw6xspHkv5zHsiSoDFZYw=
|
||||
github.com/go-ping/ping v1.1.0/go.mod h1:xIFjORFzTxqIV/tDVGO4eDy/bLuSyawEeojSm3GfRGk=
|
||||
github.com/go-playground/assert/v2 v2.0.1/go.mod h1:VDjEfimB/XKnb+ZQfWdccd7VUvScMdVu0Titje2rxJ4=
|
||||
github.com/go-playground/locales v0.13.0/go.mod h1:taPMhCMXrRLJO55olJkUXHZBHCxTMfnGwq/HNwmWNS8=
|
||||
github.com/go-playground/universal-translator v0.17.0/go.mod h1:UkSxE5sNxxRwHyU+Scu5vgOQjsIJAF8j9muTVoKLVtA=
|
||||
@@ -226,6 +228,7 @@ github.com/google/renameio v0.1.0/go.mod h1:KWCgfxg9yswjAJkECMjeO8J8rahYeXnNhOm4
|
||||
github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 h1:El6M4kTTCOh6aBiKaUGG7oYTSPP8MxqL4YI3kZKwcP4=
|
||||
github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ=
|
||||
github.com/google/uuid v1.1.2/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
||||
github.com/google/uuid v1.2.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
||||
github.com/google/uuid v1.3.0 h1:t6JiXgmwXMjEs8VusXIJk2BXHsn+wx8BZdTaoZ5fu7I=
|
||||
github.com/google/uuid v1.3.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
||||
github.com/googleapis/gax-go/v2 v2.0.4/go.mod h1:0Wqv26UfaUD9n4G6kQubkQ+KchISgw+vpHVxEJEs9eg=
|
||||
@@ -351,6 +354,8 @@ github.com/matttproud/golang_protobuf_extensions v1.0.1/go.mod h1:D8He9yQNgCq6Z5
|
||||
github.com/mediocregopher/radix/v3 v3.4.2/go.mod h1:8FL3F6UQRXHXIBSPUs5h0RybMF8i4n7wVopoX3x7Bv8=
|
||||
github.com/microcosm-cc/bluemonday v1.0.2/go.mod h1:iVP4YcDBq+n/5fb23BhYFvIMq/leAFZyRl6bYmGDlGc=
|
||||
github.com/miekg/dns v1.0.14/go.mod h1:W1PPwlIAgtquWBMBEV9nkV9Cazfe8ScdGz/Lj7v3Nrg=
|
||||
github.com/miekg/dns v1.1.25 h1:dFwPR6SfLtrSwgDcIq2bcU/gVutB4sNApq2HBdqcakg=
|
||||
github.com/miekg/dns v1.1.25/go.mod h1:bPDLeHnStXmXAq1m/Ch/hvfNHr14JKNPMBo3VZKjuso=
|
||||
github.com/mitchellh/cli v1.0.0/go.mod h1:hNIlj7HEI86fIcpObd7a0FcrxTWetlwJDGcceTlRvqc=
|
||||
github.com/mitchellh/go-homedir v1.0.0/go.mod h1:SfyaCUpYCn1Vlf4IUYiD9fPX4A5wJrkLzIz1N1q0pr0=
|
||||
github.com/mitchellh/go-homedir v1.1.0/go.mod h1:SfyaCUpYCn1Vlf4IUYiD9fPX4A5wJrkLzIz1N1q0pr0=
|
||||
@@ -514,11 +519,13 @@ golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8U
|
||||
golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||
golang.org/x/crypto v0.0.0-20190701094942-4def268fd1a4/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||
golang.org/x/crypto v0.0.0-20190820162420-60c769a6c586/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||
golang.org/x/crypto v0.0.0-20190923035154-9ee001bba392/go.mod h1:/lpIB1dKB+9EgE3H3cr1v9wB50oz8l4C4h62xy7jSTY=
|
||||
golang.org/x/crypto v0.0.0-20191011191535-87dc89f01550/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||
golang.org/x/crypto v0.0.0-20191227163750-53104e6ec876/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto=
|
||||
golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto=
|
||||
golang.org/x/crypto v0.0.0-20210322153248-0c34fe9e7dc2/go.mod h1:T9bdIzuCu7OtxOm1hfPfRQxPLYneinmdGuTeoZ9dtd4=
|
||||
golang.org/x/crypto v0.0.0-20210921155107-089bfa567519/go.mod h1:GvvjBRRGRdwPK5ydBHafDWAxML/pGHZbMvKqRZ5+Abc=
|
||||
golang.org/x/crypto v0.0.0-20211215165025-cf75a172585e h1:1SzTfNOXwIS2oWiMF+6qu0OUDKb0dauo6MoDUQyu+yU=
|
||||
golang.org/x/crypto v0.0.0-20211215165025-cf75a172585e/go.mod h1:P+XmwS30IXTQdn5tA2iutPOUgjI07+tq3H3K9MVA1s8=
|
||||
golang.org/x/exp v0.0.0-20190121172915-509febef88a4/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA=
|
||||
golang.org/x/exp v0.0.0-20190306152737-a1d7652674e8/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA=
|
||||
@@ -575,6 +582,7 @@ golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLL
|
||||
golang.org/x/net v0.0.0-20190628185345-da137c7871d7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
golang.org/x/net v0.0.0-20190827160401-ba9fcec4b297/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
golang.org/x/net v0.0.0-20190923162816-aa69164e4478/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
golang.org/x/net v0.0.0-20200114155413-6afb5195e5aa/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||
@@ -624,6 +632,7 @@ golang.org/x/sync v0.0.0-20200317015054-43a5402ce75a/go.mod h1:RxMgew5VJxzue5/jJ
|
||||
golang.org/x/sync v0.0.0-20200625203802-6e8e738ad208/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||
golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||
golang.org/x/sync v0.0.0-20201207232520-09787c993a3a/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||
golang.org/x/sync v0.0.0-20210220032951-036812b2e83c h1:5KslGYwFpkhGh+Q16bwMP3cOontH8FOep7tGV86Y7SQ=
|
||||
golang.org/x/sync v0.0.0-20210220032951-036812b2e83c/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||
golang.org/x/sys v0.0.0-20180823144017-11551d06cbcc/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||
golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||
@@ -644,6 +653,8 @@ golang.org/x/sys v0.0.0-20190626221950-04f50cda93cb/go.mod h1:h1NjWce9XRLGQEsW7w
|
||||
golang.org/x/sys v0.0.0-20190726091711-fc99dfbffb4e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20190813064441-fde4db37ae7a/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20190916202348-b4ddaad3f8a3/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20190922100055-0a153f010e69/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20190924154521-2837fb4f24fe/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20191001151750-bb3f8db39f24/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20191005200804-aed5e4c7ecf9/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
golang.org/x/sys v0.0.0-20191008105621-543471e840be/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||
@@ -723,6 +734,7 @@ golang.org/x/tools v0.0.0-20190606124116-d0a3d012864b/go.mod h1:/rFqwRUd4F7ZHNgw
|
||||
golang.org/x/tools v0.0.0-20190621195816-6e04913cbbac/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc=
|
||||
golang.org/x/tools v0.0.0-20190628153133-6cdbf07be9d0/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc=
|
||||
golang.org/x/tools v0.0.0-20190816200558-6889da9d5479/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo=
|
||||
golang.org/x/tools v0.0.0-20190907020128-2ca718005c18/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo=
|
||||
golang.org/x/tools v0.0.0-20190911174233-4f2ddba30aff/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo=
|
||||
golang.org/x/tools v0.0.0-20191012152004-8de300cfc20a/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo=
|
||||
golang.org/x/tools v0.0.0-20191112195655-aa38f8e97acc/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo=
|
||||
|
||||
109
hrp/boomer.go
109
hrp/boomer.go
@@ -2,8 +2,6 @@ package hrp
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
"golang.org/x/net/context"
|
||||
"io/ioutil"
|
||||
"os"
|
||||
"path/filepath"
|
||||
@@ -13,9 +11,11 @@ import (
|
||||
|
||||
"github.com/httprunner/funplugin"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/boomer"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/json"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/sdk"
|
||||
"github.com/rs/zerolog/log"
|
||||
"golang.org/x/net/context"
|
||||
)
|
||||
|
||||
func NewStandaloneBoomer(spawnCount int64, spawnRate float64) *HRPBoomer {
|
||||
@@ -57,7 +57,6 @@ type HRPBoomer struct {
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) InitBoomer() {
|
||||
// init output
|
||||
if !b.GetProfile().DisableConsoleOutput {
|
||||
b.AddOutput(boomer.NewConsoleOutput())
|
||||
}
|
||||
@@ -100,6 +99,16 @@ func (b *HRPBoomer) Run(testcases ...ITestCase) {
|
||||
// report execution timing event
|
||||
defer sdk.SendEvent(event.StartTiming("execution"))
|
||||
|
||||
// quit all plugins
|
||||
defer func() {
|
||||
pluginMap.Range(func(key, value interface{}) bool {
|
||||
if plugin, ok := value.(funplugin.IPlugin); ok {
|
||||
plugin.Quit()
|
||||
}
|
||||
return true
|
||||
})
|
||||
}()
|
||||
|
||||
taskSlice := b.ConvertTestCasesToBoomerTasks(testcases...)
|
||||
|
||||
b.Boomer.Run(taskSlice...)
|
||||
@@ -113,15 +122,6 @@ func (b *HRPBoomer) ConvertTestCasesToBoomerTasks(testcases ...ITestCase) (taskS
|
||||
os.Exit(1)
|
||||
}
|
||||
|
||||
// quit all plugins
|
||||
defer func() {
|
||||
if len(pluginMap) > 0 {
|
||||
for _, plugin := range pluginMap {
|
||||
plugin.Quit()
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
for _, testcase := range testCases {
|
||||
rendezvousList := initRendezvous(testcase, int64(b.GetSpawnCount()))
|
||||
task := b.convertBoomerTask(testcase, rendezvousList)
|
||||
@@ -164,7 +164,7 @@ func (b *HRPBoomer) TestCasesToBytes(testcases ...ITestCase) []byte {
|
||||
return testCasesBytes
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) BytesToTestCases(testCasesBytes []byte) []*TCase {
|
||||
func (b *HRPBoomer) BytesToTCases(testCasesBytes []byte) []*TCase {
|
||||
var testcase []*TCase
|
||||
err := json.Unmarshal(testCasesBytes, &testcase)
|
||||
if err != nil {
|
||||
@@ -177,39 +177,57 @@ func (b *HRPBoomer) Quit() {
|
||||
b.Boomer.Quit()
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) runTestCases(testCases []*TCase, profile *boomer.Profile) {
|
||||
var testcases []ITestCase
|
||||
func (b *HRPBoomer) parseTCases(testCases []*TCase) (testcases []ITestCase) {
|
||||
for _, tc := range testCases {
|
||||
tesecase, err := tc.toTestCase()
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to load testcases")
|
||||
return
|
||||
}
|
||||
// create temp dir to save testcase
|
||||
tempDir, err := ioutil.TempDir("", "hrp_testcases")
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to save testcases")
|
||||
log.Error().Err(err).Msg("failed to create hrp testcases directory")
|
||||
return
|
||||
}
|
||||
|
||||
tesecase.Config.Path = filepath.Join(tempDir, "test-case.json")
|
||||
if tesecase.Config.PluginSetting != nil {
|
||||
tesecase.Config.PluginSetting.Path = filepath.Join(tempDir, fmt.Sprintf("debugtalk.%s", tesecase.Config.PluginSetting.Type))
|
||||
err = builtin.Bytes2File(tesecase.Config.PluginSetting.Content, tesecase.Config.PluginSetting.Path)
|
||||
if tc.Config.PluginSetting != nil {
|
||||
tc.Config.PluginSetting.Path = filepath.Join(tempDir, fmt.Sprintf("debugtalk.%s", tc.Config.PluginSetting.Type))
|
||||
err = builtin.Bytes2File(tc.Config.PluginSetting.Content, tc.Config.PluginSetting.Path)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to save plugin file")
|
||||
return
|
||||
}
|
||||
tc.Config.PluginSetting.Content = nil // remove the content in testcase
|
||||
}
|
||||
err = builtin.Dump2JSON(tesecase, tesecase.Config.Path)
|
||||
|
||||
if tc.Config.Environs != nil {
|
||||
envContent := ""
|
||||
for k, v := range tc.Config.Environs {
|
||||
envContent += fmt.Sprintf("%s=%s\n", k, v)
|
||||
}
|
||||
err = os.WriteFile(filepath.Join(tempDir, ".env"), []byte(envContent), 0o644)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to dump environs")
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
tc.Config.Path = filepath.Join(tempDir, "test-case.json")
|
||||
err = builtin.Dump2JSON(tc, tc.Config.Path)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to dump testcases")
|
||||
return
|
||||
}
|
||||
|
||||
tesecase, err := tc.toTestCase()
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to load testcases")
|
||||
return
|
||||
}
|
||||
|
||||
testcases = append(testcases, tesecase)
|
||||
}
|
||||
return testcases
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) initWorker(profile *boomer.Profile) {
|
||||
// if no IP address is specified, the default IP address is that of the master
|
||||
if profile.PrometheusPushgatewayURL != "" {
|
||||
urlSlice := strings.Split(profile.PrometheusPushgatewayURL, ":")
|
||||
if len(urlSlice) != 2 {
|
||||
@@ -224,16 +242,13 @@ func (b *HRPBoomer) runTestCases(testCases []*TCase, profile *boomer.Profile) {
|
||||
|
||||
b.SetProfile(profile)
|
||||
b.InitBoomer()
|
||||
log.Info().Interface("testcases", testcases).Interface("profile", profile).Msg("run tasks successful")
|
||||
b.Run(testcases...)
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) rebalanceBoomer(profile *boomer.Profile) {
|
||||
b.SetProfile(profile)
|
||||
b.SetSpawnCount(b.GetProfile().SpawnCount)
|
||||
b.SetSpawnRate(b.GetProfile().SpawnRate)
|
||||
func (b *HRPBoomer) rebalanceRunner(profile *boomer.Profile) {
|
||||
b.SetSpawnCount(profile.SpawnCount)
|
||||
b.SetSpawnRate(profile.SpawnRate)
|
||||
b.GetRebalanceChan() <- true
|
||||
log.Info().Interface("profile", profile).Msg("rebalance tasks successful")
|
||||
log.Info().Interface("profile", profile).Msg("rebalance tasks successfully")
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) PollTasks(ctx context.Context) {
|
||||
@@ -245,11 +260,17 @@ func (b *HRPBoomer) PollTasks(ctx context.Context) {
|
||||
continue
|
||||
}
|
||||
//Todo: 过滤掉已经传输过的task
|
||||
if task.TestCases != nil {
|
||||
testCases := b.BytesToTestCases(task.TestCases)
|
||||
go b.runTestCases(testCases, task.Profile)
|
||||
if task.TestCasesBytes != nil {
|
||||
// init boomer with profile
|
||||
b.initWorker(task.Profile)
|
||||
// get testcases
|
||||
testcases := b.parseTCases(b.BytesToTCases(task.TestCasesBytes))
|
||||
log.Info().Interface("testcases", testcases).Interface("profile", b.GetProfile()).Msg("starting to run tasks")
|
||||
// run testcases
|
||||
go b.Run(testcases...)
|
||||
} else {
|
||||
go b.rebalanceBoomer(task.Profile)
|
||||
// rebalance runner with profile
|
||||
go b.rebalanceRunner(task.Profile)
|
||||
}
|
||||
|
||||
case <-b.Boomer.GetCloseChan():
|
||||
@@ -261,6 +282,16 @@ func (b *HRPBoomer) PollTasks(ctx context.Context) {
|
||||
}
|
||||
|
||||
func (b *HRPBoomer) PollTestCases(ctx context.Context) {
|
||||
// quit all plugins
|
||||
defer func() {
|
||||
pluginMap.Range(func(key, value interface{}) bool {
|
||||
if plugin, ok := value.(funplugin.IPlugin); ok {
|
||||
plugin.Quit()
|
||||
}
|
||||
return true
|
||||
})
|
||||
}()
|
||||
|
||||
for {
|
||||
select {
|
||||
case <-b.Boomer.ParseTestCasesChan():
|
||||
@@ -270,7 +301,7 @@ func (b *HRPBoomer) PollTestCases(ctx context.Context) {
|
||||
tcs = append(tcs, &tcp)
|
||||
}
|
||||
b.TestCaseBytesChan() <- b.TestCasesToBytes(tcs...)
|
||||
log.Info().Msg("put testcase successful")
|
||||
log.Info().Msg("put testcase successfully")
|
||||
case <-b.Boomer.GetCloseChan():
|
||||
return
|
||||
case <-ctx.Done():
|
||||
@@ -379,7 +410,7 @@ func (b *HRPBoomer) convertBoomerTask(testcase *TestCase, rendezvousList []*Rend
|
||||
// transaction
|
||||
// FIXME: support nested transactions
|
||||
if step.Struct().Transaction.Type == transactionEnd { // only record when transaction ends
|
||||
b.RecordTransaction(stepResult.Name, transactionSuccess, stepResult.Elapsed, 0)
|
||||
b.RecordTransaction(step.Struct().Transaction.Name, transactionSuccess, stepResult.Elapsed, 0)
|
||||
transactionSuccess = true // reset flag for next transaction
|
||||
}
|
||||
} else if stepResult.StepType == stepTypeRendezvous {
|
||||
|
||||
96
hrp/cmd/dial.go
Normal file
96
hrp/cmd/dial.go
Normal file
@@ -0,0 +1,96 @@
|
||||
package cmd
|
||||
|
||||
import (
|
||||
"runtime"
|
||||
"time"
|
||||
|
||||
"github.com/rs/zerolog/log"
|
||||
"github.com/spf13/cobra"
|
||||
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/dial"
|
||||
)
|
||||
|
||||
var (
|
||||
pingOptions dial.PingOptions
|
||||
dnsOptions dial.DnsOptions
|
||||
traceRouteOptions dial.TraceRouteOptions
|
||||
)
|
||||
|
||||
var pingCmd = &cobra.Command{
|
||||
Use: "ping $url",
|
||||
Short: "run integrated ping command",
|
||||
Args: cobra.ExactArgs(1),
|
||||
PreRun: func(cmd *cobra.Command, args []string) {
|
||||
setLogLevel(logLevel)
|
||||
},
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return dial.DoPing(&pingOptions, args)
|
||||
},
|
||||
}
|
||||
|
||||
var dnsCmd = &cobra.Command{
|
||||
Use: "dns $url",
|
||||
Short: "DNS resolution for different source and record types",
|
||||
Args: cobra.ExactArgs(1),
|
||||
PreRun: func(cmd *cobra.Command, args []string) {
|
||||
setLogLevel(logLevel)
|
||||
},
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
if dnsOptions.DnsSourceType != dial.DnsSourceTypeLocal && dnsOptions.DnsServer != "" {
|
||||
log.Warn().Msg("DNS server not supported for non-local DNS source, ignored")
|
||||
}
|
||||
if dnsOptions.DnsSourceType == dial.DnsSourceTypeHttp && dnsOptions.DnsRecordType == dial.DnsRecordTypeCNAME {
|
||||
log.Warn().Msg("CNAME record not supported for http DNS source, using default record type(A)")
|
||||
}
|
||||
return dial.DoDns(&dnsOptions, args)
|
||||
},
|
||||
}
|
||||
|
||||
var traceRouteCmd = &cobra.Command{
|
||||
Use: "traceroute $url",
|
||||
Short: "run integrated traceroute command",
|
||||
Args: cobra.ExactArgs(1),
|
||||
PreRun: func(cmd *cobra.Command, args []string) {
|
||||
setLogLevel(logLevel)
|
||||
},
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
if runtime.GOOS == "windows" {
|
||||
log.Info().Msg("using default probe number (3) on Windows")
|
||||
}
|
||||
return dial.DoTraceRoute(&traceRouteOptions, args)
|
||||
},
|
||||
}
|
||||
|
||||
var curlCmd = &cobra.Command{
|
||||
Use: "curl $url",
|
||||
Short: "run integrated curl command",
|
||||
Args: cobra.MinimumNArgs(1),
|
||||
DisableFlagParsing: true,
|
||||
PreRun: func(cmd *cobra.Command, args []string) {
|
||||
setLogLevel(logLevel)
|
||||
},
|
||||
RunE: func(cmd *cobra.Command, args []string) error {
|
||||
return dial.DoCurl(args)
|
||||
},
|
||||
}
|
||||
|
||||
func init() {
|
||||
rootCmd.AddCommand(pingCmd)
|
||||
pingCmd.Flags().IntVarP(&pingOptions.Count, "count", "c", 10, "Stop after sending (and receiving) N packets")
|
||||
pingCmd.Flags().DurationVarP(&pingOptions.Timeout, "timeout", "t", 20*time.Second, "Ping exits after N seconds")
|
||||
pingCmd.Flags().DurationVarP(&pingOptions.Interval, "interval", "i", 1*time.Second, "Wait N seconds between sending each packet")
|
||||
pingCmd.Flags().BoolVar(&pingOptions.SaveTests, "save-tests", false, "Save ping result as json")
|
||||
|
||||
rootCmd.AddCommand(dnsCmd)
|
||||
dnsCmd.Flags().IntVar(&dnsOptions.DnsSourceType, "dns-source", 0, "DNS source type\n0: local DNS\n1: http DNS\n2: google DNS")
|
||||
dnsCmd.Flags().IntVar(&dnsOptions.DnsRecordType, "dns-record", 1, "DNS record type\n1: A\n28: AAAA\n5: CNAME")
|
||||
dnsCmd.Flags().StringVar(&dnsOptions.DnsServer, "dns-server", "", "DNS server, only available for local DNS source")
|
||||
dnsCmd.Flags().BoolVar(&dnsOptions.SaveTests, "save-tests", false, "Save DNS resolution result as json")
|
||||
|
||||
rootCmd.AddCommand(traceRouteCmd)
|
||||
traceRouteCmd.Flags().IntVarP(&traceRouteOptions.MaxTTL, "max-hops", "m", 30, "Set the max number of hops (max TTL to be reached)")
|
||||
traceRouteCmd.Flags().IntVarP(&traceRouteOptions.Queries, "queries", "q", 1, "Set the number of probes per each hop")
|
||||
traceRouteCmd.Flags().BoolVar(&traceRouteOptions.SaveTests, "save-tests", false, "Save traceroute result as json")
|
||||
|
||||
rootCmd.AddCommand(curlCmd)
|
||||
}
|
||||
@@ -63,6 +63,18 @@ type Profile struct {
|
||||
DisableKeepalive bool `json:"disable-keepalive,omitempty" yaml:"disable-keepalive,omitempty" mapstructure:"disable-keepalive,omitempty"`
|
||||
}
|
||||
|
||||
func NewProfile() *Profile {
|
||||
return &Profile{
|
||||
SpawnCount: 1,
|
||||
SpawnRate: 1,
|
||||
MaxRPS: -1,
|
||||
LoopCount: -1,
|
||||
RequestIncreaseRate: "-1",
|
||||
CPUProfileDuration: 30 * time.Second,
|
||||
MemoryProfileDuration: 30 * time.Second,
|
||||
}
|
||||
}
|
||||
|
||||
func (b *Boomer) GetProfile() *Profile {
|
||||
switch b.mode {
|
||||
case DistributedMasterMode:
|
||||
@@ -158,7 +170,18 @@ func (b *Boomer) RunWorker() {
|
||||
|
||||
// TestCaseBytesChan gets test case bytes chan
|
||||
func (b *Boomer) TestCaseBytesChan() chan []byte {
|
||||
return b.masterRunner.testCaseBytes
|
||||
return b.masterRunner.testCaseBytesChan
|
||||
}
|
||||
|
||||
func (b *Boomer) GetTestCaseBytes() []byte {
|
||||
switch b.mode {
|
||||
case DistributedMasterMode:
|
||||
return b.masterRunner.testCasesBytes
|
||||
case DistributedWorkerMode:
|
||||
return b.workerRunner.testCasesBytes
|
||||
default:
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
func ProfileToBytes(profile *Profile) []byte {
|
||||
@@ -192,7 +215,7 @@ func (b *Boomer) GetTasksChan() chan *task {
|
||||
func (b *Boomer) GetRebalanceChan() chan bool {
|
||||
switch b.mode {
|
||||
case DistributedWorkerMode:
|
||||
return b.workerRunner.rebalance
|
||||
return b.workerRunner.controller.getRebalanceChan()
|
||||
default:
|
||||
return nil
|
||||
}
|
||||
@@ -469,6 +492,9 @@ func (b *Boomer) Start(Args *Profile) error {
|
||||
if b.masterRunner.isStopping() {
|
||||
return errors.New("Please wait for all workers to finish")
|
||||
}
|
||||
if int(Args.SpawnCount) < b.masterRunner.server.getAvailableClientsLength() {
|
||||
return errors.New("spawn count should be greater than available worker count")
|
||||
}
|
||||
b.SetSpawnCount(Args.SpawnCount)
|
||||
b.SetSpawnRate(Args.SpawnRate)
|
||||
b.SetProfile(Args)
|
||||
@@ -481,6 +507,9 @@ func (b *Boomer) ReBalance(Args *Profile) error {
|
||||
if !b.masterRunner.isStarting() {
|
||||
return errors.New("no start")
|
||||
}
|
||||
if int(Args.SpawnCount) < b.masterRunner.server.getAvailableClientsLength() {
|
||||
return errors.New("spawn count should be greater than available worker count")
|
||||
}
|
||||
b.SetSpawnCount(Args.SpawnCount)
|
||||
b.SetSpawnRate(Args.SpawnRate)
|
||||
b.SetProfile(Args)
|
||||
@@ -505,8 +534,9 @@ func (b *Boomer) GetWorkersInfo() []WorkerNode {
|
||||
func (b *Boomer) GetMasterInfo() map[string]interface{} {
|
||||
masterInfo := make(map[string]interface{})
|
||||
masterInfo["state"] = b.masterRunner.getState()
|
||||
masterInfo["workers"] = b.masterRunner.server.getClientsLength()
|
||||
masterInfo["workers"] = b.masterRunner.server.getAvailableClientsLength()
|
||||
masterInfo["target_users"] = b.masterRunner.getSpawnCount()
|
||||
masterInfo["current_users"] = b.masterRunner.server.getCurrentUsers()
|
||||
return masterInfo
|
||||
}
|
||||
|
||||
|
||||
@@ -313,11 +313,12 @@ func (c *grpcClient) sendMessage(msg *genericMessage) {
|
||||
return
|
||||
}
|
||||
err := c.config.getBiStreamClient().Send(&messager.StreamRequest{Type: msg.Type, Data: msg.Data, NodeID: msg.NodeID})
|
||||
switch err {
|
||||
case nil:
|
||||
if err == nil {
|
||||
atomic.StoreInt32(&c.failCount, 0)
|
||||
default:
|
||||
//log.Error().Err(err).Interface("genericMessage", *msg).Msg("failed to send message")
|
||||
return
|
||||
}
|
||||
//log.Error().Err(err).Interface("genericMessage", *msg).Msg("failed to send message")
|
||||
if msg.Type == "heartbeat" {
|
||||
atomic.AddInt32(&c.failCount, 1)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -19,8 +19,8 @@ type genericMessage struct {
|
||||
}
|
||||
|
||||
type task struct {
|
||||
Profile *Profile `json:"profile,omitempty"`
|
||||
TestCases []byte `json:"testcases,omitempty"`
|
||||
Profile *Profile `json:"profile,omitempty"`
|
||||
TestCasesBytes []byte `json:"testcases,omitempty"`
|
||||
}
|
||||
|
||||
func newGenericMessage(t string, data map[string][]byte, nodeID string) (msg *genericMessage) {
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"os"
|
||||
"sort"
|
||||
"strconv"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"github.com/google/uuid"
|
||||
@@ -454,8 +455,8 @@ var (
|
||||
)
|
||||
|
||||
var (
|
||||
minResponseTimeMap = map[string]float64{}
|
||||
maxResponseTimeMap = map[string]float64{}
|
||||
minResponseTimeMap = sync.Map{}
|
||||
maxResponseTimeMap = sync.Map{}
|
||||
)
|
||||
|
||||
// NewPrometheusPusherOutput returns a PrometheusPusherOutput.
|
||||
@@ -577,19 +578,19 @@ func (o *PrometheusPusherOutput) OnEvent(data map[string]interface{}) {
|
||||
}
|
||||
// every stat in total
|
||||
key := fmt.Sprintf("%v_%v", method, name)
|
||||
if _, ok := minResponseTimeMap[key]; !ok {
|
||||
minResponseTimeMap[key] = float64(stat.MinResponseTime)
|
||||
} else {
|
||||
minResponseTimeMap[key] = math.Min(float64(stat.MinResponseTime), minResponseTimeMap[key])
|
||||
minResponseTime, loaded := minResponseTimeMap.LoadOrStore(key, float64(stat.MinResponseTime))
|
||||
if loaded {
|
||||
minResponseTime = math.Min(minResponseTime.(float64), float64(stat.MinResponseTime))
|
||||
minResponseTimeMap.Store(key, minResponseTime)
|
||||
}
|
||||
gaugeTotalMinResponseTime.WithLabelValues(method, name).Set(minResponseTimeMap[key])
|
||||
gaugeTotalMinResponseTime.WithLabelValues(method, name).Set(minResponseTime.(float64))
|
||||
|
||||
if _, ok := maxResponseTimeMap[key]; !ok {
|
||||
maxResponseTimeMap[key] = float64(stat.MaxResponseTime)
|
||||
} else {
|
||||
maxResponseTimeMap[key] = math.Max(float64(stat.MaxResponseTime), maxResponseTimeMap[key])
|
||||
maxResponseTime, loaded := maxResponseTimeMap.LoadOrStore(key, float64(stat.MaxResponseTime))
|
||||
if loaded {
|
||||
maxResponseTime = math.Max(maxResponseTime.(float64), float64(stat.MaxResponseTime))
|
||||
maxResponseTimeMap.Store(key, maxResponseTime)
|
||||
}
|
||||
gaugeTotalMaxResponseTime.WithLabelValues(method, name).Set(maxResponseTimeMap[key])
|
||||
gaugeTotalMaxResponseTime.WithLabelValues(method, name).Set(maxResponseTime.(float64))
|
||||
|
||||
counterTotalNumRequests.WithLabelValues(method, name).Add(float64(stat.NumRequests))
|
||||
counterTotalNumFailures.WithLabelValues(method, name).Add(float64(stat.NumFailures))
|
||||
@@ -639,4 +640,7 @@ func resetPrometheusMetrics() {
|
||||
gaugeTotalFailPerSec.Set(0)
|
||||
gaugeTransactionsPassed.Set(0)
|
||||
gaugeTransactionsFailed.Set(0)
|
||||
|
||||
minResponseTimeMap = sync.Map{}
|
||||
maxResponseTimeMap = sync.Map{}
|
||||
}
|
||||
|
||||
@@ -49,10 +49,10 @@ func getStateName(state int32) (stateName string) {
|
||||
}
|
||||
|
||||
const (
|
||||
reportStatsInterval = 3 * time.Second
|
||||
heartbeatInterval = 1 * time.Second
|
||||
heartbeatLiveness = 3 * time.Second
|
||||
reconnectInterval = 3 * time.Second
|
||||
reportStatsInterval = 3 * time.Second
|
||||
heartbeatInterval = 1 * time.Second
|
||||
heartbeatLiveness = 3 * time.Second
|
||||
stateMachineInterval = 1 * time.Second
|
||||
)
|
||||
|
||||
type Loop struct {
|
||||
@@ -86,6 +86,7 @@ type Controller struct {
|
||||
currentClientsNum int64 // current clients count
|
||||
spawnCount int64 // target clients to spawn
|
||||
spawnRate float64
|
||||
rebalance chan bool // dynamically balance boomer running parameters
|
||||
spawnDone chan struct{}
|
||||
tasks []*Task
|
||||
}
|
||||
@@ -143,6 +144,12 @@ func (c *Controller) spawnCompete() {
|
||||
close(c.spawnDone)
|
||||
}
|
||||
|
||||
func (c *Controller) getRebalanceChan() chan bool {
|
||||
c.mutex.RLock()
|
||||
defer c.mutex.RUnlock()
|
||||
return c.rebalance
|
||||
}
|
||||
|
||||
func (c *Controller) isFinished() bool {
|
||||
// return true when workers acquired
|
||||
return atomic.LoadInt64(&c.currentClientsNum) == atomic.LoadInt64(&c.spawnCount)
|
||||
@@ -178,6 +185,7 @@ func (c *Controller) reset() {
|
||||
c.spawnRate = 0
|
||||
atomic.StoreInt64(&c.currentClientsNum, 0)
|
||||
c.spawnDone = make(chan struct{})
|
||||
c.rebalance = make(chan bool)
|
||||
c.tasks = []*Task{}
|
||||
c.once = sync.Once{}
|
||||
}
|
||||
@@ -199,9 +207,6 @@ type runner struct {
|
||||
controller *Controller
|
||||
loop *Loop // specify loop count for testcase, count = loopCount * spawnCount
|
||||
|
||||
// dynamically balance boomer running parameters
|
||||
rebalance chan bool
|
||||
|
||||
// stop signals the run goroutine should shutdown.
|
||||
stopChan chan bool
|
||||
// all running workers(goroutines) will select on this channel.
|
||||
@@ -358,7 +363,6 @@ func (r *runner) reportTestResult() {
|
||||
func (r *runner) reset() {
|
||||
r.controller.reset()
|
||||
r.stats.clearAll()
|
||||
r.rebalance = make(chan bool)
|
||||
r.stoppingChan = make(chan bool)
|
||||
r.doneChan = make(chan bool)
|
||||
r.reportedChan = make(chan bool)
|
||||
@@ -430,16 +434,18 @@ func (r *runner) spawnWorkers(spawnCount int64, spawnRate float64, quit chan boo
|
||||
continue
|
||||
}
|
||||
|
||||
r.controller.once.Do(func() {
|
||||
// spawning compete
|
||||
r.controller.spawnCompete()
|
||||
if spawnCompleteFunc != nil {
|
||||
spawnCompleteFunc()
|
||||
}
|
||||
r.updateState(StateRunning)
|
||||
})
|
||||
r.controller.once.Do(
|
||||
func() {
|
||||
// spawning compete
|
||||
r.controller.spawnCompete()
|
||||
if spawnCompleteFunc != nil {
|
||||
spawnCompleteFunc()
|
||||
}
|
||||
r.updateState(StateRunning)
|
||||
},
|
||||
)
|
||||
|
||||
<-r.rebalance
|
||||
<-r.controller.getRebalanceChan()
|
||||
if r.isStarting() {
|
||||
// rebalance spawn count
|
||||
r.controller.setSpawn(r.getSpawnCount(), r.getSpawnRate())
|
||||
@@ -629,7 +635,7 @@ func (r *localRunner) start() {
|
||||
r.wgMu.Lock()
|
||||
r.updateState(StateStopping)
|
||||
close(r.stoppingChan)
|
||||
close(r.rebalance)
|
||||
close(r.controller.rebalance)
|
||||
r.wgMu.Unlock()
|
||||
|
||||
// wait for goroutines before closing
|
||||
@@ -668,7 +674,8 @@ type workerRunner struct {
|
||||
masterPort int
|
||||
client *grpcClient
|
||||
|
||||
profile *Profile
|
||||
profile *Profile
|
||||
testCasesBytes []byte
|
||||
|
||||
tasksChan chan *task
|
||||
|
||||
@@ -714,10 +721,10 @@ func (r *workerRunner) onSpawnMessage(msg *genericMessage) {
|
||||
log.Error().Msg("miss tasks")
|
||||
}
|
||||
r.tasksChan <- &task{
|
||||
Profile: profile,
|
||||
TestCases: msg.Tasks,
|
||||
Profile: profile,
|
||||
TestCasesBytes: msg.Tasks,
|
||||
}
|
||||
log.Info().Msg("on spawn message successful")
|
||||
log.Info().Msg("on spawn message successfully")
|
||||
}
|
||||
|
||||
func (r *workerRunner) onRebalanceMessage(msg *genericMessage) {
|
||||
@@ -731,7 +738,7 @@ func (r *workerRunner) onRebalanceMessage(msg *genericMessage) {
|
||||
r.tasksChan <- &task{
|
||||
Profile: profile,
|
||||
}
|
||||
log.Info().Msg("on rebalance message successful")
|
||||
log.Info().Msg("on rebalance message successfully")
|
||||
}
|
||||
|
||||
// Runner acts as a state machine.
|
||||
@@ -831,6 +838,9 @@ func (r *workerRunner) run() {
|
||||
// wait for goroutines before closing
|
||||
r.wg.Wait()
|
||||
|
||||
// notify master that worker is quitting
|
||||
r.onQuiting()
|
||||
|
||||
var ticker = time.NewTicker(1 * time.Second)
|
||||
if r.client != nil {
|
||||
// waitting for quit message is sent to master
|
||||
@@ -879,6 +889,7 @@ func (r *workerRunner) run() {
|
||||
if !r.isStarting() && !r.isStopping() {
|
||||
r.updateState(StateMissing)
|
||||
}
|
||||
continue
|
||||
}
|
||||
CPUUsage := GetCurrentCPUPercent()
|
||||
MemoryUsage := GetCurrentMemoryPercent()
|
||||
@@ -918,13 +929,18 @@ func (r *workerRunner) start() {
|
||||
// block concurrent waitgroup adds in GoAttach while stopping
|
||||
r.wgMu.Lock()
|
||||
r.updateState(StateStopping)
|
||||
close(r.controller.rebalance)
|
||||
close(r.stoppingChan)
|
||||
close(r.rebalance)
|
||||
r.wgMu.Unlock()
|
||||
|
||||
// wait for goroutines before closing
|
||||
r.wg.Wait()
|
||||
|
||||
// reset loop
|
||||
if r.loop != nil {
|
||||
r.loop = nil
|
||||
}
|
||||
|
||||
close(r.doneChan)
|
||||
|
||||
// wait until all stats are reported successfully
|
||||
@@ -951,7 +967,6 @@ func (r *workerRunner) stop() {
|
||||
}
|
||||
|
||||
func (r *workerRunner) close() {
|
||||
r.onQuiting()
|
||||
close(r.closeChan)
|
||||
}
|
||||
|
||||
@@ -970,8 +985,8 @@ type masterRunner struct {
|
||||
profile *Profile
|
||||
|
||||
parseTestCasesChan chan bool
|
||||
testCaseBytes chan []byte
|
||||
tcb []byte
|
||||
testCaseBytesChan chan []byte
|
||||
testCasesBytes []byte
|
||||
}
|
||||
|
||||
func newMasterRunner(masterBindHost string, masterBindPort int) *masterRunner {
|
||||
@@ -988,7 +1003,7 @@ func newMasterRunner(masterBindHost string, masterBindPort int) *masterRunner {
|
||||
masterBindPort: masterBindPort,
|
||||
server: newServer(masterBindHost, masterBindPort),
|
||||
parseTestCasesChan: make(chan bool),
|
||||
testCaseBytes: make(chan []byte),
|
||||
testCaseBytesChan: make(chan []byte),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1011,23 +1026,15 @@ func (r *masterRunner) heartbeatWorker() {
|
||||
if !ok {
|
||||
log.Error().Msg("failed to get worker information")
|
||||
}
|
||||
if atomic.LoadInt32(&workerInfo.Heartbeat) < 0 {
|
||||
if workerInfo.getState() == StateQuitting {
|
||||
return true
|
||||
}
|
||||
if workerInfo.getState() != StateMissing {
|
||||
workerInfo.setState(StateMissing)
|
||||
}
|
||||
if r.isStopping() {
|
||||
// all running workers missed, setting state to stopped
|
||||
if r.server.getClientsLength() <= 0 {
|
||||
r.updateState(StateStopped)
|
||||
go func() {
|
||||
if atomic.LoadInt32(&workerInfo.Heartbeat) < 0 {
|
||||
if workerInfo.getState() != StateMissing {
|
||||
workerInfo.setState(StateMissing)
|
||||
}
|
||||
return true
|
||||
} else {
|
||||
atomic.AddInt32(&workerInfo.Heartbeat, -1)
|
||||
}
|
||||
} else {
|
||||
atomic.AddInt32(&workerInfo.Heartbeat, -1)
|
||||
}
|
||||
}()
|
||||
return true
|
||||
})
|
||||
case <-reportTicker.C:
|
||||
@@ -1051,72 +1058,85 @@ func (r *masterRunner) clientListener() {
|
||||
if !ok {
|
||||
continue
|
||||
}
|
||||
switch msg.Type {
|
||||
case typeClientReady:
|
||||
workerInfo.setState(StateInit)
|
||||
if r.getState() == StateRunning {
|
||||
log.Warn().Str("worker id", workerInfo.ID).Msg("worker joined, ready to rebalance the load of each worker")
|
||||
go func() {
|
||||
switch msg.Type {
|
||||
case typeClientReady:
|
||||
workerInfo.setState(StateInit)
|
||||
case typeClientStopped:
|
||||
workerInfo.setState(StateStopped)
|
||||
case typeHeartbeat:
|
||||
if workerInfo.getState() == StateMissing {
|
||||
workerInfo.setState(int32(builtin.BytesToInt64(msg.Data["state"])))
|
||||
}
|
||||
workerInfo.updateHeartbeat(3)
|
||||
currentCPUUsage, ok := msg.Data["current_cpu_usage"]
|
||||
if ok {
|
||||
workerInfo.updateCPUUsage(builtin.ByteToFloat64(currentCPUUsage))
|
||||
}
|
||||
currentPidCpuUsage, ok := msg.Data["current_pid_cpu_usage"]
|
||||
if ok {
|
||||
workerInfo.updateWorkerCPUUsage(builtin.ByteToFloat64(currentPidCpuUsage))
|
||||
}
|
||||
currentMemoryUsage, ok := msg.Data["current_memory_usage"]
|
||||
if ok {
|
||||
workerInfo.updateMemoryUsage(builtin.ByteToFloat64(currentMemoryUsage))
|
||||
}
|
||||
currentPidMemoryUsage, ok := msg.Data["current_pid_memory_usage"]
|
||||
if ok {
|
||||
workerInfo.updateWorkerMemoryUsage(builtin.ByteToFloat64(currentPidMemoryUsage))
|
||||
}
|
||||
currentUsers, ok := msg.Data["current_users"]
|
||||
if ok {
|
||||
workerInfo.updateUserCount(builtin.BytesToInt64(currentUsers))
|
||||
}
|
||||
case typeSpawning:
|
||||
workerInfo.setState(StateSpawning)
|
||||
case typeSpawningComplete:
|
||||
workerInfo.setState(StateRunning)
|
||||
case typeQuit:
|
||||
if workerInfo.getState() == StateQuitting {
|
||||
break
|
||||
}
|
||||
workerInfo.setState(StateQuitting)
|
||||
case typeException:
|
||||
// Todo
|
||||
default:
|
||||
}
|
||||
}()
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func (r *masterRunner) stateMachine() {
|
||||
ticker := time.NewTicker(stateMachineInterval)
|
||||
for {
|
||||
select {
|
||||
case <-r.closeChan:
|
||||
return
|
||||
case <-ticker.C:
|
||||
switch r.getState() {
|
||||
case StateSpawning:
|
||||
if r.server.getCurrentUsers() == int(r.getSpawnCount()) {
|
||||
log.Warn().Msg("all workers spawn done, setting state as running")
|
||||
r.updateState(StateRunning)
|
||||
}
|
||||
case StateRunning:
|
||||
if r.server.getStartingClientsLength() == 0 {
|
||||
r.updateState(StateStopped)
|
||||
continue
|
||||
}
|
||||
if r.server.getWorkersLengthByState(StateInit) != 0 {
|
||||
err := r.rebalance()
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to rebalance")
|
||||
}
|
||||
}
|
||||
case typeClientStopped:
|
||||
workerInfo.setState(StateStopped)
|
||||
if r.server.getWorkersLengthByState(StateStopped)+r.server.getWorkersLengthByState(StateInit) == r.server.getClientsLength() {
|
||||
case StateStopping:
|
||||
if r.server.getReadyClientsLength() == r.server.getAvailableClientsLength() {
|
||||
r.updateState(StateStopped)
|
||||
}
|
||||
case typeHeartbeat:
|
||||
if workerInfo.getState() != int32(builtin.BytesToInt64(msg.Data["state"])) {
|
||||
workerInfo.setState(int32(builtin.BytesToInt64(msg.Data["state"])))
|
||||
}
|
||||
workerInfo.updateHeartbeat(3)
|
||||
currentCPUUsage, ok := msg.Data["current_cpu_usage"]
|
||||
if ok {
|
||||
workerInfo.updateCPUUsage(builtin.ByteToFloat64(currentCPUUsage))
|
||||
}
|
||||
currentPidCpuUsage, ok := msg.Data["current_pid_cpu_usage"]
|
||||
if ok {
|
||||
workerInfo.updateWorkerCPUUsage(builtin.ByteToFloat64(currentPidCpuUsage))
|
||||
}
|
||||
currentMemoryUsage, ok := msg.Data["current_memory_usage"]
|
||||
if ok {
|
||||
workerInfo.updateMemoryUsage(builtin.ByteToFloat64(currentMemoryUsage))
|
||||
}
|
||||
currentPidMemoryUsage, ok := msg.Data["current_pid_memory_usage"]
|
||||
if ok {
|
||||
workerInfo.updateWorkerMemoryUsage(builtin.ByteToFloat64(currentPidMemoryUsage))
|
||||
}
|
||||
currentUsers, ok := msg.Data["current_users"]
|
||||
if ok {
|
||||
workerInfo.updateUserCount(builtin.BytesToInt64(currentUsers))
|
||||
}
|
||||
case typeSpawning:
|
||||
workerInfo.setState(StateSpawning)
|
||||
case typeSpawningComplete:
|
||||
workerInfo.setState(StateRunning)
|
||||
if r.server.getWorkersLengthByState(StateRunning) == r.server.getClientsLength() {
|
||||
log.Warn().Msg("all workers spawn done, setting state as running")
|
||||
r.updateState(StateRunning)
|
||||
}
|
||||
case typeQuit:
|
||||
if workerInfo.getState() == StateQuitting {
|
||||
break
|
||||
}
|
||||
workerInfo.setState(StateQuitting)
|
||||
if r.isStarting() {
|
||||
if r.server.getClientsLength() > 0 {
|
||||
log.Warn().Str("worker id", workerInfo.ID).Msg("worker quited, ready to rebalance the load of each worker")
|
||||
err := r.rebalance()
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("failed to rebalance")
|
||||
}
|
||||
}
|
||||
}
|
||||
case typeException:
|
||||
// Todo
|
||||
default:
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1146,7 +1166,7 @@ func (r *masterRunner) run() {
|
||||
case <-r.closeChan:
|
||||
return
|
||||
case <-ticker.C:
|
||||
c := r.server.getClientsLength()
|
||||
c := r.server.getAvailableClientsLength()
|
||||
log.Info().Msg(fmt.Sprintf("expected worker number: %v, current worker count: %v", r.expectWorkers, c))
|
||||
if c >= r.expectWorkers {
|
||||
err = r.start()
|
||||
@@ -1165,6 +1185,9 @@ func (r *masterRunner) run() {
|
||||
}()
|
||||
}
|
||||
|
||||
// master state machine
|
||||
r.goAttach(r.stateMachine)
|
||||
|
||||
// listen and deal message from worker
|
||||
r.goAttach(r.clientListener)
|
||||
|
||||
@@ -1174,13 +1197,13 @@ func (r *masterRunner) run() {
|
||||
}
|
||||
|
||||
func (r *masterRunner) start() error {
|
||||
numWorkers := r.server.getClientsLength()
|
||||
numWorkers := r.server.getAvailableClientsLength()
|
||||
if numWorkers == 0 {
|
||||
return errors.New("current available workers: 0")
|
||||
}
|
||||
|
||||
// fetching testcase
|
||||
testcase, err := r.fetchTestCase()
|
||||
// fetching testcases
|
||||
testCasesBytes, err := r.fetchTestCases()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
@@ -1223,19 +1246,19 @@ func (r *masterRunner) start() error {
|
||||
Type: "spawn",
|
||||
Profile: ProfileToBytes(workerProfile),
|
||||
NodeID: workerInfo.ID,
|
||||
Tasks: testcase,
|
||||
Tasks: testCasesBytes,
|
||||
}
|
||||
cur++
|
||||
}
|
||||
return true
|
||||
})
|
||||
|
||||
log.Warn().Interface("profile", r.profile).Msg("send spawn data to worker successful")
|
||||
log.Warn().Interface("profile", r.profile).Msg("send spawn data to worker successfully")
|
||||
return nil
|
||||
}
|
||||
|
||||
func (r *masterRunner) rebalance() error {
|
||||
numWorkers := r.server.getClientsLength()
|
||||
numWorkers := r.server.getAvailableClientsLength()
|
||||
if numWorkers == 0 {
|
||||
return errors.New("current available workers: 0")
|
||||
}
|
||||
@@ -1276,7 +1299,7 @@ func (r *masterRunner) rebalance() error {
|
||||
Type: "spawn",
|
||||
Profile: ProfileToBytes(workerProfile),
|
||||
NodeID: workerInfo.ID,
|
||||
Tasks: r.tcb,
|
||||
Tasks: r.testCasesBytes,
|
||||
}
|
||||
} else {
|
||||
workerInfo.getStream() <- &messager.StreamResponse{
|
||||
@@ -1290,22 +1313,22 @@ func (r *masterRunner) rebalance() error {
|
||||
return true
|
||||
})
|
||||
|
||||
log.Warn().Msg("send rebalance data to worker successful")
|
||||
log.Warn().Msg("send rebalance data to worker successfully")
|
||||
return nil
|
||||
}
|
||||
|
||||
func (r *masterRunner) fetchTestCase() ([]byte, error) {
|
||||
func (r *masterRunner) fetchTestCases() ([]byte, error) {
|
||||
ticker := time.NewTicker(30 * time.Second)
|
||||
if len(r.testCaseBytes) > 0 {
|
||||
<-r.testCaseBytes
|
||||
if len(r.testCaseBytesChan) > 0 {
|
||||
<-r.testCaseBytesChan
|
||||
}
|
||||
r.parseTestCasesChan <- true
|
||||
select {
|
||||
case <-ticker.C:
|
||||
return nil, errors.New("parse testcases timeout")
|
||||
case tcb := <-r.testCaseBytes:
|
||||
r.tcb = tcb
|
||||
return tcb, nil
|
||||
case testCasesBytes := <-r.testCaseBytesChan:
|
||||
r.testCasesBytes = testCasesBytes
|
||||
return testCasesBytes, nil
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1336,22 +1359,23 @@ func (r *masterRunner) close() {
|
||||
func (r *masterRunner) reportStats() {
|
||||
currentTime := time.Now()
|
||||
println()
|
||||
println("========================= HttpRunner Master for Distributed Load Testing ========================= ")
|
||||
println(fmt.Sprintf("Current time: %s, State: %v, Current Available Workers: %v, Target Users: %v",
|
||||
currentTime.Format("2006/01/02 15:04:05"), getStateName(r.getState()), r.server.getClientsLength(), r.getSpawnCount()))
|
||||
println("==================================== HttpRunner Master for Distributed Load Testing ==================================== ")
|
||||
println(fmt.Sprintf("Current time: %s, State: %v, Current Available Workers: %v, Target Users: %v, Current Users: %v",
|
||||
currentTime.Format("2006/01/02 15:04:05"), getStateName(r.getState()), r.server.getAvailableClientsLength(), r.getSpawnCount(), r.server.getCurrentUsers()))
|
||||
table := tablewriter.NewWriter(os.Stdout)
|
||||
table.SetColMinWidth(0, 20)
|
||||
table.SetColMinWidth(0, 40)
|
||||
table.SetColMinWidth(1, 10)
|
||||
table.SetColMinWidth(2, 10)
|
||||
table.SetHeader([]string{"Worker ID", "IP", "State", "Current Users", "CPU Usage (%)", "Memory Usage (%)"})
|
||||
|
||||
for _, worker := range r.server.getAllWorkers() {
|
||||
row := make([]string, 6)
|
||||
row[0] = worker.ID
|
||||
row[1] = worker.IP
|
||||
row[2] = fmt.Sprintf("%v", getStateName(worker.getState()))
|
||||
row[3] = fmt.Sprintf("%v", worker.getUserCount())
|
||||
row[4] = fmt.Sprintf("%.2f", worker.getCPUUsage())
|
||||
row[5] = fmt.Sprintf("%.2f", worker.getMemoryUsage())
|
||||
row[2] = fmt.Sprintf("%v", getStateName(worker.State))
|
||||
row[3] = fmt.Sprintf("%v", worker.UserCount)
|
||||
row[4] = fmt.Sprintf("%.2f", worker.CPUUsage)
|
||||
row[5] = fmt.Sprintf("%.2f", worker.MemoryUsage)
|
||||
table.Append(row)
|
||||
}
|
||||
table.Render()
|
||||
|
||||
@@ -338,6 +338,7 @@ func TestOnQuitMessage(t *testing.T) {
|
||||
go runner.onMessage(newGenericMessage("quit", nil, runner.nodeID))
|
||||
close(runner.doneChan)
|
||||
<-runner.closeChan
|
||||
runner.onQuiting()
|
||||
if runner.getState() != StateQuitting {
|
||||
t.Error("Runner's state should be StateQuitting")
|
||||
}
|
||||
@@ -348,6 +349,7 @@ func TestOnQuitMessage(t *testing.T) {
|
||||
runner.client.shutdownChan = make(chan bool)
|
||||
runner.onMessage(newGenericMessage("quit", nil, runner.nodeID))
|
||||
<-runner.closeChan
|
||||
runner.onQuiting()
|
||||
if runner.getState() != StateQuitting {
|
||||
t.Error("Runner's state should be StateQuitting")
|
||||
}
|
||||
@@ -395,7 +397,7 @@ func TestOnMessage(t *testing.T) {
|
||||
|
||||
// increase goroutines while running
|
||||
runner.onMessage(newMessageToWorker("rebalance", ProfileToBytes(&Profile{SpawnCount: 15, SpawnRate: 15}), nil, nil))
|
||||
runner.rebalance <- true
|
||||
runner.controller.rebalance <- true
|
||||
|
||||
time.Sleep(2 * time.Second)
|
||||
if runner.getState() != StateRunning {
|
||||
@@ -460,6 +462,7 @@ func TestClientListener(t *testing.T) {
|
||||
runner.updateState(StateInit)
|
||||
runner.setSpawnCount(10)
|
||||
runner.setSpawnRate(10)
|
||||
go runner.stateMachine()
|
||||
go runner.clientListener()
|
||||
runner.server.clients.Store("testID1", &WorkerNode{ID: "testID1", Heartbeat: 3, stream: make(chan *messager.StreamResponse, 10)})
|
||||
runner.server.clients.Store("testID2", &WorkerNode{ID: "testID2", Heartbeat: 3, stream: make(chan *messager.StreamResponse, 10)})
|
||||
@@ -483,6 +486,7 @@ func TestClientListener(t *testing.T) {
|
||||
Type: typeClientStopped,
|
||||
NodeID: "testID2",
|
||||
}
|
||||
runner.updateState(StateRunning)
|
||||
worker2, ok := runner.server.getClients().Load("testID2")
|
||||
if !ok {
|
||||
t.Fatal("error")
|
||||
@@ -515,7 +519,7 @@ func TestHeartbeatWorker(t *testing.T) {
|
||||
runner.server.clients.Store("testID2", &WorkerNode{ID: "testID2", Heartbeat: 1, State: StateInit, stream: make(chan *messager.StreamResponse, 10)})
|
||||
go runner.clientListener()
|
||||
go runner.heartbeatWorker()
|
||||
time.Sleep(3 * time.Second)
|
||||
time.Sleep(4 * time.Second)
|
||||
worker1, ok := runner.server.getClients().Load("testID1")
|
||||
if !ok {
|
||||
t.Fatal()
|
||||
@@ -525,7 +529,7 @@ func TestHeartbeatWorker(t *testing.T) {
|
||||
t.Fatal()
|
||||
}
|
||||
if workerInfo1.getState() != StateMissing {
|
||||
t.Error("expected state of worker runner is missing, but got", workerInfo1.getState())
|
||||
t.Error("expected state of worker runner is missing, but got", getStateName(workerInfo1.getState()))
|
||||
}
|
||||
runner.server.recvChannel() <- &genericMessage{
|
||||
Type: typeHeartbeat,
|
||||
|
||||
@@ -53,6 +53,24 @@ func (w *WorkerNode) setState(state int32) {
|
||||
atomic.StoreInt32(&w.State, state)
|
||||
}
|
||||
|
||||
func (w *WorkerNode) isStarting() bool {
|
||||
return w.getState() == StateRunning || w.getState() == StateSpawning
|
||||
}
|
||||
|
||||
func (w *WorkerNode) isStopping() bool {
|
||||
return w.getState() == StateStopping
|
||||
}
|
||||
|
||||
func (w *WorkerNode) isAvailable() bool {
|
||||
state := w.getState()
|
||||
return state != StateMissing && state != StateQuitting
|
||||
}
|
||||
|
||||
func (w *WorkerNode) isReady() bool {
|
||||
state := w.getState()
|
||||
return state == StateInit || state == StateStopped
|
||||
}
|
||||
|
||||
func (w *WorkerNode) updateHeartbeat(heartbeat int32) {
|
||||
atomic.StoreInt32(&w.Heartbeat, heartbeat)
|
||||
}
|
||||
@@ -130,8 +148,8 @@ func (w *WorkerNode) getMemoryUsage() float64 {
|
||||
}
|
||||
|
||||
func (w *WorkerNode) setStream(stream chan *messager.StreamResponse) {
|
||||
w.mutex.RLock()
|
||||
defer w.mutex.RUnlock()
|
||||
w.mutex.Lock()
|
||||
defer w.mutex.Unlock()
|
||||
w.stream = stream
|
||||
}
|
||||
|
||||
@@ -302,26 +320,19 @@ func (s *grpcServer) Register(ctx context.Context, req *messager.RegisterRequest
|
||||
wn := newWorkerNode(req.NodeID, clientIp, req.Os, req.Arch)
|
||||
s.clients.Store(req.NodeID, wn)
|
||||
log.Warn().Str("worker id", req.NodeID).Msg("worker joined")
|
||||
return &messager.RegisterResponse{Code: "0", Message: "register successfully"}, nil
|
||||
return &messager.RegisterResponse{Code: "0", Message: "register successful"}, nil
|
||||
}
|
||||
|
||||
func (s *grpcServer) SignOut(_ context.Context, req *messager.SignOutRequest) (*messager.SignOutResponse, error) {
|
||||
// delete worker information
|
||||
s.clients.Delete(req.NodeID)
|
||||
log.Warn().Str("worker id", req.NodeID).Msg("worker quited")
|
||||
return &messager.SignOutResponse{Code: "0", Message: "sign out successfully"}, nil
|
||||
return &messager.SignOutResponse{Code: "0", Message: "sign out successful"}, nil
|
||||
}
|
||||
|
||||
func (s *grpcServer) valid(token string) (isValid bool) {
|
||||
s.clients.Range(func(key, value interface{}) bool {
|
||||
if workerInfo, ok := value.(*WorkerNode); ok {
|
||||
if workerInfo.ID == token {
|
||||
isValid = true
|
||||
}
|
||||
}
|
||||
return true
|
||||
})
|
||||
return
|
||||
func (s *grpcServer) validClientToken(token string) bool {
|
||||
_, ok := s.clients.Load(token)
|
||||
return ok
|
||||
}
|
||||
|
||||
func (s *grpcServer) BidirectionalStreamingMessage(srv messager.Message_BidirectionalStreamingMessageServer) error {
|
||||
@@ -332,7 +343,7 @@ func (s *grpcServer) BidirectionalStreamingMessage(srv messager.Message_Bidirect
|
||||
return status.Error(codes.Unauthenticated, "missing token header")
|
||||
}
|
||||
|
||||
ok = s.valid(token)
|
||||
ok = s.validClientToken(token)
|
||||
if !ok {
|
||||
return status.Error(codes.Unauthenticated, "invalid token")
|
||||
}
|
||||
@@ -404,7 +415,7 @@ func (s *grpcServer) sendMsg(srv messager.Message_BidirectionalStreamingMessageS
|
||||
func (s *grpcServer) sendBroadcasts(msg *genericMessage) {
|
||||
s.clients.Range(func(key, value interface{}) bool {
|
||||
if workerInfo, ok := value.(*WorkerNode); ok {
|
||||
if workerInfo.getState() == StateQuitting || workerInfo.getState() == StateMissing {
|
||||
if !workerInfo.isAvailable() {
|
||||
return true
|
||||
}
|
||||
workerInfo.getStream() <- &messager.StreamResponse{
|
||||
@@ -517,10 +528,10 @@ func (s *grpcServer) getClients() *sync.Map {
|
||||
return s.clients
|
||||
}
|
||||
|
||||
func (s *grpcServer) getClientsLength() (l int) {
|
||||
func (s *grpcServer) getAvailableClientsLength() (l int) {
|
||||
s.clients.Range(func(key, value interface{}) bool {
|
||||
if workerInfo, ok := value.(*WorkerNode); ok {
|
||||
if workerInfo.getState() != StateQuitting && workerInfo.getState() != StateMissing {
|
||||
if workerInfo.isAvailable() {
|
||||
l++
|
||||
}
|
||||
}
|
||||
@@ -528,3 +539,39 @@ func (s *grpcServer) getClientsLength() (l int) {
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
func (s *grpcServer) getReadyClientsLength() (l int) {
|
||||
s.clients.Range(func(key, value interface{}) bool {
|
||||
if workerInfo, ok := value.(*WorkerNode); ok {
|
||||
if workerInfo.isReady() {
|
||||
l++
|
||||
}
|
||||
}
|
||||
return true
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
func (s *grpcServer) getStartingClientsLength() (l int) {
|
||||
s.clients.Range(func(key, value interface{}) bool {
|
||||
if workerInfo, ok := value.(*WorkerNode); ok {
|
||||
if workerInfo.isStarting() {
|
||||
l++
|
||||
}
|
||||
}
|
||||
return true
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
func (s *grpcServer) getCurrentUsers() (l int) {
|
||||
s.clients.Range(func(key, value interface{}) bool {
|
||||
if workerInfo, ok := value.(*WorkerNode); ok {
|
||||
if workerInfo.isStarting() {
|
||||
l += int(workerInfo.getUserCount())
|
||||
}
|
||||
}
|
||||
return true
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
72
hrp/internal/dial/curl.go
Normal file
72
hrp/internal/dial/curl.go
Normal file
@@ -0,0 +1,72 @@
|
||||
package dial
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"fmt"
|
||||
"os"
|
||||
"os/exec"
|
||||
"path/filepath"
|
||||
"time"
|
||||
|
||||
"github.com/rs/zerolog/log"
|
||||
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
)
|
||||
|
||||
const (
|
||||
normalResult = "STDOUT"
|
||||
errorResult = "STDERR"
|
||||
failedResult = "FAILED"
|
||||
)
|
||||
|
||||
type CurlResult struct {
|
||||
Result string `json:"result"`
|
||||
ErrorMsg string `json:"errorMsg"`
|
||||
ResultType string `json:"resultType"`
|
||||
}
|
||||
|
||||
func DoCurl(args []string) (err error) {
|
||||
var saveTests bool
|
||||
for i, arg := range args {
|
||||
if arg == "--save-tests" {
|
||||
args = append(args[:i], args[i+1:]...)
|
||||
saveTests = true
|
||||
}
|
||||
}
|
||||
var curlResult CurlResult
|
||||
defer func() {
|
||||
if saveTests {
|
||||
dir, _ := os.Getwd()
|
||||
curlResultName := fmt.Sprintf("curl_result_%v.json", time.Now().Format("20060102150405"))
|
||||
curlResultPath := filepath.Join(dir, curlResultName)
|
||||
err = builtin.Dump2JSON(curlResult, curlResultPath)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("save dns resolution result failed")
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
cmd := exec.Command("curl", args...)
|
||||
var stdout, stderr bytes.Buffer
|
||||
cmd.Stdout = &stdout
|
||||
cmd.Stderr = &stderr
|
||||
|
||||
err = cmd.Run()
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msgf("fail to run curl command")
|
||||
curlResult.ErrorMsg = err.Error()
|
||||
curlResult.Result = stderr.String()
|
||||
curlResult.ResultType = errorResult
|
||||
return
|
||||
}
|
||||
if stdout.String() != "" {
|
||||
fmt.Printf(stdout.String())
|
||||
curlResult.Result = stdout.String()
|
||||
curlResult.ResultType = normalResult
|
||||
} else if stderr.String() != "" {
|
||||
fmt.Printf(stderr.String())
|
||||
curlResult.ErrorMsg = stderr.String()
|
||||
curlResult.ResultType = errorResult
|
||||
}
|
||||
return
|
||||
}
|
||||
251
hrp/internal/dial/dns.go
Normal file
251
hrp/internal/dial/dns.go
Normal file
@@ -0,0 +1,251 @@
|
||||
package dial
|
||||
|
||||
import (
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"io/ioutil"
|
||||
"net"
|
||||
"net/http"
|
||||
"net/url"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"strconv"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/miekg/dns"
|
||||
"github.com/pkg/errors"
|
||||
"github.com/rs/zerolog/log"
|
||||
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
)
|
||||
|
||||
const (
|
||||
httpDnsUrl = "https://dig.bdurl.net/q"
|
||||
googleDnsUrl = "https://dns.google/resolve"
|
||||
)
|
||||
|
||||
const (
|
||||
DnsSourceTypeLocal = iota
|
||||
DnsSourceTypeHttp
|
||||
DnsSourceTypeGoogle
|
||||
)
|
||||
|
||||
const (
|
||||
DnsRecordTypeA = 1
|
||||
DnsRecordTypeAAAA = 28
|
||||
DnsRecordTypeCNAME = 5
|
||||
)
|
||||
|
||||
var dnsHttpClient = &http.Client{
|
||||
Timeout: 5 * time.Minute,
|
||||
}
|
||||
|
||||
type DnsOptions struct {
|
||||
DnsSourceType int
|
||||
DnsRecordType int
|
||||
DnsServer string
|
||||
SaveTests bool
|
||||
}
|
||||
|
||||
type DnsResult struct {
|
||||
DnsList []string `json:"dnsList"`
|
||||
DnsSource int `json:"dnsType"`
|
||||
DnsRecordType int `json:"dnsRecordType"`
|
||||
DnsServer string `json:"dnsServer,omitempty"`
|
||||
Ttl int `json:"ttl"`
|
||||
Suc bool `json:"suc"`
|
||||
ErrMsg string `json:"errMsg"`
|
||||
}
|
||||
|
||||
type googleDnsResp struct {
|
||||
Answer []googleDnsAnswer `json:"Answer"`
|
||||
}
|
||||
|
||||
type httpDnsResp struct {
|
||||
Ips []string `json:"ips"`
|
||||
Ttl int `json:"ttl"`
|
||||
}
|
||||
|
||||
type googleDnsAnswer struct {
|
||||
Name string `json:"name"`
|
||||
Type int `json:"type"`
|
||||
TTL int `json:"TTL"`
|
||||
Data string `json:"data"`
|
||||
}
|
||||
|
||||
func ParseIP(s string) (net.IP, int) {
|
||||
ip := net.ParseIP(s)
|
||||
if ip == nil {
|
||||
return nil, 0
|
||||
}
|
||||
for i := 0; i < len(s); i++ {
|
||||
switch s[i] {
|
||||
case '.':
|
||||
return ip, 4
|
||||
case ':':
|
||||
return ip, 6
|
||||
}
|
||||
}
|
||||
return nil, 0
|
||||
}
|
||||
|
||||
func localDns(src string, dnsRecordType int, dnsServer string) (dnsResult DnsResult, err error) {
|
||||
dnsResult.DnsSource = DnsSourceTypeLocal
|
||||
dnsResult.DnsRecordType = dnsRecordType
|
||||
|
||||
if dnsServer == "" {
|
||||
config, _ := dns.ClientConfigFromFile("/etc/resolv.conf")
|
||||
dnsServer = config.Servers[0]
|
||||
} else {
|
||||
dnsResult.DnsServer = dnsServer
|
||||
}
|
||||
|
||||
_, ipType := ParseIP(dnsServer)
|
||||
if ipType == 4 {
|
||||
dnsServer += ":53"
|
||||
}
|
||||
|
||||
c := dns.Client{
|
||||
Timeout: 5 * time.Second,
|
||||
}
|
||||
m := dns.Msg{}
|
||||
|
||||
m.SetQuestion(src+".", uint16(dnsRecordType))
|
||||
r, _, err := c.Exchange(&m, dnsServer)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
for _, ans := range r.Answer {
|
||||
switch dnsRecordType {
|
||||
case DnsRecordTypeA:
|
||||
record, isType := ans.(*dns.A)
|
||||
if isType {
|
||||
dnsResult.Ttl = int(record.Hdr.Ttl)
|
||||
dnsResult.DnsList = append(dnsResult.DnsList, record.A.String())
|
||||
}
|
||||
case DnsRecordTypeAAAA:
|
||||
record, isType := ans.(*dns.AAAA)
|
||||
if isType {
|
||||
dnsResult.Ttl = int(record.Hdr.Ttl)
|
||||
dnsResult.DnsList = append(dnsResult.DnsList, record.AAAA.String())
|
||||
}
|
||||
case DnsRecordTypeCNAME:
|
||||
record, isType := ans.(*dns.CNAME)
|
||||
if isType {
|
||||
dnsResult.Ttl = int(record.Hdr.Ttl)
|
||||
dnsResult.DnsList = append(dnsResult.DnsList, record.Target)
|
||||
}
|
||||
}
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func httpDns(url string, dnsRecordType int) (dnsResult DnsResult, err error) {
|
||||
target := httpDnsUrl + "?host=" + url
|
||||
if dnsRecordType == DnsRecordTypeAAAA {
|
||||
target += "&aid=13&f=2"
|
||||
}
|
||||
resp, err := dnsHttpClient.Get(target)
|
||||
|
||||
dnsResult.DnsSource = DnsSourceTypeHttp
|
||||
dnsResult.DnsRecordType = dnsRecordType
|
||||
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
var buf []byte
|
||||
buf, err = ioutil.ReadAll(resp.Body)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
var result httpDnsResp
|
||||
err = json.Unmarshal(buf, &result)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
dnsResult.DnsList = result.Ips
|
||||
dnsResult.Ttl = result.Ttl
|
||||
return
|
||||
}
|
||||
|
||||
func googleDns(url string, dnsRecordType int) (dnsResult DnsResult, err error) {
|
||||
resp, err := dnsHttpClient.Get(googleDnsUrl + "?name=" + url + "&type=" + strconv.Itoa(dnsRecordType))
|
||||
|
||||
dnsResult.DnsSource = DnsSourceTypeGoogle
|
||||
dnsResult.DnsRecordType = dnsRecordType
|
||||
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
var buf []byte
|
||||
buf, err = ioutil.ReadAll(resp.Body)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
var result googleDnsResp
|
||||
err = json.Unmarshal(buf, &result)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
if len(result.Answer) == 0 {
|
||||
return
|
||||
}
|
||||
for _, answer := range result.Answer {
|
||||
if answer.Type == dnsRecordType {
|
||||
dnsResult.Ttl = answer.TTL
|
||||
dnsResult.DnsList = append(dnsResult.DnsList, answer.Data)
|
||||
}
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func DoDns(dnsOptions *DnsOptions, args []string) (err error) {
|
||||
if len(args) != 1 {
|
||||
return errors.New("there should be one argument")
|
||||
}
|
||||
|
||||
var dnsResult DnsResult
|
||||
defer func() {
|
||||
if dnsOptions.SaveTests {
|
||||
dir, _ := os.Getwd()
|
||||
dnsResultName := fmt.Sprintf("dns_result_%v.json", time.Now().Format("20060102150405"))
|
||||
dnsResultPath := filepath.Join(dir, dnsResultName)
|
||||
err = builtin.Dump2JSON(dnsResult, dnsResultPath)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("save dns resolution result failed")
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
dnsTarget := args[0]
|
||||
|
||||
parsedURL, err := url.Parse(dnsTarget)
|
||||
if err == nil && parsedURL.Host != "" {
|
||||
log.Info().Msgf("parse input url %v and extract host %v", dnsTarget, parsedURL.Host)
|
||||
dnsTarget = strings.Split(parsedURL.Host, ":")[0]
|
||||
}
|
||||
log.Info().Msgf("resolve DNS for %v", dnsTarget)
|
||||
dnsRecordType := dnsOptions.DnsRecordType
|
||||
dnsServer := dnsOptions.DnsServer
|
||||
switch dnsOptions.DnsSourceType {
|
||||
case DnsSourceTypeLocal:
|
||||
dnsResult, err = localDns(dnsTarget, dnsRecordType, dnsServer)
|
||||
case DnsSourceTypeHttp:
|
||||
dnsResult, err = httpDns(dnsTarget, dnsRecordType)
|
||||
case DnsSourceTypeGoogle:
|
||||
dnsResult, err = googleDns(dnsTarget, dnsRecordType)
|
||||
}
|
||||
if err != nil {
|
||||
dnsResult.Suc = false
|
||||
dnsResult.ErrMsg = err.Error()
|
||||
log.Error().Err(err).Msgf("fail to do DNS for %s", dnsTarget)
|
||||
} else {
|
||||
dnsResult.Suc = true
|
||||
dnsResult.ErrMsg = ""
|
||||
fmt.Printf("\nDNS resolution done, result IP list: %v\n", dnsResult.DnsList)
|
||||
}
|
||||
return
|
||||
}
|
||||
116
hrp/internal/dial/ping.go
Normal file
116
hrp/internal/dial/ping.go
Normal file
@@ -0,0 +1,116 @@
|
||||
package dial
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"net/url"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/go-ping/ping"
|
||||
"github.com/pkg/errors"
|
||||
"github.com/rs/zerolog/log"
|
||||
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
)
|
||||
|
||||
type PingOptions struct {
|
||||
Count int
|
||||
Timeout time.Duration
|
||||
Interval time.Duration
|
||||
SaveTests bool
|
||||
}
|
||||
|
||||
type PingResult struct {
|
||||
Suc bool `json:"suc"`
|
||||
ErrMsg string `json:"errMsg"`
|
||||
Ip string `json:"ip"`
|
||||
AvgCost int `json:"avgCost"`
|
||||
MaxCost int `json:"maxCost"`
|
||||
MinCost int `json:"minCost"`
|
||||
Lost int `json:"lost"`
|
||||
PingCount int `json:"pingCount"`
|
||||
PacketSize int `json:"packetSize"`
|
||||
ReceivePacketCount int `json:"receivePacketCount"`
|
||||
SendPacketCount int `json:"sendPacketCount"`
|
||||
SuccessCount int `json:"successCount"`
|
||||
DebugLog string `json:"debugLog"`
|
||||
}
|
||||
|
||||
func DoPing(pingOptions *PingOptions, args []string) (err error) {
|
||||
if len(args) != 1 {
|
||||
return errors.New("there should be one argument")
|
||||
}
|
||||
|
||||
var pingResult PingResult
|
||||
defer func() {
|
||||
if pingOptions.SaveTests {
|
||||
dir, _ := os.Getwd()
|
||||
pingResultName := fmt.Sprintf("ping_result_%v.json", time.Now().Format("20060102150405"))
|
||||
pingResultPath := filepath.Join(dir, pingResultName)
|
||||
err = builtin.Dump2JSON(pingResult, pingResultPath)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("save ping result failed")
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
pingTarget := args[0]
|
||||
|
||||
parsedURL, err := url.Parse(pingTarget)
|
||||
if err == nil && parsedURL.Host != "" {
|
||||
log.Info().Msgf("parse input url %v and extract host %v", pingTarget, parsedURL.Host)
|
||||
pingTarget = strings.Split(parsedURL.Host, ":")[0]
|
||||
}
|
||||
|
||||
log.Info().Msgf("ping host %v", pingTarget)
|
||||
pinger, err := ping.NewPinger(pingTarget)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msgf("fail to get pinger for %s", pingTarget)
|
||||
pingResult.Suc = false
|
||||
pingResult.ErrMsg = err.Error()
|
||||
pingResult.DebugLog = err.Error()
|
||||
return
|
||||
}
|
||||
pinger.Count = pingOptions.Count
|
||||
pinger.Timeout = pingOptions.Timeout
|
||||
pinger.Interval = pingOptions.Interval
|
||||
|
||||
pinger.OnRecv = func(pkt *ping.Packet) {
|
||||
pingResult.DebugLog += fmt.Sprintf("%d bytes from %s: icmp_seq=%d time=%v\n",
|
||||
pkt.Nbytes, pkt.IPAddr, pkt.Seq, pkt.Rtt)
|
||||
}
|
||||
pinger.OnFinish = func(stats *ping.Statistics) {
|
||||
pingResult.DebugLog += fmt.Sprintf("\n--- %s ping statistics ---\n", stats.Addr)
|
||||
pingResult.DebugLog += fmt.Sprintf("%d packets transmitted, %d packets received, %v%% packet loss\n",
|
||||
stats.PacketsSent, stats.PacketsRecv, stats.PacketLoss)
|
||||
pingResult.DebugLog += fmt.Sprintf("round-trip min/avg/max/stddev = %v/%v/%v/%v\n",
|
||||
stats.MinRtt, stats.AvgRtt, stats.MaxRtt, stats.StdDevRtt)
|
||||
}
|
||||
pingResult.DebugLog += fmt.Sprintf("PING %s (%s):\n", pinger.Addr(), pinger.IPAddr())
|
||||
|
||||
err = pinger.Run() // blocks until finished
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msgf("fail to run ping for %s", parsedURL)
|
||||
pingResult.Suc = false
|
||||
pingResult.ErrMsg = err.Error()
|
||||
pingResult.DebugLog = err.Error()
|
||||
return
|
||||
}
|
||||
fmt.Print(pingResult.DebugLog)
|
||||
stats := pinger.Statistics() // get send/receive/rtt stats
|
||||
pingResult.Ip = pinger.IPAddr().String()
|
||||
pingResult.AvgCost = int(stats.AvgRtt / time.Millisecond)
|
||||
pingResult.MaxCost = int(stats.MaxRtt / time.Millisecond)
|
||||
pingResult.MinCost = int(stats.MinRtt / time.Millisecond)
|
||||
pingResult.Lost = int(stats.PacketLoss)
|
||||
pingResult.PingCount = pingOptions.Count
|
||||
pingResult.PacketSize = pinger.Size
|
||||
pingResult.ReceivePacketCount = stats.PacketsRecv
|
||||
pingResult.SendPacketCount = stats.PacketsSent
|
||||
pingResult.SuccessCount = stats.PacketsRecv
|
||||
pingResult.Suc = true
|
||||
pingResult.ErrMsg = ""
|
||||
return
|
||||
}
|
||||
20
hrp/internal/dial/traceroute.go
Normal file
20
hrp/internal/dial/traceroute.go
Normal file
@@ -0,0 +1,20 @@
|
||||
package dial
|
||||
|
||||
type TraceRouteOptions struct {
|
||||
MaxTTL int
|
||||
Queries int
|
||||
SaveTests bool
|
||||
}
|
||||
|
||||
type TraceRouteResult struct {
|
||||
IP string `json:"ip"`
|
||||
Details []TraceRouteResultNode `json:"details"`
|
||||
Suc bool `json:"suc"`
|
||||
ErrMsg string `json:"errMsg"`
|
||||
}
|
||||
|
||||
type TraceRouteResultNode struct {
|
||||
Id int `json:"id"`
|
||||
Ip string `json:"ip"`
|
||||
Time string `json:"time"`
|
||||
}
|
||||
106
hrp/internal/dial/traceroute_unix.go
Normal file
106
hrp/internal/dial/traceroute_unix.go
Normal file
@@ -0,0 +1,106 @@
|
||||
//go:build darwin || linux
|
||||
// +build darwin linux
|
||||
|
||||
package dial
|
||||
|
||||
import (
|
||||
"bufio"
|
||||
"fmt"
|
||||
"net/url"
|
||||
"os"
|
||||
"os/exec"
|
||||
"path/filepath"
|
||||
"regexp"
|
||||
"strconv"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/pkg/errors"
|
||||
"github.com/rs/zerolog/log"
|
||||
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
)
|
||||
|
||||
var (
|
||||
regexIPAddr = regexp.MustCompile(`([\d.]+)`)
|
||||
regexElapsedTime = regexp.MustCompile(`(\d+\.\d+)`)
|
||||
regexTraceroutePass = regexp.MustCompile(fmt.Sprintf(`(\d+)[\s*]+(\S+)\s+\(%s\)\s+%s\s+ms`, regexIPAddr, regexElapsedTime))
|
||||
regexTracerouteFailure = regexp.MustCompile(`(\d+)[\s*]+$`)
|
||||
)
|
||||
|
||||
func DoTraceRoute(traceRouteOptions *TraceRouteOptions, args []string) (err error) {
|
||||
if len(args) != 1 {
|
||||
return errors.New("there should be one argument")
|
||||
}
|
||||
var traceRouteResult TraceRouteResult
|
||||
defer func() {
|
||||
if traceRouteOptions.SaveTests {
|
||||
dir, _ := os.Getwd()
|
||||
traceRouteResultName := fmt.Sprintf("traceroute_result_%v.json", time.Now().Format("20060102150405"))
|
||||
traceRouteResultPath := filepath.Join(dir, traceRouteResultName)
|
||||
err = builtin.Dump2JSON(traceRouteResult, traceRouteResultPath)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("save traceroute result failed")
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
traceRouteTarget := args[0]
|
||||
parsedURL, err := url.Parse(traceRouteTarget)
|
||||
if err == nil && parsedURL.Host != "" {
|
||||
log.Info().Msgf("parse input url %v and extract host %v", traceRouteTarget, parsedURL.Host)
|
||||
traceRouteTarget = strings.Split(parsedURL.Host, ":")[0]
|
||||
}
|
||||
|
||||
cmd := exec.Command("traceroute", "-m", strconv.Itoa(traceRouteOptions.MaxTTL),
|
||||
"-q", strconv.Itoa(traceRouteOptions.Queries), traceRouteTarget)
|
||||
stdout, _ := cmd.StdoutPipe()
|
||||
|
||||
startT := time.Now()
|
||||
defer func() {
|
||||
log.Info().Msgf("for target %s, traceroute costs %v", traceRouteTarget, time.Since(startT))
|
||||
}()
|
||||
|
||||
log.Info().Msgf("start to traceroute %v", traceRouteTarget)
|
||||
err = cmd.Start()
|
||||
if err != nil {
|
||||
traceRouteResult.Suc = false
|
||||
traceRouteResult.ErrMsg = "execute traceroute failed"
|
||||
log.Error().Err(err).Msg("start command failed")
|
||||
return
|
||||
}
|
||||
|
||||
scanner := bufio.NewScanner(stdout)
|
||||
for scanner.Scan() {
|
||||
hopLine := scanner.Text()
|
||||
fmt.Println(hopLine)
|
||||
failureLine := regexTracerouteFailure.FindStringSubmatch(hopLine)
|
||||
if len(failureLine) == 2 {
|
||||
hopID, _ := strconv.Atoi(failureLine[1])
|
||||
traceRouteResult.Details = append(traceRouteResult.Details, TraceRouteResultNode{
|
||||
Id: hopID,
|
||||
})
|
||||
continue
|
||||
}
|
||||
passLine := regexTraceroutePass.FindStringSubmatch(hopLine)
|
||||
if len(passLine) == 5 {
|
||||
hopID, _ := strconv.Atoi(passLine[1])
|
||||
traceRouteResult.Details = append(traceRouteResult.Details, TraceRouteResultNode{
|
||||
Id: hopID,
|
||||
Ip: passLine[3],
|
||||
Time: passLine[4],
|
||||
})
|
||||
traceRouteResult.Suc = true
|
||||
}
|
||||
}
|
||||
hopCount := len(traceRouteResult.Details)
|
||||
traceRouteResult.IP = traceRouteResult.Details[hopCount-1].Ip
|
||||
err = cmd.Wait()
|
||||
if err != nil {
|
||||
traceRouteResult.Suc = false
|
||||
traceRouteResult.ErrMsg = "wait traceroute finish failed"
|
||||
log.Error().Err(err).Msg("wait command failed")
|
||||
return
|
||||
}
|
||||
return
|
||||
}
|
||||
105
hrp/internal/dial/traceroute_windows.go
Normal file
105
hrp/internal/dial/traceroute_windows.go
Normal file
@@ -0,0 +1,105 @@
|
||||
//go:build windows
|
||||
// +build windows
|
||||
|
||||
package dial
|
||||
|
||||
import (
|
||||
"bufio"
|
||||
"fmt"
|
||||
"net/url"
|
||||
"os"
|
||||
"os/exec"
|
||||
"path/filepath"
|
||||
"regexp"
|
||||
"strconv"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/pkg/errors"
|
||||
"github.com/rs/zerolog/log"
|
||||
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
)
|
||||
|
||||
var (
|
||||
regexTracertPass = regexp.MustCompile(`(\d+)[\s*<]+(\d+)\s+ms`)
|
||||
regexTracertFailure = regexp.MustCompile(`(\d+)[\s*]+Request timed out`)
|
||||
)
|
||||
|
||||
func DoTraceRoute(traceRouteOptions *TraceRouteOptions, args []string) (err error) {
|
||||
if len(args) != 1 {
|
||||
return errors.New("there should be one argument")
|
||||
}
|
||||
var traceRouteResult TraceRouteResult
|
||||
defer func() {
|
||||
if traceRouteOptions.SaveTests {
|
||||
dir, _ := os.Getwd()
|
||||
traceRouteResultName := fmt.Sprintf("traceroute_result_%v.json", time.Now().Format("20060102150405"))
|
||||
traceRouteResultPath := filepath.Join(dir, traceRouteResultName)
|
||||
err = builtin.Dump2JSON(traceRouteResult, traceRouteResultPath)
|
||||
if err != nil {
|
||||
log.Error().Err(err).Msg("save traceroute result failed")
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
traceRouteTarget := args[0]
|
||||
parsedURL, err := url.Parse(traceRouteTarget)
|
||||
if err == nil && parsedURL.Host != "" {
|
||||
log.Info().Msgf("parse input url %v and extract host %v", traceRouteTarget, parsedURL.Host)
|
||||
traceRouteTarget = strings.Split(parsedURL.Host, ":")[0]
|
||||
}
|
||||
|
||||
cmd := exec.Command("tracert", "-h", strconv.Itoa(traceRouteOptions.MaxTTL), traceRouteTarget)
|
||||
stdout, _ := cmd.StdoutPipe()
|
||||
|
||||
startT := time.Now()
|
||||
defer func() {
|
||||
log.Info().Msgf("for target %s, traceroute costs %v", traceRouteTarget, time.Since(startT))
|
||||
}()
|
||||
|
||||
log.Info().Msgf("start to traceroute %v", traceRouteTarget)
|
||||
err = cmd.Start()
|
||||
if err != nil {
|
||||
traceRouteResult.Suc = false
|
||||
traceRouteResult.ErrMsg = "execute traceroute failed"
|
||||
log.Error().Err(err).Msg("start command failed")
|
||||
return
|
||||
}
|
||||
|
||||
scanner := bufio.NewScanner(stdout)
|
||||
for scanner.Scan() {
|
||||
hopLine := scanner.Text()
|
||||
fmt.Println(hopLine)
|
||||
failureLine := regexTracertFailure.FindStringSubmatch(hopLine)
|
||||
if len(failureLine) == 2 {
|
||||
hopID, _ := strconv.Atoi(failureLine[1])
|
||||
traceRouteResult.Details = append(traceRouteResult.Details, TraceRouteResultNode{
|
||||
Id: hopID,
|
||||
})
|
||||
continue
|
||||
}
|
||||
passLine := regexTracertPass.FindStringSubmatch(hopLine)
|
||||
if len(passLine) == 3 {
|
||||
hopID, _ := strconv.Atoi(passLine[1])
|
||||
fields := strings.Fields(hopLine)
|
||||
hopIP := strings.Trim(fields[len(fields)-1], "[]")
|
||||
traceRouteResult.Details = append(traceRouteResult.Details, TraceRouteResultNode{
|
||||
Id: hopID,
|
||||
Ip: hopIP,
|
||||
Time: passLine[2],
|
||||
})
|
||||
traceRouteResult.Suc = true
|
||||
}
|
||||
}
|
||||
hopCount := len(traceRouteResult.Details)
|
||||
traceRouteResult.IP = traceRouteResult.Details[hopCount-1].Ip
|
||||
err = cmd.Wait()
|
||||
if err != nil {
|
||||
traceRouteResult.Suc = false
|
||||
traceRouteResult.ErrMsg = "wait traceroute finish failed"
|
||||
log.Error().Err(err).Msg("wait command failed")
|
||||
return
|
||||
}
|
||||
return
|
||||
}
|
||||
@@ -5,6 +5,7 @@ import (
|
||||
"os"
|
||||
"path/filepath"
|
||||
"strings"
|
||||
"sync"
|
||||
|
||||
"github.com/httprunner/funplugin"
|
||||
"github.com/httprunner/funplugin/fungo"
|
||||
@@ -24,7 +25,7 @@ const (
|
||||
|
||||
const projectInfoFile = "proj.json" // used for ensuring root project
|
||||
|
||||
var pluginMap = map[string]funplugin.IPlugin{} // used for reusing plugin instance
|
||||
var pluginMap = sync.Map{} // used for reusing plugin instance
|
||||
|
||||
func initPlugin(path, venv string, logOn bool) (plugin funplugin.IPlugin, err error) {
|
||||
// plugin file not found
|
||||
@@ -37,8 +38,8 @@ func initPlugin(path, venv string, logOn bool) (plugin funplugin.IPlugin, err er
|
||||
}
|
||||
|
||||
// reuse plugin instance if it already initialized
|
||||
if p, ok := pluginMap[pluginPath]; ok {
|
||||
return p, nil
|
||||
if p, ok := pluginMap.Load(pluginPath); ok {
|
||||
return p.(funplugin.IPlugin), nil
|
||||
}
|
||||
|
||||
pluginOptions := []funplugin.Option{funplugin.WithLogOn(logOn)}
|
||||
@@ -74,7 +75,7 @@ func initPlugin(path, venv string, logOn bool) (plugin funplugin.IPlugin, err er
|
||||
}
|
||||
|
||||
// add plugin instance to plugin map
|
||||
pluginMap[pluginPath] = plugin
|
||||
pluginMap.Store(pluginPath, plugin)
|
||||
|
||||
// report event for initializing plugin
|
||||
event := sdk.EventTracking{
|
||||
|
||||
@@ -17,6 +17,7 @@ import (
|
||||
"github.com/rs/zerolog/log"
|
||||
"golang.org/x/net/http2"
|
||||
|
||||
"github.com/httprunner/funplugin"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/builtin"
|
||||
"github.com/httprunner/httprunner/v4/hrp/internal/sdk"
|
||||
)
|
||||
@@ -190,11 +191,12 @@ func (r *HRPRunner) Run(testcases ...ITestCase) error {
|
||||
|
||||
// quit all plugins
|
||||
defer func() {
|
||||
if len(pluginMap) > 0 {
|
||||
for _, plugin := range pluginMap {
|
||||
pluginMap.Range(func(key, value interface{}) bool {
|
||||
if plugin, ok := value.(funplugin.IPlugin); ok {
|
||||
plugin.Quit()
|
||||
}
|
||||
}
|
||||
return true
|
||||
})
|
||||
}()
|
||||
|
||||
var runErr error
|
||||
@@ -287,15 +289,17 @@ func (r *HRPRunner) newCaseRunner(testcase *TestCase) (*testCaseRunner, error) {
|
||||
// load plugin info to testcase config
|
||||
if plugin != nil {
|
||||
pluginPath, _ := locatePlugin(testcase.Config.Path)
|
||||
pluginContent, err := builtin.ReadFile(pluginPath)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
tp := strings.Split(plugin.Path(), ".")
|
||||
runner.parsedConfig.PluginSetting = &PluginConfig{
|
||||
Path: pluginPath,
|
||||
Content: pluginContent,
|
||||
Type: tp[len(tp)-1],
|
||||
if runner.parsedConfig.PluginSetting == nil {
|
||||
pluginContent, err := builtin.ReadFile(pluginPath)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
tp := strings.Split(plugin.Path(), ".")
|
||||
runner.parsedConfig.PluginSetting = &PluginConfig{
|
||||
Path: pluginPath,
|
||||
Content: pluginContent,
|
||||
Type: tp[len(tp)-1],
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -24,7 +24,7 @@ func removeHashicorpGoPlugin() {
|
||||
log.Info().Msg("[teardown] remove hashicorp go plugin")
|
||||
os.Remove(tmpl("debugtalk.bin"))
|
||||
pluginPath, _ := filepath.Abs(tmpl("debugtalk.bin"))
|
||||
delete(pluginMap, pluginPath)
|
||||
pluginMap.Delete(pluginPath)
|
||||
}
|
||||
|
||||
func buildHashicorpPyPlugin() {
|
||||
|
||||
@@ -190,7 +190,7 @@ func (api *apiHandler) Start(w http.ResponseWriter, r *http.Request) {
|
||||
return
|
||||
}
|
||||
req := StartRequestBody{
|
||||
Profile: *api.boomer.GetProfile(),
|
||||
Profile: *boomer.NewProfile(),
|
||||
}
|
||||
err = mapstructure.Decode(data, &req)
|
||||
if err != nil {
|
||||
|
||||
@@ -387,13 +387,16 @@ func runStepRequest(r *SessionRunner, step *TStep) (stepResult *StepResult, err
|
||||
if err != nil {
|
||||
return stepResult, errors.Wrap(err, "do request failed")
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
if resp != nil {
|
||||
defer resp.Body.Close()
|
||||
}
|
||||
|
||||
// decode response body in br/gzip/deflate formats
|
||||
err = decodeResponseBody(resp)
|
||||
if err != nil {
|
||||
return stepResult, errors.Wrap(err, "decode response body failed")
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
|
||||
// log & print response
|
||||
if r.LogOn() {
|
||||
|
||||
@@ -104,10 +104,6 @@ func (tc *TCase) ToTestCase(casePath string) (*TestCase, error) {
|
||||
return nil, errors.New("invalid testcase format, missing teststeps!")
|
||||
}
|
||||
|
||||
err := tc.MakeCompat()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if tc.Config == nil {
|
||||
tc.Config = &TConfig{Name: "please input testcase name"}
|
||||
}
|
||||
@@ -121,6 +117,11 @@ func (tc *TCase) toTestCase() (*TestCase, error) {
|
||||
Config: tc.Config,
|
||||
}
|
||||
|
||||
err := tc.MakeCompat()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
// locate project root dir by plugin path
|
||||
projectRootDir, err := GetProjectRootDirPath(tc.Config.Path)
|
||||
if err != nil {
|
||||
|
||||
@@ -37,6 +37,7 @@ class DBEngine(object):
|
||||
pass
|
||||
|
||||
def _fetch(self, query, size=-1, commit=True):
|
||||
query = query.strip()
|
||||
result = self.session.execute(query)
|
||||
if query.upper()[:6] == "SELECT":
|
||||
if size < 0:
|
||||
@@ -80,5 +81,6 @@ class DBEngine(object):
|
||||
if __name__ == "__main__":
|
||||
# db = DBEngine(f"mysql+pymysql://xxxxx:xxxxx@10.0.0.1:3306/dbname?charset=utf8mb4")
|
||||
db = DBEngine(f"sqlite:////Users/bytedance/HttpRunner/examples/data/sqlite.db")
|
||||
print(db.fetchmany("select* from student", 5))
|
||||
print(db.fetchmany("""
|
||||
select* from student""", 5))
|
||||
print(db.fetchmany("select* from student", 5))
|
||||
|
||||
Reference in New Issue
Block a user