Skip to content

Commit

Permalink
Merge branch 'master' into balancer-set
Browse files Browse the repository at this point in the history
  • Loading branch information
lhy1024 committed Apr 20, 2023
2 parents 8b049dd + 3cc745a commit 65003da
Show file tree
Hide file tree
Showing 8 changed files with 108 additions and 11 deletions.
10 changes: 7 additions & 3 deletions client/resource_group/controller/controller.go
Expand Up @@ -38,6 +38,7 @@ const (
retryInterval = 50 * time.Millisecond
maxNotificationChanLen = 200
needTokensAmplification = 1.1
trickleReserveDuration = 1250 * time.Millisecond
)

type selectType int
Expand Down Expand Up @@ -763,7 +764,10 @@ func (gc *groupCostController) shouldReportConsumption() bool {
failpoint.Inject("acceleratedReportingPeriod", func() {
timeSinceLastRequest = extendedReportingPeriodFactor * defaultTargetPeriod
})
if timeSinceLastRequest >= defaultTargetPeriod {
// Due to `gc.run.lastRequestTime` update operations late in this logic,
// so `timeSinceLastRequest` is less than defaultGroupStateUpdateInterval a little bit, lead to actual report period is greater than defaultTargetPeriod.
// Add defaultGroupStateUpdateInterval/2 as duration buffer to avoid it.
if timeSinceLastRequest+defaultGroupStateUpdateInterval/2 >= defaultTargetPeriod {
if timeSinceLastRequest >= extendedReportingPeriodFactor*defaultTargetPeriod {
return true
}
Expand Down Expand Up @@ -882,9 +886,9 @@ func (gc *groupCostController) modifyTokenCounter(counter *tokenCounter, bucket
deadline := gc.run.now.Add(trickleDuration)
cfg.NewRate = float64(bucket.GetSettings().FillRate) + granted/trickleDuration.Seconds()

timerDuration := trickleDuration - time.Second
timerDuration := trickleDuration - trickleReserveDuration
if timerDuration <= 0 {
timerDuration = (trickleDuration + time.Second) / 2
timerDuration = (trickleDuration + trickleReserveDuration) / 2
}
counter.notify.mu.Lock()
counter.notify.setupNotificationTimer = time.NewTimer(timerDuration)
Expand Down
2 changes: 1 addition & 1 deletion client/resource_group/controller/metrics.go
Expand Up @@ -38,7 +38,7 @@ var (
Namespace: namespace,
Subsystem: requestSubsystem,
Name: "success",
Buckets: prometheus.ExponentialBuckets(0.001, 4, 8), // 0.001 ~ 40.96
Buckets: []float64{.005, .01, .05, .1, .5, 1, 5, 10, 20, 25, 30}, // 0.005 ~ 30
Help: "Bucketed histogram of wait duration of successful request.",
}, []string{resourceGroupNameLabel})

Expand Down
25 changes: 20 additions & 5 deletions pkg/keyspace/tso_keyspace_group.go
Expand Up @@ -77,6 +77,8 @@ type GroupManager struct {
// TODO: add user kind with different balancer
// when we ensure where the correspondence between tso node and user kind will be found
nodesBalancer balancer.Balancer[string]
// serviceRegistryMap stores the mapping from the service registry key to the service address.
serviceRegistryMap map[string]string
}

// NewKeyspaceGroupManager creates a Manager of keyspace group related data.
Expand Down Expand Up @@ -117,6 +119,7 @@ func (m *GroupManager) Bootstrap() error {
// If the etcd client is not nil, start the watch loop.
if m.client != nil {
m.nodesBalancer = balancer.GenByPolicy[string](m.policy)
m.serviceRegistryMap = make(map[string]string)
m.wg.Add(1)
go m.startWatchLoop()
}
Expand Down Expand Up @@ -199,6 +202,7 @@ func (m *GroupManager) startWatchLoop() {
continue
}
m.nodesBalancer.Put(s.ServiceAddr)
m.serviceRegistryMap[string(item.Key)] = s.ServiceAddr
}
break
}
Expand Down Expand Up @@ -254,17 +258,28 @@ func (m *GroupManager) watchServiceAddrs(ctx context.Context, revision int64) (i
return revision, wresp.Err()
}
for _, event := range wresp.Events {
s := &discovery.ServiceRegistryEntry{}
if err := json.Unmarshal(event.Kv.Value, s); err != nil {
log.Warn("failed to unmarshal service registry entry", zap.Error(err), zap.ByteString("value", event.Kv.Value))
}
switch event.Type {
case clientv3.EventTypePut:
s := &discovery.ServiceRegistryEntry{}
if err := json.Unmarshal(event.Kv.Value, s); err != nil {
log.Warn("failed to unmarshal service registry entry",
zap.String("event-kv-key", string(event.Kv.Key)), zap.Error(err))
break
}
m.nodesBalancer.Put(s.ServiceAddr)
m.serviceRegistryMap[string(event.Kv.Key)] = s.ServiceAddr
case clientv3.EventTypeDelete:
m.nodesBalancer.Delete(s.ServiceAddr)
key := string(event.Kv.Key)
if serviceAddr, ok := m.serviceRegistryMap[key]; ok {
delete(m.serviceRegistryMap, key)
m.nodesBalancer.Delete(serviceAddr)
} else {
log.Warn("can't retrieve service addr from service registry map",
zap.String("event-kv-key", key))
}
}
}
revision = wresp.Header.Revision + 1
}
}
}
Expand Down
3 changes: 2 additions & 1 deletion pkg/tso/keyspace_group_manager.go
Expand Up @@ -485,6 +485,7 @@ func (kgm *KeyspaceGroupManager) watchKeyspaceGroupsMetaChange(revision int64) (
log.Warn("failed to unmarshal keyspace group",
zap.Uint32("keyspace-group-id", groupID),
zap.Error(errs.ErrJSONUnmarshal.Wrap(err).FastGenWithCause()))
break
}
kgm.updateKeyspaceGroup(group)
case clientv3.EventTypeDelete:
Expand All @@ -499,7 +500,7 @@ func (kgm *KeyspaceGroupManager) watchKeyspaceGroupsMetaChange(revision int64) (
}
}
}
revision = wresp.Header.Revision
revision = wresp.Header.Revision + 1
}

select {
Expand Down
56 changes: 56 additions & 0 deletions server/api/region.go
Expand Up @@ -22,6 +22,7 @@ import (
"net/url"
"sort"
"strconv"
"strings"

"github.com/gorilla/mux"
"github.com/pingcap/failpoint"
Expand Down Expand Up @@ -836,6 +837,61 @@ func (h *regionsHandler) AccelerateRegionsScheduleInRange(w http.ResponseWriter,
h.rd.Text(w, http.StatusOK, fmt.Sprintf("Accelerate regions scheduling in a given range [%s,%s)", rawStartKey, rawEndKey))
}

// @Tags region
// @Summary Accelerate regions scheduling in given ranges, only receive hex format for keys
// @Accept json
// @Param body body object true "json params"
// @Param limit query integer false "Limit count" default(256)
// @Produce json
// @Success 200 {string} string "Accelerate regions scheduling in given ranges [startKey1, endKey1), [startKey2, endKey2), ..."
// @Failure 400 {string} string "The input is invalid."
// @Router /regions/accelerate-schedule/batch [post]
func (h *regionsHandler) AccelerateRegionsScheduleInRanges(w http.ResponseWriter, r *http.Request) {
rc := getCluster(r)
var input []map[string]interface{}
if err := apiutil.ReadJSONRespondError(h.rd, w, r.Body, &input); err != nil {
return
}
limit := 256
if limitStr := r.URL.Query().Get("limit"); limitStr != "" {
var err error
limit, err = strconv.Atoi(limitStr)
if err != nil {
h.rd.JSON(w, http.StatusBadRequest, err.Error())
return
}
}
if limit > maxRegionLimit {
limit = maxRegionLimit
}
var msgBuilder strings.Builder
msgBuilder.Grow(128)
msgBuilder.WriteString("Accelerate regions scheduling in given ranges: ")
var regions []*core.RegionInfo
for _, rg := range input {
startKey, rawStartKey, err := apiutil.ParseKey("start_key", rg)
if err != nil {
h.rd.JSON(w, http.StatusBadRequest, err.Error())
return
}
endKey, rawEndKey, err := apiutil.ParseKey("end_key", rg)
if err != nil {
h.rd.JSON(w, http.StatusBadRequest, err.Error())
return
}
regions = append(regions, rc.ScanRegions(startKey, endKey, limit)...)
msgBuilder.WriteString(fmt.Sprintf("[%s,%s), ", rawStartKey, rawEndKey))
}
if len(regions) > 0 {
regionsIDList := make([]uint64, 0, len(regions))
for _, region := range regions {
regionsIDList = append(regionsIDList, region.GetID())
}
rc.AddSuspectRegions(regionsIDList...)
}
h.rd.Text(w, http.StatusOK, msgBuilder.String())
}

func (h *regionsHandler) GetTopNRegions(w http.ResponseWriter, r *http.Request, less func(a, b *core.RegionInfo) bool) {
rc := getCluster(r)
limit := defaultRegionLimit
Expand Down
20 changes: 20 additions & 0 deletions server/api/region_test.go
Expand Up @@ -332,6 +332,26 @@ func (suite *regionTestSuite) TestAccelerateRegionsScheduleInRange() {
suite.Len(idList, 2)
}

func (suite *regionTestSuite) TestAccelerateRegionsScheduleInRanges() {
re := suite.Require()
r1 := core.NewTestRegionInfo(557, 13, []byte("a1"), []byte("a2"))
r2 := core.NewTestRegionInfo(558, 14, []byte("a2"), []byte("a3"))
r3 := core.NewTestRegionInfo(559, 15, []byte("a3"), []byte("a4"))
r4 := core.NewTestRegionInfo(560, 16, []byte("a4"), []byte("a5"))
r5 := core.NewTestRegionInfo(561, 17, []byte("a5"), []byte("a6"))
mustRegionHeartbeat(re, suite.svr, r1)
mustRegionHeartbeat(re, suite.svr, r2)
mustRegionHeartbeat(re, suite.svr, r3)
mustRegionHeartbeat(re, suite.svr, r4)
mustRegionHeartbeat(re, suite.svr, r5)
body := fmt.Sprintf(`[{"start_key":"%s", "end_key": "%s"}, {"start_key":"%s", "end_key": "%s"}]`, hex.EncodeToString([]byte("a1")), hex.EncodeToString([]byte("a3")), hex.EncodeToString([]byte("a4")), hex.EncodeToString([]byte("a6")))

err := tu.CheckPostJSON(testDialClient, fmt.Sprintf("%s/regions/accelerate-schedule/batch", suite.urlPrefix), []byte(body), tu.StatusOK(re))
suite.NoError(err)
idList := suite.svr.GetRaftCluster().GetSuspectRegions()
suite.Len(idList, 4)
}

func (suite *regionTestSuite) TestScatterRegions() {
re := suite.Require()
r1 := core.NewTestRegionInfo(601, 13, []byte("b1"), []byte("b2"))
Expand Down
1 change: 1 addition & 0 deletions server/api/router.go
Expand Up @@ -267,6 +267,7 @@ func createRouter(prefix string, svr *server.Server) *mux.Router {
registerFunc(clusterRouter, "/regions/check/hist-keys", regionsHandler.GetKeysHistogram, setMethods(http.MethodGet), setAuditBackend(prometheus))
registerFunc(clusterRouter, "/regions/sibling/{id}", regionsHandler.GetRegionSiblings, setMethods(http.MethodGet), setAuditBackend(prometheus))
registerFunc(clusterRouter, "/regions/accelerate-schedule", regionsHandler.AccelerateRegionsScheduleInRange, setMethods(http.MethodPost), setAuditBackend(localLog, prometheus))
registerFunc(clusterRouter, "/regions/accelerate-schedule/batch", regionsHandler.AccelerateRegionsScheduleInRanges, setMethods(http.MethodPost), setAuditBackend(localLog, prometheus))
registerFunc(clusterRouter, "/regions/scatter", regionsHandler.ScatterRegions, setMethods(http.MethodPost), setAuditBackend(localLog, prometheus))
registerFunc(clusterRouter, "/regions/split", regionsHandler.SplitRegions, setMethods(http.MethodPost), setAuditBackend(localLog, prometheus))
registerFunc(clusterRouter, "/regions/range-holes", regionsHandler.GetRangeHoles, setMethods(http.MethodGet), setAuditBackend(prometheus))
Expand Down
2 changes: 1 addition & 1 deletion server/server.go
Expand Up @@ -1820,7 +1820,7 @@ func (s *Server) watchServicePrimaryAddr(ctx context.Context, serviceName string
s.servicePrimaryMap.Delete(serviceName)
}
}
revision = wresp.Header.Revision
revision = wresp.Header.Revision + 1
}
}
}
Expand Down

0 comments on commit 65003da

Please sign in to comment.