Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions internal/servers/plugin/v3/plugin.go
Original file line number Diff line number Diff line change
Expand Up @@ -210,6 +210,10 @@ func (s *Server) Sync(req *pb.Sync_Request, stream pb.Plugin_SyncServer) error {
}
}

if err := s.Plugin.OnSyncFinish(ctx); err != nil {
return status.Errorf(codes.Internal, "failed to finish sync: %v", err)
}

return syncErr
}

Expand Down
13 changes: 13 additions & 0 deletions plugin/plugin.go
Original file line number Diff line number Diff line change
Expand Up @@ -141,6 +141,19 @@ func (p *Plugin) OnBeforeSend(ctx context.Context, msg message.SyncMessage) (mes
return msg, nil
}

// OnSyncFinisher is an interface that can be implemented by a plugin client to be notified when a sync finishes.
type OnSyncFinisher interface {
OnSyncFinish(context.Context) error
}

// OnSyncFinish gets called after a sync finishes.
func (p *Plugin) OnSyncFinish(ctx context.Context) error {
if v, ok := p.client.(OnSyncFinisher); ok {
return v.OnSyncFinish(ctx)
}
return nil
}

// IsStaticLinkingEnabled whether static linking is to be enabled
func (p *Plugin) IsStaticLinkingEnabled() bool {
return p.staticLinking
Expand Down
19 changes: 14 additions & 5 deletions scheduler/scheduler.go
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ import (
"context"
"errors"
"fmt"
"github.com/apache/arrow/go/v14/arrow"
"runtime/debug"
"sync/atomic"
"time"
Expand Down Expand Up @@ -182,15 +183,23 @@ func (s *Scheduler) Sync(ctx context.Context, client schema.ClientMeta, tables s
}
}()
for resource := range resources {
vector := resource.GetValues()
bldr := array.NewRecordBuilder(memory.DefaultAllocator, resource.Table.ToArrowSchema())
scalar.AppendToRecordBuilder(bldr, vector)
rec := bldr.NewRecord()
res <- &message.SyncInsert{Record: rec}
select {
case res <- &message.SyncInsert{Record: resourceToRecord(resource)}:
case <-ctx.Done():
return ctx.Err()
}
}
return nil
}

func resourceToRecord(resource *schema.Resource) arrow.Record {
vector := resource.GetValues()
bldr := array.NewRecordBuilder(memory.DefaultAllocator, resource.Table.ToArrowSchema())
scalar.AppendToRecordBuilder(bldr, vector)
rec := bldr.NewRecord()
return rec
}

func (s *syncClient) logTablesMetrics(tables schema.Tables, client Client) {
clientName := client.ID()
for _, table := range tables {
Expand Down
5 changes: 4 additions & 1 deletion scheduler/scheduler_dfs.go
Original file line number Diff line number Diff line change
Expand Up @@ -176,7 +176,10 @@ func (s *syncClient) resolveResourcesDfs(ctx context.Context, table *schema.Tabl
atomic.AddUint64(&tableMetrics.Errors, 1)
return
}
resourcesChan <- resolvedResource
select {
case resourcesChan <- resolvedResource:
case <-ctx.Done():
}
}()
}
wg.Wait()
Expand Down
85 changes: 85 additions & 0 deletions scheduler/scheduler_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,9 @@ package scheduler

import (
"context"
"fmt"
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
"testing"

"github.com/apache/arrow/go/v14/arrow"
Expand Down Expand Up @@ -40,6 +43,22 @@ func testColumnResolverPanic(context.Context, schema.ClientMeta, *schema.Resourc
panic("ColumnResolver")
}

func testTableSuccessWithData(data []any) *schema.Table {
return &schema.Table{
Name: "test_table_success",
Resolver: func(_ context.Context, _ schema.ClientMeta, _ *schema.Resource, res chan<- any) error {
res <- data
return nil
},
Columns: []schema.Column{
{
Name: "test_column",
Type: arrow.PrimitiveTypes.Int64,
},
},
}
}

func testTableSuccess() *schema.Table {
return &schema.Table{
Name: "test_table_success",
Expand Down Expand Up @@ -233,6 +252,72 @@ func TestScheduler(t *testing.T) {
}
}

func TestScheduler_Cancellation(t *testing.T) {
data := make([]any, 100)

tests := []struct {
name string
data []any
cancel bool
messageCount int
}{
{
name: "should consume all message",
data: data,
cancel: false,
messageCount: len(data) + 1, // 9 data + 1 migration message
},
{
name: "should not consume all message on cancel",
data: data,
cancel: true,
messageCount: len(data) + 1, // 9 data + 1 migration message
},
}

for _, strategy := range AllStrategies {
for _, tc := range tests {
tc := tc
t.Run(fmt.Sprintf("%s_%s", tc.name, strategy.String()), func(t *testing.T) {
sc := NewScheduler(WithLogger(zerolog.New(zerolog.NewTestWriter(t))), WithStrategy(strategy))

messages := make(chan message.SyncMessage)
ctx, cancel := context.WithCancel(context.Background())
defer cancel()

go func() {
err := sc.Sync(
ctx,
&testExecutionClient{},
[]*schema.Table{testTableSuccessWithData(tc.data)},
messages,
)
if tc.cancel {
assert.Equal(t, err, context.Canceled)
} else {
require.NoError(t, err)
}
close(messages)
}()

messageConsumed := 0
for range messages {
if tc.cancel {
cancel()
}
messageConsumed++
}

if tc.cancel {
assert.NotEqual(t, tc.messageCount, messageConsumed)
} else {
assert.Equal(t, tc.messageCount, messageConsumed)
}
})
}
}
}

func testSyncTable(t *testing.T, tc syncTestCase, strategy Strategy, deterministicCQID bool) {
ctx := context.Background()
tables := []*schema.Table{}
Expand Down