Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Roadmap][Volunteer-Wanted] SeaTunnel 2.0 Roadmap Task #720

Open
24 of 54 tasks
CalvinKirs opened this issue Nov 29, 2021 · 27 comments
Open
24 of 54 tasks

[Roadmap][Volunteer-Wanted] SeaTunnel 2.0 Roadmap Task #720

CalvinKirs opened this issue Nov 29, 2021 · 27 comments

Comments

@CalvinKirs
Copy link
Member

CalvinKirs commented Nov 29, 2021

Hello community, we have recently made a 2.0 roadmap, welcome everyone to discuss and supplement.

Update some latest messages:
Hi guys, this is the roadmap draft according to the mindmap, for the following 3 months, please feel free to share your ideas and welcome to contribute to SeaTunnel and join the community.

Roadmap for 12/2021 ~ 03/2022:

@xleoken
Copy link
Member

xleoken commented Nov 29, 2021

The roadmap 2.0 is excellent 👍.

I want to add a new feature to roadmap, it is very important for data integration. We use flume long long ago, but it's not friendly for us to write the agent file. Then we use flinksql replace the flume, it have provided great convenience to insert、etl data, etc.

project weakness
flume 1) need to learn how to write agent file 2) can not work with resource manager like yarn, k8s
sqoop 1) don't have rich connectors 2) based on mapreduce

flinksql

  • has rich connectors, cdc
  • friendly for us, we can use sql grammar easily, like +, -, etc
  • easy to work with resource manager, yarn, k8s

This point of this feature is introduce a new workflow template, show you the workflow demo.

The origin workflow

env {
  execution.parallelism = 1
}

source {
    FakeSourceStream {
      result_table_name = "fake"
      field_name = "name,age"
    }
}

transform {
    sql {
      sql = "select name,age from fake"
    }
}

sink {
  ConsoleSink {}
}

The feature workflow

CREATE TABLE fake_source (
  name string,
  age int
) with (
  'connector.type' = 'fakestream',
  'format.type' = 'json'
);

CREATE TABLE print_sink (
  name string,
  age int
) with (
  'connector.type' = 'print',
);

INSERT INTO print_sink
SELECT * FROM fake_source;

@wntp
Copy link
Contributor

wntp commented Nov 29, 2021

The roadmap 2.0 is excellent 👍.
I wanted After Process and Per Process of source/sink/transform is also in, Let's do it together, come on!

@charlesy6
Copy link

charlesy6 commented Nov 29, 2021

The roadmap 2.0 is excellent 👍.

I have a suggestion, it seems we can focus on one type underline engine. It will be mord hard to maintain flink or spark in further.

@kalencaya
Copy link
Contributor

The roadmap 2.0 is excellent 👍.
Configurable is programmable, why not provide useful config dsl assemble source、transform、sink?

@chenhu
Copy link
Contributor

chenhu commented Nov 30, 2021

The roadmap 2.0 is excellent 👍
The new configuration is easier to accept .

@chenhu
Copy link
Contributor

chenhu commented Nov 30, 2021

We could add the scheduler info like this :
CREATE TABLE fake_source (
name string,
age int
) with (
'connector.type' = 'fakestream',
'format.type' = 'json'
'scheduler.cron' = '* * * * *'
);

CREATE TABLE print_sink (
name string,
age int
) with (
'connector.type' = 'print',
);

INSERT INTO print_sink
SELECT * FROM fake_source;

@CalvinKirs CalvinKirs pinned this issue Dec 7, 2021
@wolfboys
Copy link
Member

I want to contribute to the Core Framework and Plugin Framework part, I am interested in Flink Backend Integration and Plugin Framework, I have relevant implementation experience.

@davidzollo
Copy link
Contributor

I want to contribute to the Core Framework and Plugin Framework part, I am interested in Flink Backend Integration and Plugin Framework, I have relevant implementation experience.

good job, Do you want to implement the entire parts of Flink Backend Integration and Plugin Framework?

@davidzollo
Copy link
Contributor

davidzollo commented Dec 11, 2021

@garyelephant good news. great work

@davidzollo
Copy link
Contributor

Everyone is welcome to share your opinions and suggestions if you have any suggestions about the roadmap, welcome to join the open source community, thx

@wuchunfu
Copy link
Member

I want to contribute to the plugins part, I am interested in plugins.

@simon824
Copy link
Member

I can contribute Project Structure ,Plugins,ability to define variables

@davidzollo

This comment has been minimized.

@wntp

This comment has been minimized.

@davidzollo

This comment has been minimized.

@wolfboys

This comment has been minimized.

@davidzollo

This comment has been minimized.

@davidzollo

This comment has been minimized.

@davidzollo davidzollo changed the title [Roadmap]SeaTunnel 2.0 Roadmap [Roadmap] SeaTunnel 2.0 Roadmap Dec 16, 2021
@davidzollo davidzollo unpinned this issue Dec 17, 2021
@davidzollo davidzollo pinned this issue Dec 17, 2021
@davidzollo davidzollo unpinned this issue Dec 17, 2021
@davidzollo davidzollo pinned this issue Dec 19, 2021
@calvinjiang
Copy link

I'm mainly interested in the Flink Backend Integration. I'd like to contribute more features for this.

@kezhenxu94
Copy link
Member

  • support install and try SeaTunnel by docker container.

This one is finished in #815

@davidzollo
Copy link
Contributor

I'm mainly interested in the Flink Backend Integration. I'd like to contribute more features for this.

good job, please create an related issue first

@CalvinKirs CalvinKirs changed the title [Roadmap] SeaTunnel 2.0 Roadmap [Roadmap][Volunteer-Wanted] SeaTunnel 2.0 Roadmap Task Dec 26, 2021
@apache apache deleted a comment from garyelephant Dec 26, 2021
@xbkaishui

This comment has been minimized.

@CalvinKirs CalvinKirs mentioned this issue Jan 1, 2022
2 tasks
@davidzollo
Copy link
Contributor

if anybody who want to do some contributions, please leave a message.

@yuangjiang
Copy link

the roadmap 2.0
It is recommended to disassemble flink into two different execution modes. From the current test, the flink datastream api and table api cannot be unified. Then we need to support the sql connector that can be submitted to the community. The plug-ins that have been developed by themselves are not two Different modes of execution

@yuangjiang
Copy link

It is recommended to support the submission of spark and flink scripts. The plugin can be used to submit the extended data of spark hive.

@BruceWong96
Copy link

I want to contribute to the Plugin Framework part, I am interested in Flink Backend Integration and Plugin Framework, I have relevant implementation experience.

@Yves-yuan
Copy link
Contributor

The roadmap 2.0 is excellent 👍
I'm interested in [Flink Backend Integration],[Configuration Management Framework] and [plugins] modules,and i will try to make some contributions.

@CalvinKirs CalvinKirs unpinned this issue Jul 12, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests