Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Need a mechanism to compress data to reduce storage costs #126

Closed
laptrinhbockchain opened this issue Oct 13, 2023 · 13 comments
Closed

Need a mechanism to compress data to reduce storage costs #126

laptrinhbockchain opened this issue Oct 13, 2023 · 13 comments
Labels
state: backlog This is pending further review type: enhancement New feature or request

Comments

@laptrinhbockchain
Copy link

Currently the Base node need about additional 100G to store data every day. With data growth so fast, it causes huge obstacles in terms of infrastructure costs.

I think that need a mechanism to compress data to reduce storage costs. This helps many Nodes join the network to increase the decentralization of the project.

Currently, I don't know if this mechanism exists. But is there any way like Node configuration, or some solution to minimize this capacity. Such as:

  1. Compress old data
  2. Eliminate old data (Node only queries recent data)
  3. Move old data to cheaper Storage like HDD
  4. ...
@darren2025
Copy link

I also think it's quite necessary to act

@wbnns wbnns added state: backlog This is pending further review type: enhancement New feature or request labels Oct 16, 2023
@wbnns
Copy link
Member

wbnns commented Oct 16, 2023

@laptrinhbockchain

Thanks for the feedback and for opening this issue. Yes, currently it's not possible to prune the node, which would decrease the required storage space.

Leaving this open and logging this feature request as I'm sure others would also like to have this option should it become available in the future.

@mdehoog
Copy link
Contributor

mdehoog commented Nov 7, 2023

You can run the op-geth with --gcmode=full (see #146) which will greatly reduce disk space requirements, but not allow you to read blocks older than 128 from chain tip.

We don't yet have a snapshot available for full nodes so it will take a while to sync. Will update this issue if and when we have full snapshots.

@HiroyukiNaito
Copy link

@mdehoog I see. Thank you!

@wangtao19911111
Copy link

@mdehoog ,are there any updates here?

@wbnns
Copy link
Member

wbnns commented Feb 16, 2024

@wangtao19911111 Hello, full node snapshots are coming very soon.

@MindlessSteel
Copy link

What about cloud or Ai computation assistance
Computing transactions in the background is the fastest and efficient way to verify everything
Is legit, in my opinion

@kmalloc
Copy link

kmalloc commented Mar 21, 2024

@wangtao19911111 Hello, full node snapshots are coming very soon.

any update? maintaining a full node these day no long an easy task.

@wbnns
Copy link
Member

wbnns commented Mar 22, 2024

@laptrinhbockchain @MindlessSteel @wangtao19911111

Hi there! Just wanted to update that Base Node now supports snap sync by default! Simply:

  1. docker compose stopyour Node
  2. Pull the latest code from https://github.com/base-org/node
  3. Make sure your ENVs have the correct L1 and Beacon endpoints, then
  4. docker compose up -- build.

Whereas the current archive chain data is about 2.5TB, you'll be able to snap sync for less than 250GB.

If you all need any help or support, please let us know!

@wbnns wbnns closed this as completed Mar 22, 2024
@adapt7
Copy link

adapt7 commented Apr 19, 2024

@wbnns hi there. So I had fully synced full node but almost got out of disk space so I've just used your instruction above and set sync mode to snap and rebuilt the docker.

How long do I wait till geth data dir usage changes from 4TB to 0.25TB? It happens automatically I assume?

@roberto-bayardo
Copy link
Contributor

Fastest way would be to just delete the old DB and start a fresh snap-sync. I am not sure it will go back in time and prune the old data automatically?

@fridary
Copy link

fridary commented Apr 26, 2024

@laptrinhbockchain @MindlessSteel @wangtao19911111
Whereas the current archive chain data is about 2.5TB, you'll be able to snap sync for less than 250GB.

@wbnns why main github page says a high performance SSD drive with at least 4 TB free (NVME recommended) instead of 2.5TB?

@wbnns
Copy link
Member

wbnns commented Apr 26, 2024

@fridary

Heya, this is factoring in room for the archive chain data to grow, plus unallocated space in the environment (in case it needs to be utilized).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
state: backlog This is pending further review type: enhancement New feature or request
Projects
None yet
Development

No branches or pull requests