You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In absence of transaction, incorrect queries can lead to loss of data
Often this needs to be done on tables that are receiving new inserts during the update.
Are there recommended ways to do so safely while minimizing I/O load? Could you please comment on alternatives e.g.
a) Copy the whole table and run a mutation on the copy, then exchange the tables
b) Select+insert the affected rows into a new table, then copy them back and run OPTIMIZE TABLE ... FINAL assuming deduplication can be used
c) Select+insert the affected rows into a new table, run a DELETE mutation on the original table and copy the rows back
Thanks!
The text was updated successfully, but these errors were encountered:
@den-crane FWIW we do the same but we are not confident that data corruption bugs or stuck mutations are 100% reproducible across testbeds and production, hence my question.
Sometimes it is necessary to update the value of a column across a large number of row.
The
ALTER TABLE ... UPDATE
command presents some risks:Often this needs to be done on tables that are receiving new inserts during the update.
Are there recommended ways to do so safely while minimizing I/O load? Could you please comment on alternatives e.g.
a) Copy the whole table and run a mutation on the copy, then exchange the tables
b) Select+insert the affected rows into a new table, then copy them back and run OPTIMIZE TABLE ... FINAL assuming deduplication can be used
c) Select+insert the affected rows into a new table, run a DELETE mutation on the original table and copy the rows back
Thanks!
The text was updated successfully, but these errors were encountered: