-
Notifications
You must be signed in to change notification settings - Fork 24.3k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Add reserved snapshot/repo action (#89601)
Adds ability to configure snapshot/repo in file based settings.
- Loading branch information
Showing
11 changed files
with
727 additions
and
89 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,5 @@ | ||
pr: 89601 | ||
summary: Add reserved snapshot/repo action | ||
area: Infra/Core | ||
type: enhancement | ||
issues: [] |
248 changes: 248 additions & 0 deletions
248
...lClusterTest/java/org/elasticsearch/reservedstate/service/RepositoriesFileSettingsIT.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,248 @@ | ||
/* | ||
* Copyright Elasticsearch B.V. and/or licensed to Elasticsearch B.V. under one | ||
* or more contributor license agreements. Licensed under the Elastic License | ||
* 2.0 and the Server Side Public License, v 1; you may not use this file except | ||
* in compliance with, at your election, the Elastic License 2.0 or the Server | ||
* Side Public License, v 1. | ||
*/ | ||
|
||
package org.elasticsearch.reservedstate.service; | ||
|
||
import org.elasticsearch.action.admin.cluster.repositories.get.GetRepositoriesAction; | ||
import org.elasticsearch.action.admin.cluster.repositories.get.GetRepositoriesRequest; | ||
import org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryAction; | ||
import org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryRequest; | ||
import org.elasticsearch.action.admin.cluster.repositories.reservedstate.ReservedRepositoryAction; | ||
import org.elasticsearch.client.internal.Client; | ||
import org.elasticsearch.cluster.ClusterChangedEvent; | ||
import org.elasticsearch.cluster.ClusterStateListener; | ||
import org.elasticsearch.cluster.metadata.ReservedStateErrorMetadata; | ||
import org.elasticsearch.cluster.metadata.ReservedStateHandlerMetadata; | ||
import org.elasticsearch.cluster.metadata.ReservedStateMetadata; | ||
import org.elasticsearch.cluster.service.ClusterService; | ||
import org.elasticsearch.common.settings.Settings; | ||
import org.elasticsearch.core.Strings; | ||
import org.elasticsearch.core.Tuple; | ||
import org.elasticsearch.repositories.RepositoryMissingException; | ||
import org.elasticsearch.test.ESIntegTestCase; | ||
import org.elasticsearch.xcontent.XContentParserConfiguration; | ||
|
||
import java.io.ByteArrayInputStream; | ||
import java.nio.charset.StandardCharsets; | ||
import java.nio.file.Files; | ||
import java.nio.file.Path; | ||
import java.nio.file.StandardCopyOption; | ||
import java.util.concurrent.CountDownLatch; | ||
import java.util.concurrent.ExecutionException; | ||
import java.util.concurrent.TimeUnit; | ||
import java.util.concurrent.atomic.AtomicLong; | ||
import java.util.stream.Collectors; | ||
|
||
import static org.elasticsearch.test.NodeRoles.dataOnlyNode; | ||
import static org.elasticsearch.xcontent.XContentType.JSON; | ||
import static org.hamcrest.Matchers.allOf; | ||
import static org.hamcrest.Matchers.containsInAnyOrder; | ||
import static org.hamcrest.Matchers.containsString; | ||
import static org.hamcrest.Matchers.equalTo; | ||
import static org.hamcrest.Matchers.hasSize; | ||
import static org.hamcrest.Matchers.notNullValue; | ||
|
||
@ESIntegTestCase.ClusterScope(scope = ESIntegTestCase.Scope.TEST, numDataNodes = 0, autoManageMasterNodes = false) | ||
public class RepositoriesFileSettingsIT extends ESIntegTestCase { | ||
|
||
private static AtomicLong versionCounter = new AtomicLong(1); | ||
|
||
private static String testJSON = """ | ||
{ | ||
"metadata": { | ||
"version": "%s", | ||
"compatibility": "8.4.0" | ||
}, | ||
"state": { | ||
"snapshot_repositories": { | ||
"repo": { | ||
"type": "fs", | ||
"settings": { | ||
"location": "my_backup_location" | ||
} | ||
}, | ||
"repo1": { | ||
"type": "fs", | ||
"settings": { | ||
"location": "my_backup_location_1" | ||
} | ||
} | ||
} | ||
} | ||
}"""; | ||
|
||
private static String testErrorJSON = """ | ||
{ | ||
"metadata": { | ||
"version": "%s", | ||
"compatibility": "8.4.0" | ||
}, | ||
"state": { | ||
"snapshot_repositories": { | ||
"err-repo": { | ||
"type": "interstelar", | ||
"settings": { | ||
"location": "my_backup_location" | ||
} | ||
} | ||
} | ||
} | ||
}"""; | ||
|
||
private void assertMasterNode(Client client, String node) throws ExecutionException, InterruptedException { | ||
assertThat(client.admin().cluster().prepareState().execute().get().getState().nodes().getMasterNode().getName(), equalTo(node)); | ||
} | ||
|
||
private void writeJSONFile(String node, String json) throws Exception { | ||
long version = versionCounter.incrementAndGet(); | ||
|
||
FileSettingsService fileSettingsService = internalCluster().getInstance(FileSettingsService.class, node); | ||
|
||
Files.createDirectories(fileSettingsService.operatorSettingsDir()); | ||
Path tempFilePath = createTempFile(); | ||
|
||
Files.write(tempFilePath, Strings.format(json, version).getBytes(StandardCharsets.UTF_8)); | ||
Files.move(tempFilePath, fileSettingsService.operatorSettingsFile(), StandardCopyOption.ATOMIC_MOVE); | ||
} | ||
|
||
private Tuple<CountDownLatch, AtomicLong> setupClusterStateListener(String node) { | ||
ClusterService clusterService = internalCluster().clusterService(node); | ||
CountDownLatch savedClusterState = new CountDownLatch(1); | ||
AtomicLong metadataVersion = new AtomicLong(-1); | ||
clusterService.addListener(new ClusterStateListener() { | ||
@Override | ||
public void clusterChanged(ClusterChangedEvent event) { | ||
ReservedStateMetadata reservedState = event.state().metadata().reservedStateMetadata().get(FileSettingsService.NAMESPACE); | ||
if (reservedState != null) { | ||
ReservedStateHandlerMetadata handlerMetadata = reservedState.handlers().get(ReservedRepositoryAction.NAME); | ||
if (handlerMetadata != null && handlerMetadata.keys().contains("repo")) { | ||
clusterService.removeListener(this); | ||
metadataVersion.set(event.state().metadata().version()); | ||
savedClusterState.countDown(); | ||
} | ||
} | ||
} | ||
}); | ||
|
||
return new Tuple<>(savedClusterState, metadataVersion); | ||
} | ||
|
||
private void assertClusterStateSaveOK(CountDownLatch savedClusterState, AtomicLong metadataVersion) throws Exception { | ||
boolean awaitSuccessful = savedClusterState.await(20, TimeUnit.SECONDS); | ||
assertTrue(awaitSuccessful); | ||
|
||
final var reposResponse = client().execute( | ||
GetRepositoriesAction.INSTANCE, | ||
new GetRepositoriesRequest(new String[] { "repo", "repo1" }) | ||
).get(); | ||
|
||
assertThat( | ||
reposResponse.repositories().stream().map(r -> r.name()).collect(Collectors.toSet()), | ||
containsInAnyOrder("repo", "repo1") | ||
); | ||
|
||
assertEquals( | ||
"Failed to process request " | ||
+ "[org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryRequest/unset] " | ||
+ "with errors: [[repo] set as read-only by [file_settings]]", | ||
expectThrows( | ||
IllegalArgumentException.class, | ||
() -> client().execute(PutRepositoryAction.INSTANCE, sampleRestRequest("repo")).actionGet() | ||
).getMessage() | ||
); | ||
} | ||
|
||
public void testSettingsApplied() throws Exception { | ||
internalCluster().setBootstrapMasterNodeIndex(0); | ||
logger.info("--> start data node / non master node"); | ||
var dataNode = internalCluster().startNode(Settings.builder().put(dataOnlyNode()).put("discovery.initial_state_timeout", "1s")); | ||
|
||
var savedClusterState = setupClusterStateListener(dataNode); | ||
// In internal cluster tests, the nodes share the config directory, so when we write with the data node path | ||
// the master will pick it up on start | ||
writeJSONFile(dataNode, testJSON); | ||
|
||
logger.info("--> start master node"); | ||
final String masterNode = internalCluster().startMasterOnlyNode(); | ||
assertMasterNode(internalCluster().nonMasterClient(), masterNode); | ||
|
||
assertClusterStateSaveOK(savedClusterState.v1(), savedClusterState.v2()); | ||
} | ||
|
||
private Tuple<CountDownLatch, AtomicLong> setupClusterStateListenerForError(String node) { | ||
ClusterService clusterService = internalCluster().clusterService(node); | ||
CountDownLatch savedClusterState = new CountDownLatch(1); | ||
AtomicLong metadataVersion = new AtomicLong(-1); | ||
clusterService.addListener(new ClusterStateListener() { | ||
@Override | ||
public void clusterChanged(ClusterChangedEvent event) { | ||
ReservedStateMetadata reservedState = event.state().metadata().reservedStateMetadata().get(FileSettingsService.NAMESPACE); | ||
if (reservedState != null && reservedState.errorMetadata() != null) { | ||
assertEquals(ReservedStateErrorMetadata.ErrorKind.VALIDATION, reservedState.errorMetadata().errorKind()); | ||
assertThat(reservedState.errorMetadata().errors(), allOf(notNullValue(), hasSize(1))); | ||
assertThat( | ||
reservedState.errorMetadata().errors().get(0), | ||
containsString("[err-repo] repository type [interstelar] does not exist") | ||
); | ||
clusterService.removeListener(this); | ||
metadataVersion.set(event.state().metadata().version()); | ||
savedClusterState.countDown(); | ||
} | ||
} | ||
}); | ||
|
||
return new Tuple<>(savedClusterState, metadataVersion); | ||
} | ||
|
||
private void assertClusterStateNotSaved(CountDownLatch savedClusterState, AtomicLong metadataVersion) throws Exception { | ||
boolean awaitSuccessful = savedClusterState.await(20, TimeUnit.SECONDS); | ||
assertTrue(awaitSuccessful); | ||
|
||
assertEquals( | ||
"[err-repo] missing", | ||
expectThrows( | ||
RepositoryMissingException.class, | ||
() -> client().execute(GetRepositoriesAction.INSTANCE, new GetRepositoriesRequest(new String[] { "err-repo" })).actionGet() | ||
).getMessage() | ||
); | ||
|
||
// This should succeed, nothing was reserved | ||
client().execute(PutRepositoryAction.INSTANCE, sampleRestRequest("err-repo")).get(); | ||
} | ||
|
||
public void testErrorSaved() throws Exception { | ||
internalCluster().setBootstrapMasterNodeIndex(0); | ||
logger.info("--> start data node / non master node"); | ||
internalCluster().startNode(Settings.builder().put(dataOnlyNode()).put("discovery.initial_state_timeout", "1s")); | ||
|
||
logger.info("--> start master node"); | ||
final String masterNode = internalCluster().startMasterOnlyNode(); | ||
assertMasterNode(internalCluster().nonMasterClient(), masterNode); | ||
var savedClusterState = setupClusterStateListenerForError(masterNode); | ||
|
||
writeJSONFile(masterNode, testErrorJSON); | ||
assertClusterStateNotSaved(savedClusterState.v1(), savedClusterState.v2()); | ||
} | ||
|
||
private PutRepositoryRequest sampleRestRequest(String name) throws Exception { | ||
var json = """ | ||
{ | ||
"type": "fs", | ||
"settings": { | ||
"location": "my_backup_location_2" | ||
} | ||
}"""; | ||
|
||
try ( | ||
var bis = new ByteArrayInputStream(json.getBytes(StandardCharsets.UTF_8)); | ||
var parser = JSON.xContent().createParser(XContentParserConfiguration.EMPTY, bis) | ||
) { | ||
return new PutRepositoryRequest(name).source(parser.map()); | ||
} | ||
} | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.