This repository has been archived by the owner on Jul 11, 2022. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 83
/
MigrateData.java
136 lines (116 loc) · 5.04 KB
/
MigrateData.java
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
package org.rhq.cassandra.schema;
import java.util.ArrayList;
import java.util.Date;
import java.util.List;
import com.datastax.driver.core.ResultSet;
import com.datastax.driver.core.ResultSetFuture;
import com.datastax.driver.core.Row;
import com.datastax.driver.core.Session;
import com.datastax.driver.core.SimpleStatement;
import com.datastax.driver.core.Statement;
import com.datastax.driver.core.querybuilder.Batch;
import com.datastax.driver.core.querybuilder.QueryBuilder;
import com.google.common.util.concurrent.AsyncFunction;
import com.google.common.util.concurrent.Futures;
import com.google.common.util.concurrent.ListenableFuture;
import com.google.common.util.concurrent.RateLimiter;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.joda.time.DateTime;
import org.joda.time.Seconds;
/**
* @author John Sanda
*/
public class MigrateData implements AsyncFunction<ResultSet, List<ResultSet>> {
private static final Log log = LogFactory.getLog(MigrateData.class);
private static final int BATCH_SIZE = 45;
private Integer scheduleId;
private MigrateAggregateMetrics.Bucket bucket;
private RateLimiter writePermits;
private Session session;
private Seconds ttl;
public MigrateData(Integer scheduleId, MigrateAggregateMetrics.Bucket bucket, RateLimiter writePermits,
Session session, Seconds ttl) {
this.scheduleId = scheduleId;
this.bucket = bucket;
this.writePermits = writePermits;
this.session = session;
this.ttl = ttl;
}
@Override
public ListenableFuture<List<ResultSet>> apply(ResultSet resultSet) throws Exception {
try {
List<ResultSetFuture> insertFutures = new ArrayList<ResultSetFuture>();
if (resultSet.isExhausted()) {
return Futures.allAsList(insertFutures);
}
List<Row> rows = resultSet.all();
Date time = rows.get(0).getDate(0);
Date nextTime;
Double max = null;
Double min = null;
Double avg = null;
Seconds elapsedSeconds = Seconds.secondsBetween(DateTime.now(), new DateTime(time));
List<Statement> statements = new ArrayList<Statement>(BATCH_SIZE);
for (Row row : rows) {
nextTime = row.getDate(0);
if (nextTime.equals(time)) {
int type = row.getInt(1);
switch (type) {
case 0:
max = row.getDouble(2);
break;
case 1:
min = row.getDouble(2);
break;
default:
avg = row.getDouble(2);
}
} else {
if (elapsedSeconds.isLessThan(ttl)) {
if (isDataMissing(avg, max, min)) {
if (log.isDebugEnabled()) {
log.debug("We only have a partial " + bucket + " metric for {scheduleId: " +
scheduleId + ", time: " + time.getTime() + "}. It will not be migrated.");
}
} else {
int newTTL = ttl.getSeconds() - elapsedSeconds.getSeconds();
statements.add(createInsertStatement(time, avg, max, min, newTTL));
if (statements.size() == BATCH_SIZE) {
insertFutures.add(writeBatch(statements));
statements.clear();
}
}
time = nextTime;
max = row.getDouble(2);
min = null;
avg = null;
}
}
}
if (!statements.isEmpty()) {
insertFutures.add(writeBatch(statements));
}
return Futures.allAsList(insertFutures);
} catch (Exception e) {
log.warn("An error occurred while migrating data", e);
throw e;
}
}
private boolean isDataMissing(Double avg, Double max, Double min) {
if (avg == null || Double.isNaN(avg)) return true;
if (max == null || Double.isNaN(max)) return true;
if (min == null || Double.isNaN(min)) return true;
return false;
}
private ResultSetFuture writeBatch(List<Statement> statements) {
Batch batch = QueryBuilder.batch(statements.toArray(new Statement[statements.size()]));
writePermits.acquire();
return session.executeAsync(batch);
}
private SimpleStatement createInsertStatement(Date time, Double avg, Double max, Double min, int newTTL) {
return new SimpleStatement("INSERT INTO rhq.aggregate_metrics(schedule_id, bucket, time, avg, max, min) " +
"VALUES (" + scheduleId + ", '" + bucket + "', " + time.getTime() + ", " + avg + ", " + max + ", " + min +
") USING TTL " + newTTL);
}
}