-
Notifications
You must be signed in to change notification settings - Fork 552
/
result_merger.rb
125 lines (110 loc) · 3.72 KB
/
result_merger.rb
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
# frozen_string_literal: true
require "json"
module SimpleCov
#
# Singleton that is responsible for caching, loading and merging
# SimpleCov::Results into a single result for coverage analysis based
# upon multiple test suites.
#
module ResultMerger
class << self
# The path to the .resultset.json cache file
def resultset_path
File.join(SimpleCov.coverage_path, ".resultset.json")
end
def resultset_writelock
File.join(SimpleCov.coverage_path, ".resultset.json.lock")
end
# Loads the cached resultset from JSON and returns it as a Hash,
# caching it for subsequent accesses.
def resultset
@resultset ||= begin
data = stored_data
if data
begin
JSON.parse(data) || {}
rescue StandardError
{}
end
else
{}
end
end
end
# Returns the contents of the resultset cache as a string or if the file is missing or empty nil
def stored_data
synchronize_resultset do
return unless File.exist?(resultset_path)
data = File.read(resultset_path)
return if data.nil? || data.length < 2
data
end
end
# Gets the resultset hash and re-creates all included instances
# of SimpleCov::Result from that.
# All results that are above the SimpleCov.merge_timeout will be
# dropped. Returns an array of SimpleCov::Result items.
def results
results = Result.from_hash(resultset)
results.select { |result| result.time_since_creation < SimpleCov.merge_timeout }
end
def merge_and_store(*results)
result = merge_results(*results)
store_result(result) if result
result
end
# Merge two or more SimpleCov::Results into a new one with merged
# coverage data and the command_name for the result consisting of a join
# on all source result's names
def merge_results(*results)
parsed_results = JSON.parse(JSON.dump(results.map(&:original_result)))
combined_result = SimpleCov::Combine::ResultsCombiner.combine(*parsed_results)
result = SimpleCov::Result.new(combined_result)
# Specify the command name
result.command_name = results.map(&:command_name).sort.join(", ")
result
end
#
# Gets all SimpleCov::Results from cache, merges them and produces a new
# SimpleCov::Result with merged coverage data and the command_name
# for the result consisting of a join on all source result's names
#
def merged_result
merge_results(*results)
end
# Saves the given SimpleCov::Result in the resultset cache
def store_result(result)
synchronize_resultset do
# Ensure we have the latest, in case it was already cached
clear_resultset
new_set = resultset
command_name, data = result.to_hash.first
new_set[command_name] = data
File.open(resultset_path, "w+") do |f_|
f_.puts JSON.pretty_generate(new_set)
end
end
true
end
# Ensure only one process is reading or writing the resultset at any
# given time
def synchronize_resultset
# make it reentrant
return yield if defined?(@resultset_locked) && @resultset_locked
begin
@resultset_locked = true
File.open(resultset_writelock, "w+") do |f|
f.flock(File::LOCK_EX)
yield
end
ensure
@resultset_locked = false
end
end
# Clear out the previously cached .resultset
def clear_resultset
@resultset = nil
end
end
end
end