In this Section we present, (1) execution time of FETA's analysis of all traces, for queries executed in isolation and in concurrence (2) precision/recall of deduced triple patterns and joins, for queries executed in isolation and (3) precision/recall of joins, for queries executed in concurrence. All experiments are made with the same IP Address host, using either Anapsid or FedX query engines.
Summary
In order to evaluate FETA's performance we provide the execution times of FETA's analysis, for both queries executed in isolation and in concurrence. Experiments were run on Ubuntu 14.04 LTS, with 8 CPUs and 32G RAM.
Next, we present execution times of FETA analysis of queries executed in isolation, for Cross Domain (CD) and Life Science (LS) collections.
Query | Anapsid | FedX | Query | Anapsid | FedX |
---|---|---|---|---|---|
CD1 | 0s | 0s | LS1 | 0s | 0s |
CD2 | 0s | 0s | LS2 | 0s | 0s |
CD3 | 0s | 0s | LS3 | 4s | 10s |
CD4 | 0s | 0s | LS4 | 14s | 0s |
CD5 | 0s | 0s | LS5 | 1s | 2s |
CD6 | 33s | 0s | LS6 | 10s | 4m31s |
CD7 | 0s | 0s | LS7 | 1m38s | 2s |
Next, we present execution times of FETA analysis of queries executed in concurrence, for Cross Domain (CD), Life Science (LS) and Mixed (MX) collections.
Query | Anapsid | FedX | Query | Anapsid | FedX | Query | Anapsid | FedX |
---|---|---|---|---|---|---|---|---|
CD mixage1 | 43s | 2s | LS mixage1 | 4m14s | 8m38s | MX mixage1 | 54s | 17s |
CD mixage2 | 44s | 2s | LS mixage2 | 3m5s | 9m25s | MX mixage2 | 47s | 17s |
CD mixage3 | 45s | 2s | LS mixage3 | 4m28s | 9m38s | MX mixage3 | 48s | 17s |
CD mixage4 | 42s | 2s | LS mixage4 | 3m5s | 9m59s | MX mixage4 | 48s | 15s |
First, we evaluate FETA by measuring precision/recall of both deduced triple patterns and joins, for queries executed in isolation comparing to those identified in original federated queries, for Cross Domain and Life Science collections.
Next, we present precision/recall of deduced triple patterns, for Anapsid and FedX traces.
Next, we present precision/recall of deduced joins, for Anapsid and FedX traces.
Next, we evaluate FETA by measuring precision/recall of deduced joins, for queries executed in concurrence comparing to those executed in isolation, for Cross Domain (CD), Life Science (LS) and Mixed (MX) collections.
This experiment was made by varying the user-defined gap threshold, i.e. the maximum temporal distance between two queries to consider them possibly joinables, namely 1%, 10%, 25%, 50%, 75% and 100% of the duration of the input capture trace. For each collection, we generated four different shufflings to simulate a concurrent execution, varying (i) the order of execution of isolated queries, (ii) the number of subqueries, of the same federated query, appearing continuously (blocks of 1 to 16 subqueries), and (iii) the delay between each subquery (from 1 to 16 units of time) of the produced shuffling.
Next, we present precision/recall of deduced joins, concerning the concurrent execution of all CD collection's queries (CD1 to CD7).
ANAPSID traces' precision
ANAPSID traces' recall
FedX traces' precision
FedX traces' recall
Next, we present precision/recall of deduced joins, concerning the concurrent execution of all LS collection queries (LS1 to LS7).
Anapsid traces' precision
Anapsid traces' recall
FedX traces' precision
FedX traces' recall
Next, we present precision/recall of deduced joins, concerning the concurrent execution of a mixed collection. This set is composed of non similar queries from CD and LS collections, i.e. queries that have a null resultset for their common projected variables, namely: CD3, CD4, CD5, CD6, LS2 and LS3.
Anapsid traces' precision
We omit this figure, as precision for this collection is always equal to 1.
Anapsid traces' recall
FedX traces' precision
We omit this figure, as precision for this collection is always equal to 1.
FedX traces' recall