Skip to content

Latest commit

 

History

History
136 lines (77 loc) · 7.82 KB

experiments_with_fedbench.md

File metadata and controls

136 lines (77 loc) · 7.82 KB

FETA: experiments with FedBench

In this Section we present, (1) execution time of FETA's analysis of all traces, for queries executed in isolation and in concurrence (2) precision/recall of deduced triple patterns and joins, for queries executed in isolation and (3) precision/recall of joins, for queries executed in concurrence. All experiments are made with the same IP Address host, using either Anapsid or FedX query engines.

Summary

  1. Execution analysis time
  1. Isolated execution

  2. Concurrent execution

Execution analysis time

In order to evaluate FETA's performance we provide the execution times of FETA's analysis, for both queries executed in isolation and in concurrence. Experiments were run on Ubuntu 14.04 LTS, with 8 CPUs and 32G RAM.

Isolated analysis time

Next, we present execution times of FETA analysis of queries executed in isolation, for Cross Domain (CD) and Life Science (LS) collections.

Query Anapsid FedX Query Anapsid FedX
CD1 0s 0s LS1 0s 0s
CD2 0s 0s LS2 0s 0s
CD3 0s 0s LS3 4s 10s
CD4 0s 0s LS4 14s 0s
CD5 0s 0s LS5 1s 2s
CD6 33s 0s LS6 10s 4m31s
CD7 0s 0s LS7 1m38s 2s

Concurrent analysis time

Next, we present execution times of FETA analysis of queries executed in concurrence, for Cross Domain (CD), Life Science (LS) and Mixed (MX) collections.

Query Anapsid FedX Query Anapsid FedX Query Anapsid FedX
CD mixage1 43s 2s LS mixage1 4m14s 8m38s MX mixage1 54s 17s
CD mixage2 44s 2s LS mixage2 3m5s 9m25s MX mixage2 47s 17s
CD mixage3 45s 2s LS mixage3 4m28s 9m38s MX mixage3 48s 17s
CD mixage4 42s 2s LS mixage4 3m5s 9m59s MX mixage4 48s 15s

Isolated execution

First, we evaluate FETA by measuring precision/recall of both deduced triple patterns and joins, for queries executed in isolation comparing to those identified in original federated queries, for Cross Domain and Life Science collections.

Precision and recall of tps

Next, we present precision/recall of deduced triple patterns, for Anapsid and FedX traces.

GitHub Logo GitHub Logo

Precision and recall of joins

Next, we present precision/recall of deduced joins, for Anapsid and FedX traces.

GitHub Logo GitHub Logo

Concurrent execution

Next, we evaluate FETA by measuring precision/recall of deduced joins, for queries executed in concurrence comparing to those executed in isolation, for Cross Domain (CD), Life Science (LS) and Mixed (MX) collections.

This experiment was made by varying the user-defined gap threshold, i.e. the maximum temporal distance between two queries to consider them possibly joinables, namely 1%, 10%, 25%, 50%, 75% and 100% of the duration of the input capture trace. For each collection, we generated four different shufflings to simulate a concurrent execution, varying (i) the order of execution of isolated queries, (ii) the number of subqueries, of the same federated query, appearing continuously (blocks of 1 to 16 subqueries), and (iii) the delay between each subquery (from 1 to 16 units of time) of the produced shuffling.

Cross domain

Next, we present precision/recall of deduced joins, concerning the concurrent execution of all CD collection's queries (CD1 to CD7).

ANAPSID traces' precision

GitHub Logo

ANAPSID traces' recall

GitHub Logo

FedX traces' precision

GitHub Logo

FedX traces' recall

GitHub Logo

Life science

Next, we present precision/recall of deduced joins, concerning the concurrent execution of all LS collection queries (LS1 to LS7).

Anapsid traces' precision

GitHub Logo

Anapsid traces' recall

GitHub Logo

FedX traces' precision

GitHub Logo

FedX traces' recall

GitHub Logo

Mixed collection

Next, we present precision/recall of deduced joins, concerning the concurrent execution of a mixed collection. This set is composed of non similar queries from CD and LS collections, i.e. queries that have a null resultset for their common projected variables, namely: CD3, CD4, CD5, CD6, LS2 and LS3.

Anapsid traces' precision

We omit this figure, as precision for this collection is always equal to 1.

Anapsid traces' recall

GitHub Logo

FedX traces' precision

We omit this figure, as precision for this collection is always equal to 1.

FedX traces' recall

GitHub Logo