Please cite this paper if you use this library in your work:
Erkang Zhu, Fatemeh Nargesian, Ken Q. Pu, Renée J. Miller: LSH Ensemble: Internet-Scale Domain Search. PVLDB 9(12): 1185-1196 (2016) Link
Presentation slides @ VLDB 2016, New Delhi.
We used two datasets for evaluation. The datasets are all from public domains and can be downloaded directly from the original publisher.
- Canadian Open Data, tabular domains only (as of June 2015):
Each file corresponds to a single domain extracted from a column of
a table, which could be a spreadsheet or an CSV file.
The filenames follow the
<data file name>.<column id>
format. - 2015 WDC Web Tables, English Relational, 51 compressed files: See the data format here.
By using these datasets you agree to use them for academic research purpose only, and we shall not be held responisble for any inaccuracy or error that may exist in the datasets, nor we shall be responsible for any consequence of usage of these datasets.
Install this library by running:
go get github.com/ekzhu/lshensemble
Import the library in your import
:
import (
"github.com/ekzhu/lshensemble"
)
First you need to obtain the domains, and each domain should have a string ID.
For simplicity we represent a domain as map[string]bool
, whose keys are distinct values.
Assuming you have obtained the domains from some dataset,
you can generate the MinHash signatures from the domains.
domains []map[string]bool
keys []string // each key corresponds to the domain at the same index
// ...
// obtaining domains and keys
// ...
// initializing the domain records to hold the MinHash signatures
domainRecords := make([]*lshensemble.DomainRecord, len(domains))
// set the minhash seed
seed := 42
// set the number of hash functions
numHash := 256
// create the domain records with the signatures
for i := range domains {
mh := lshensemble.NewMinhash(seed, numHash)
for v := range domains[i] {
mh.Push([]byte(v))
}
domainRecords[i] := &lshensemble.DomainRecord {
Key : keys[i],
Size : len(domains[i]),
Signature : mh.Signature()
}
}
Before you can index the domains, you need to sort them in increasing order by
their sizes. BySize
wrapper allows the domains to tbe sorted using the build-in sort
package.
sort.Sort(lshensemble.BySize(domainRecords))
Now you can use BootstrapLshEnsembleOptimal
/BootstrapLshEnsembleEquiDepth
(or BootstrapLshEnsemblePlusOptimal
/BootstrapLshEnsemblePlusEquiDepth
)
for better accuracy at higher memory cost*)
to create an LSH Ensemble index.
BootstrapLshEnsembleOptimal
uses dynamic programming to create partitions that
are optimal in the sense that the total number of false positives generated from
all the partitions are minimized. This method can be
a bit slower due to the dynamic programming overhead, however, it creates
optimized partitions for any kind of data distribution.
BootstrapLshEnsembleEquiDepth
uses simple equi-depth -- same number of domains
in every partition. This is method is described in the original
paper as suitable for power-law
distributed domain sizes, which is common in real-world domains.
You need to
specify the number of partitions to use and some other parameters.
The LSH parameter K (number of hash functions per band) is dynamically tuned at query-time,
but the maximum value should be specified here.
* See explanation for the reason for the "Plus" version.
// Set the number of partitions
numPart := 8
// Set the maximum value for the MinHash LSH parameter K
// (number of hash functions per band).
maxK := 4
// Create index using equi-depth partitioning
// You can also use BootstrapLshEnsemblePlusEquiDepth for better accuracy
index_eqd, err := lshensemble.BootstrapLshEnsembleEquiDepth(numPart, numHash, maxK,
len(domainRecords), lshensemble.Recs2Chan(domainRecords))
if err != nil {
panic(err)
}
// Create index using optimal partitioning
// You can also use BootstrapLshEnsemblePlusOptimal for better accuracy
index_opt, err := lshensemble.BootstrapLshEnsembleOptimal(numPart, numHash, maxK,
func () <-chan *lshensemble.DomainRecord {
return lshensemble.Recs2Chan(domainRecords);
})
if err != nil {
panic(err)
}
For better memory efficiency when the number of domains is large,
it's wiser to use Golang channels and goroutines
to pipeline the generation of the signatures, and then use disk-based sorting to sort the domain records.
This is why BootstrapLshEnsembleEquiDepth
accepts a channel of *DomainRecord
as input.
For a small number of domains, you simply use Recs2Chan
to convert the sorted slice of *DomainRecord
into a chan *DomainRecord
.
To help serializing the domain records to disk, you can use SerializeSignature
to serialize the signatures.
You need to come up with your own serialization schema for the keys and sizes.
Lastly, you can use Query
function, which returns a Golang channel of
the keys of the candidates domains, which may contain false positives - domains that do not
meet the containment threshold.
Therefore, you can optionally include a post-processing step to remove
the false positive domains using the original domain values.
// pick a domain to use as the query
querySig := domainRecords[0].Signature
querySize := domainRecords[0].Size
// set the containment threshold
threshold := 0.5
// get the keys of the candidate domains (may contain false positives)
// through a channel with option to cancel early.
done := make(chan struct{})
defer close(done) // Important!!
results := index.Query(querySig, querySize, threshold, done)
for key := range results {
// ...
// You may want to include a post-processing step here to remove
// false positive domains using the actual domain values.
// ...
// You can call break here to stop processing results.
}
First you need to download the Canadian Open Data domains
and extract the domain files into a directory called _cod_domains
by running the following command.
tar xzf canadian_open_data_tabular_domains_only.tar.gz
Use Golang's test
tool to start the benchmark:
go test -bench=Benchmark_CanadianOpenData -timeout=24h
The benchmark process is in the following order:
- Read the domain files into memory
- Run Linear Scan to get the ground truth
- Run LSH Ensemble to get the query results
- Run the accuracy analysis to generate a report on precisions and recalls
MinHash LSH has two parameters K
and L
(in the
paper
I used r
and b
respectively).
L
is the number of "bands" and K
is the number of hash functions per band.
The details about the two parameters can be found in
Chapter 3 of the textbook,
Mining of Massive Datasets.
In LSH Ensemble, we want to allow the K
and L
of the LSH index in every partition to
vary at query time, in order to optimize them for any given query
(see Section 5.5 of the paper).
We can use achive this by using multiple MinHash LSH, one for each value of K
.
This allows us to vary the parameter K
and L
in the following space:
K * L <= number of hash functions (let this be H)
1 <= K <= H
However, this means that for every possible value of K
from 1 to H
,
we need to create a MinHash LSH -- very expensive.
So it is not wise to allow K
to vary from 1 to H
,
and that's why we have a MaxK
parameter, which bounds K
and saves memory.
So the new parameter space is:
K * L <= H
1 <= K <= MaxK
It is important to note that it is not the case for L
,
because we can choose how many "bands" to use at query time.
Now, if we use LSH Forest,
we can vary the parameter K
from 1 to MaxK
at query time with just one LSH.
You can read the paper to understand how this can be done
(hint: prefix tree).
This comes at a price -- the parameter space is more restricted:
MaxK * L <= H
1 <= K <= MaxK
Essentially, we have less freedom in varying L
, as
1 <= L <= min{H / MaxK, H}
base on the above constraints.
In this library for LSH Ensemble, we provide both implmentations (LSH Forest and "vanilla" MinHash LSH ). Specifically,
BootstrapLshEnsembleEquiDepth
andBootstrapLshEnsembleOptimal
build the index using the LSH Forest implementation, which use less memory but with a more restricted parameter space for optimization.BootstrapLshEnsemblePlusEquiDepth
andBootstrapLshEnsemblePlusOptimal
build the index using the "vanilla" MinHash LSH implementation (one LSH for everyK
), which uses more memory (bounded byMaxK
) but with no restriction onL
.
We found that the optimal K
for most queries are less than 4. So in practice you
can just set MaxK
to 4.