== Estimate time for Improving deduplication: 14 MD =
Degree of uncertainty: B
Responsible: SVC
Improving deduplication
Introduction
Generating deduplicating indices is currently error-prone and time-consuming. It is however my belief, that the OOM bugs in the batch monitoring software may limit the errors until now seen while generating indices.
Task 1: Try to retrieve data from other bitarchive replica if the local bitarchive replica is unresponsive
Refer to FR XXXXX.
Task 2: Optimize the fetching process.
To generate the deduplication indices, it needs to fetch a lot of metadata-records, specifically the records containing the crawl-logs, and cdx-entries.
Currently, our software for each job-id included in the dedup index, launches two batch-jobs, one for the cdxes, and one for the crawllog. This could maybe be reduced to one job, that maybe also does some of the processing locally on the storage-machines.
Task 3: Optimize the indexing process.
Currently, part of indexing process is a merging of the information found in the CDX-file with the information found in the crawllog. Investigate if this merging of information could be done as part of the post-processing, or as a pre-processing step, if this information is absent in the metadata-records.