dcd8284393
(to be imported soon).
23 lines
1.0 KiB
Plaintext
23 lines
1.0 KiB
Plaintext
1. provide output diagnostics that explain how many input keys total,
|
|
how many after dealing with static links, and finally, after the
|
|
algorithm is complete, how many dynamic duplicates do we now
|
|
have.
|
|
2. fix up GATHER_STATISTICS for all instrumentation.
|
|
3. Useful idea:
|
|
|
|
a. Generate the wordlist as a contiguous block of keywords, as before.
|
|
This wordlist *must* be sorted by hash value.
|
|
|
|
b. generate the lookup_array, which are an array of signed {chars,shorts,ints},
|
|
which ever allows full coverage of the wordlist dimensions. If the
|
|
value v, where v = lookup_array[hash(str,len)], is >= 0, then we
|
|
simply use this result as a direct access into wordlist to snag
|
|
the keyword for comparison.
|
|
|
|
c. Otherwise, if v is < 0 this is an indication that we'll need to
|
|
search through some number of duplicates hash values. Using a
|
|
hash linking scheme we'd then index into a duplicate_address
|
|
table that would provide the starting index and total length of
|
|
the duplicate entries to consider sequentially.
|
|
|