I'm faced with the problem of a requirement to index approximately
20 to the seventh power of roughly 60 byte string data fields .. per
record.
Yet the average complete record will contain no more than a hundred
populated fields. The rest of all the million plus whatever possible
matrix chunklets - will all be totally blank. Yet in any given record,
there will be the possibility that any mish-mash of field order might
be expected.
The string data is going to be totally variable length, but not more
than 60 bytes per field. It likely may be also internally internally
delimited within each string for pointering purposes, but that's not
an issue for the main task of serving up the chunklets.
The expected number of possible indexes and data to recover might well
exceed 100,000 different record indexes/data matrixes, per computer server
site. The file(s?) and index operations will have to be network
available - perhaps 50 to 100 simultaneous workstations will have to
be able to use the data at peak load. Average connected load may be
on the order of one or two dozen boxes, max.
This isn't a transaction processing operation making direct use of
the proposed index/data hodge-podge. What I speak to will only be
expected to furnish pointers to the transaction processing operation
based on what is in the string data here. Hence the issue of what
I know as roll-forward and roll-backward from the Btrieve game is
maybe not as critical here as it is in some other work I've done.
Any suggestions on how to approach this?
In PB 3.5 ... ?)
------------------
Mike Luther
[email protected]
20 to the seventh power of roughly 60 byte string data fields .. per
record.
Yet the average complete record will contain no more than a hundred
populated fields. The rest of all the million plus whatever possible
matrix chunklets - will all be totally blank. Yet in any given record,
there will be the possibility that any mish-mash of field order might
be expected.
The string data is going to be totally variable length, but not more
than 60 bytes per field. It likely may be also internally internally
delimited within each string for pointering purposes, but that's not
an issue for the main task of serving up the chunklets.
The expected number of possible indexes and data to recover might well
exceed 100,000 different record indexes/data matrixes, per computer server
site. The file(s?) and index operations will have to be network
available - perhaps 50 to 100 simultaneous workstations will have to
be able to use the data at peak load. Average connected load may be
on the order of one or two dozen boxes, max.
This isn't a transaction processing operation making direct use of
the proposed index/data hodge-podge. What I speak to will only be
expected to furnish pointers to the transaction processing operation
based on what is in the string data here. Hence the issue of what
I know as roll-forward and roll-backward from the Btrieve game is
maybe not as critical here as it is in some other work I've done.
Any suggestions on how to approach this?
In PB 3.5 ... ?)
------------------
Mike Luther
[email protected]
Comment