HDF5
1.14.4.3
API Reference
|
Boolean field used to enable and disable the default reporting function. This function is invoked every time the automatic cache resize code is run, and reports on its activities.
This is a debugging function, and should normally be turned off.
Boolean field indicating whether the trace_file_name field should be used to open a trace file for the cache.
*** DEPRECATED *** Use H5Fstart/stop
logging functions instead
The trace file is a debugging feature that allow the capture of top level metadata cache requests for purposes of debugging and/or optimization. This field should normally be set to FALSE
, as trace file collection imposes considerable overhead.
This field should only be set to TRUE
when the trace_file_name contains the full path of the desired trace file, and either there is no open trace file on the cache, or the close_trace_file
field is also TRUE
.
Boolean field indicating whether the current trace file (if any) should be closed.
*** DEPRECATED *** Use H5Fstart/stop
logging functions instead
See the above comments on the open_trace_file field. This field should be set to FALSE
unless there is an open trace file on the cache that you wish to close.
Full path of the trace file to be opened if the open_trace_file field is TRUE
.
*** DEPRECATED *** Use H5Fstart/stop
logging functions instead
In the parallel case, an ascii representation of the mpi rank of the process will be appended to the file name to yield a unique trace file name for each process.
The length of the path must not exceed H5AC__MAX_TRACE_FILE_NAME_LEN characters.
Boolean field used to either report the current evictions enabled status of the cache, or to set the cache's evictions enabled status.
In general, the metadata cache should always be allowed to evict entries. However, in some cases it is advantageous to disable evictions briefly, and thereby postpone metadata writes. However, this must be done with care, as the cache can grow quickly. If you do this, re-enable evictions as soon as possible and monitor cache size.
At present, evictions can only be disabled if automatic cache resizing is also disabled (that is, (incr_mode == H5C_incr__off ) && ( decr_mode == H5C_decr__off )
). There is no logical reason why this should be so, but it simplifies implementation and testing, and I can't think of any reason why it would be desirable. If you can think of one, I'll revisit the issue. (JM)
FALSE
, the initial_size field is ignored.[min_size, max_size]
.double
in the range 0 to 1 indicating the fraction of the cache that is to be kept clean. This field is only used in parallel mode. Typical values are 0.1 to 0.5.[MIN_MAX_CACHE_SIZE, MAX_MAX_CACHE_SIZE]
. Also, max_size
must be greater than or equal to min_size
.[H5C__MIN_MAX_CACHE_SIZE, H5C__MAX_MAX_CACHE_SIZE]
. Also, min_size
must be less than or equal to max_size
.Number of accesses on the cache over which to collect hit rate stats before running the automatic cache resize code, if it is enabled.
At the end of an epoch, we discard prior hit rate data and start collecting afresh. The epoch_length must lie in the closed interval [H5C__MIN_AR_EPOCH_LENGTH, H5C__MAX_AR_EPOCH_LENGTH]
.
H5C_cache_incr_mode
enumerated type whose value indicates how we determine whether the cache size should be increased. At present there are two possible values: H5C_incr__off:
Don't attempt to increase the size of the cache automatically.H5C_incr__threshold:
Attempt to increase the size of the cache whenever the average hit rate over the last epoch drops below the value supplied in the lower_hr_threshold
field.decr_mode
to H5C_incr__off
if you disable metadata cache entry evictions.Lower hit rate threshold. If the increment mode (incr_mode
) is H5C_incr__threshold
and the hit rate drops below the value supplied in this field in an epoch, increment the cache size by size_increment
. Note that cache size may not be incremented above max_size
, and that the increment may be further restricted by the max_increment
field if it is enabled.
When enabled, this field must contain a value in the range [0.0, 1.0]. Depending on the incr_mode
selected, it may also have to be less than upper_hr_threshold
.
Double containing the multiplier used to derive the new cache size from the old if a cache size increment is triggered. The increment must be greater than 1.0, and should not exceed 2.0.
The new cache size is obtained my multiplying the current max cache size by the increment, and then clamping to max_size
and to stay within the max_increment
as necessary.
max_increment
field should be used to limit the maximum cache size increment.apply_max_increment
field described above, this field contains the maximum number of bytes by which the cache size can be increased in a single re-size.Instance of the H5C_cache_flash_incr_mode
enumerated type whose value indicates whether and by which algorithm we should make flash increases in the size of the cache to accommodate insertion of large entries and large increases in the size of a single entry.
The addition of the flash increment mode was occasioned by performance problems that appear when a local heap is increased to a size in excess of the current cache size. While the existing re-size code dealt with this eventually, performance was very bad for the remainder of the epoch.
At present, there are two possible values for the flash_incr_mode:
H5C_flash_incr__off:
Don't perform flash increases in the size of the cache.H5C_flash_incr__add_space:
Let x
be either the size of a newly newly inserted entry, or the number of bytes by which the size of an existing entry has been increased.x > flash_threshold * current max cache size
, increase the current maximum cache size by x * flash_multiple
less any free space in the cache, and star a new epoch. For now at least, pay no attention to the maximum increment.In both of the above cases, the flash increment pays no attention to the maximum increment (at least in this first incarnation), but DOES stay within max_size.
With a little thought, it should be obvious that the above flash cache size increase algorithm is not sufficient for all circumstances – for example, suppose the user round robins through (1/flash_threshold) +1
groups, adding one data set to each on each pass. Then all will increase in size at about the same time, requiring the max cache size to at least double to maintain acceptable performance, however the above flash increment algorithm will not be triggered.
Hopefully, the add space algorithms detailed above will be sufficient for the performance problems encountered to date. However, we should expect to revisit the issue.
H5C_flash_incr__add_space
section of the discussion of the flash_incr_mode
section. This field is ignored unless flash_incr_mode
is H5C_flash_incr__add_space
.flash_incr_mode
is H5C_flash_incr__add_space
.Instance of the H5C_cache_decr_mode
enumerated type whose value indicates how we determine whether the cache size should be decreased. At present there are four possibilities.
H5C_decr__off:
Don't attempt to decrease the size of the cache automatically.H5C_decr__threshold:
Attempt to decrease the size of the cache whenever the average hit rate over the last epoch rises above the value supplied in the upper_hr_threshold
field. H5C_decr__age_out:
At the end of each epoch, search the cache for entries that have not been accessed for at least the number of epochs specified in the epochs_before_eviction field, and evict these entries. Conceptually, the maximum cache size is then decreased to match the new actual cache size. However, this reduction may be modified by the min_size
, the max_decrement
, and/or the empty_reserve
. H5C_decr__age_out_with_threshold:
Same as age_out, but we only attempt to reduce the cache size when the hit rate observed over the last epoch exceeds the value provided in the upper_hr_threshold
field.Note that you must set decr_mode
to H5C_decr__off
if you disable metadata cache entry evictions.
Upper hit rate threshold. The use of this field varies according to the current decr_mode
:
H5C_decr__off
or H5C_decr__age_out:
The value of this field is ignored.
H5C_decr__threshold:
If the hit rate exceeds this threshold in any epoch, attempt to decrement the cache size by size_decrement.min_size
.upper_threshold
is 1.0, the cache sizeH5C_decr__age_out_with_threshold:
If the hit rate exceeds this threshold in any epoch, attempt to reduce the cache size by evicting entries that have not been accessed for more than the specified number of epochs. This field is only used when the decr_mode is H5C_decr__threshold
.
The field is a double containing the multiplier used to derive the new cache size from the old if a cache size decrement is triggered. The decrement must be in the range 0.0 (in which case the cache will try to contract to its minimum size) to 1.0 (in which case the cache will never shrink).
max_decrement
field.min_size
of the cache, and (in age out modes) by the empty_reserve
field.Integer field used in H5C_decr__age_out
and H5C_decr__age_out_with_threshold
decrement modes.
This field contains the number of epochs an entry must remain unaccessed before it is evicted in an attempt to reduce the cache size. If applicable, this field must lie in the range [1, H5C__MAX_EPOCH_MARKERS]
.
To avoid a constant racheting down of cache size by small amounts in the H5C_decr__age_out
and H5C_decr__age_out_with_threshold
modes, this field allows one to require that any cache size reductions leave the specified fraction of unused space in the cache.
The value of this field must be in the range [0.0, 1.0]. I would expect typical values to be in the range of 0.01 to 0.1.
In PHDF5, all operations that modify metadata must be executed collectively.
We used to think that this was enough to ensure consistency across the metadata caches, but since we allow processes to read metadata individually, the order of dirty entries in the LRU list can vary across processes, which can result in inconsistencies between the caches.
PHDF5 uses several strategies to prevent such inconsistencies in metadata, all of which use the fact that the same stream of dirty metadata is seen by all processes for purposes of synchronization. This is done by having each process count the number of bytes of dirty metadata generated, and then running a "sync point" whenever this count exceeds a user specified threshold (see dirty_bytes_threshold
below).
The current metadata write strategy is indicated by the metadata_write_strategy
field. The possible values of this field, along with the associated metadata write strategies are discussed below.
Threshold of dirty byte creation used to synchronize updates between caches. (See above for outline and motivation.)
This value MUST be consistent across all processes accessing the file. This field is ignored unless HDF5 has been compiled for parallel.