1277 lines
		
	
	
		
			53 KiB
		
	
	
	
		
			TypeScript
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			1277 lines
		
	
	
		
			53 KiB
		
	
	
	
		
			TypeScript
		
	
	
		
			Executable File
		
	
	
	
	
| /**
 | |
|  * @module LRUCache
 | |
|  */
 | |
| declare const TYPE: unique symbol;
 | |
| export type PosInt = number & {
 | |
|     [TYPE]: 'Positive Integer';
 | |
| };
 | |
| export type Index = number & {
 | |
|     [TYPE]: 'LRUCache Index';
 | |
| };
 | |
| export type UintArray = Uint8Array | Uint16Array | Uint32Array;
 | |
| export type NumberArray = UintArray | number[];
 | |
| declare class ZeroArray extends Array<number> {
 | |
|     constructor(size: number);
 | |
| }
 | |
| export type { ZeroArray };
 | |
| export type { Stack };
 | |
| export type StackLike = Stack | Index[];
 | |
| declare class Stack {
 | |
|     #private;
 | |
|     heap: NumberArray;
 | |
|     length: number;
 | |
|     static create(max: number): StackLike;
 | |
|     constructor(max: number, HeapCls: {
 | |
|         new (n: number): NumberArray;
 | |
|     });
 | |
|     push(n: Index): void;
 | |
|     pop(): Index;
 | |
| }
 | |
| /**
 | |
|  * Promise representing an in-progress {@link LRUCache#fetch} call
 | |
|  */
 | |
| export type BackgroundFetch<V> = Promise<V | undefined> & {
 | |
|     __returned: BackgroundFetch<V> | undefined;
 | |
|     __abortController: AbortController;
 | |
|     __staleWhileFetching: V | undefined;
 | |
| };
 | |
| export type DisposeTask<K, V> = [
 | |
|     value: V,
 | |
|     key: K,
 | |
|     reason: LRUCache.DisposeReason
 | |
| ];
 | |
| export declare namespace LRUCache {
 | |
|     /**
 | |
|      * An integer greater than 0, reflecting the calculated size of items
 | |
|      */
 | |
|     type Size = number;
 | |
|     /**
 | |
|      * Integer greater than 0, representing some number of milliseconds, or the
 | |
|      * time at which a TTL started counting from.
 | |
|      */
 | |
|     type Milliseconds = number;
 | |
|     /**
 | |
|      * An integer greater than 0, reflecting a number of items
 | |
|      */
 | |
|     type Count = number;
 | |
|     /**
 | |
|      * The reason why an item was removed from the cache, passed
 | |
|      * to the {@link Disposer} methods.
 | |
|      *
 | |
|      * - `evict`: The item was evicted because it is the least recently used,
 | |
|      *   and the cache is full.
 | |
|      * - `set`: A new value was set, overwriting the old value being disposed.
 | |
|      * - `delete`: The item was explicitly deleted, either by calling
 | |
|      *   {@link LRUCache#delete}, {@link LRUCache#clear}, or
 | |
|      *   {@link LRUCache#set} with an undefined value.
 | |
|      * - `expire`: The item was removed due to exceeding its TTL.
 | |
|      * - `fetch`: A {@link OptionsBase#fetchMethod} operation returned
 | |
|      *   `undefined` or was aborted, causing the item to be deleted.
 | |
|      */
 | |
|     type DisposeReason = 'evict' | 'set' | 'delete' | 'expire' | 'fetch';
 | |
|     /**
 | |
|      * A method called upon item removal, passed as the
 | |
|      * {@link OptionsBase.dispose} and/or
 | |
|      * {@link OptionsBase.disposeAfter} options.
 | |
|      */
 | |
|     type Disposer<K, V> = (value: V, key: K, reason: DisposeReason) => void;
 | |
|     /**
 | |
|      * A function that returns the effective calculated size
 | |
|      * of an entry in the cache.
 | |
|      */
 | |
|     type SizeCalculator<K, V> = (value: V, key: K) => Size;
 | |
|     /**
 | |
|      * Options provided to the
 | |
|      * {@link OptionsBase.fetchMethod} function.
 | |
|      */
 | |
|     interface FetcherOptions<K, V, FC = unknown> {
 | |
|         signal: AbortSignal;
 | |
|         options: FetcherFetchOptions<K, V, FC>;
 | |
|         /**
 | |
|          * Object provided in the {@link FetchOptions.context} option to
 | |
|          * {@link LRUCache#fetch}
 | |
|          */
 | |
|         context: FC;
 | |
|     }
 | |
|     /**
 | |
|      * Occasionally, it may be useful to track the internal behavior of the
 | |
|      * cache, particularly for logging, debugging, or for behavior within the
 | |
|      * `fetchMethod`. To do this, you can pass a `status` object to the
 | |
|      * {@link LRUCache#fetch}, {@link LRUCache#get}, {@link LRUCache#set},
 | |
|      * {@link LRUCache#memo}, and {@link LRUCache#has} methods.
 | |
|      *
 | |
|      * The `status` option should be a plain JavaScript object. The following
 | |
|      * fields will be set on it appropriately, depending on the situation.
 | |
|      */
 | |
|     interface Status<V> {
 | |
|         /**
 | |
|          * The status of a set() operation.
 | |
|          *
 | |
|          * - add: the item was not found in the cache, and was added
 | |
|          * - update: the item was in the cache, with the same value provided
 | |
|          * - replace: the item was in the cache, and replaced
 | |
|          * - miss: the item was not added to the cache for some reason
 | |
|          */
 | |
|         set?: 'add' | 'update' | 'replace' | 'miss';
 | |
|         /**
 | |
|          * the ttl stored for the item, or undefined if ttls are not used.
 | |
|          */
 | |
|         ttl?: Milliseconds;
 | |
|         /**
 | |
|          * the start time for the item, or undefined if ttls are not used.
 | |
|          */
 | |
|         start?: Milliseconds;
 | |
|         /**
 | |
|          * The timestamp used for TTL calculation
 | |
|          */
 | |
|         now?: Milliseconds;
 | |
|         /**
 | |
|          * the remaining ttl for the item, or undefined if ttls are not used.
 | |
|          */
 | |
|         remainingTTL?: Milliseconds;
 | |
|         /**
 | |
|          * The calculated size for the item, if sizes are used.
 | |
|          */
 | |
|         entrySize?: Size;
 | |
|         /**
 | |
|          * The total calculated size of the cache, if sizes are used.
 | |
|          */
 | |
|         totalCalculatedSize?: Size;
 | |
|         /**
 | |
|          * A flag indicating that the item was not stored, due to exceeding the
 | |
|          * {@link OptionsBase.maxEntrySize}
 | |
|          */
 | |
|         maxEntrySizeExceeded?: true;
 | |
|         /**
 | |
|          * The old value, specified in the case of `set:'update'` or
 | |
|          * `set:'replace'`
 | |
|          */
 | |
|         oldValue?: V;
 | |
|         /**
 | |
|          * The results of a {@link LRUCache#has} operation
 | |
|          *
 | |
|          * - hit: the item was found in the cache
 | |
|          * - stale: the item was found in the cache, but is stale
 | |
|          * - miss: the item was not found in the cache
 | |
|          */
 | |
|         has?: 'hit' | 'stale' | 'miss';
 | |
|         /**
 | |
|          * The status of a {@link LRUCache#fetch} operation.
 | |
|          * Note that this can change as the underlying fetch() moves through
 | |
|          * various states.
 | |
|          *
 | |
|          * - inflight: there is another fetch() for this key which is in process
 | |
|          * - get: there is no {@link OptionsBase.fetchMethod}, so
 | |
|          *   {@link LRUCache#get} was called.
 | |
|          * - miss: the item is not in cache, and will be fetched.
 | |
|          * - hit: the item is in the cache, and was resolved immediately.
 | |
|          * - stale: the item is in the cache, but stale.
 | |
|          * - refresh: the item is in the cache, and not stale, but
 | |
|          *   {@link FetchOptions.forceRefresh} was specified.
 | |
|          */
 | |
|         fetch?: 'get' | 'inflight' | 'miss' | 'hit' | 'stale' | 'refresh';
 | |
|         /**
 | |
|          * The {@link OptionsBase.fetchMethod} was called
 | |
|          */
 | |
|         fetchDispatched?: true;
 | |
|         /**
 | |
|          * The cached value was updated after a successful call to
 | |
|          * {@link OptionsBase.fetchMethod}
 | |
|          */
 | |
|         fetchUpdated?: true;
 | |
|         /**
 | |
|          * The reason for a fetch() rejection.  Either the error raised by the
 | |
|          * {@link OptionsBase.fetchMethod}, or the reason for an
 | |
|          * AbortSignal.
 | |
|          */
 | |
|         fetchError?: Error;
 | |
|         /**
 | |
|          * The fetch received an abort signal
 | |
|          */
 | |
|         fetchAborted?: true;
 | |
|         /**
 | |
|          * The abort signal received was ignored, and the fetch was allowed to
 | |
|          * continue.
 | |
|          */
 | |
|         fetchAbortIgnored?: true;
 | |
|         /**
 | |
|          * The fetchMethod promise resolved successfully
 | |
|          */
 | |
|         fetchResolved?: true;
 | |
|         /**
 | |
|          * The fetchMethod promise was rejected
 | |
|          */
 | |
|         fetchRejected?: true;
 | |
|         /**
 | |
|          * The status of a {@link LRUCache#get} operation.
 | |
|          *
 | |
|          * - fetching: The item is currently being fetched.  If a previous value
 | |
|          *   is present and allowed, that will be returned.
 | |
|          * - stale: The item is in the cache, and is stale.
 | |
|          * - hit: the item is in the cache
 | |
|          * - miss: the item is not in the cache
 | |
|          */
 | |
|         get?: 'stale' | 'hit' | 'miss';
 | |
|         /**
 | |
|          * A fetch or get operation returned a stale value.
 | |
|          */
 | |
|         returnedStale?: true;
 | |
|     }
 | |
|     /**
 | |
|      * options which override the options set in the LRUCache constructor
 | |
|      * when calling {@link LRUCache#fetch}.
 | |
|      *
 | |
|      * This is the union of {@link GetOptions} and {@link SetOptions}, plus
 | |
|      * {@link OptionsBase.noDeleteOnFetchRejection},
 | |
|      * {@link OptionsBase.allowStaleOnFetchRejection},
 | |
|      * {@link FetchOptions.forceRefresh}, and
 | |
|      * {@link FetcherOptions.context}
 | |
|      *
 | |
|      * Any of these may be modified in the {@link OptionsBase.fetchMethod}
 | |
|      * function, but the {@link GetOptions} fields will of course have no
 | |
|      * effect, as the {@link LRUCache#get} call already happened by the time
 | |
|      * the fetchMethod is called.
 | |
|      */
 | |
|     interface FetcherFetchOptions<K, V, FC = unknown> extends Pick<OptionsBase<K, V, FC>, 'allowStale' | 'updateAgeOnGet' | 'noDeleteOnStaleGet' | 'sizeCalculation' | 'ttl' | 'noDisposeOnSet' | 'noUpdateTTL' | 'noDeleteOnFetchRejection' | 'allowStaleOnFetchRejection' | 'ignoreFetchAbort' | 'allowStaleOnFetchAbort'> {
 | |
|         status?: Status<V>;
 | |
|         size?: Size;
 | |
|     }
 | |
|     /**
 | |
|      * Options that may be passed to the {@link LRUCache#fetch} method.
 | |
|      */
 | |
|     interface FetchOptions<K, V, FC> extends FetcherFetchOptions<K, V, FC> {
 | |
|         /**
 | |
|          * Set to true to force a re-load of the existing data, even if it
 | |
|          * is not yet stale.
 | |
|          */
 | |
|         forceRefresh?: boolean;
 | |
|         /**
 | |
|          * Context provided to the {@link OptionsBase.fetchMethod} as
 | |
|          * the {@link FetcherOptions.context} param.
 | |
|          *
 | |
|          * If the FC type is specified as unknown (the default),
 | |
|          * undefined or void, then this is optional.  Otherwise, it will
 | |
|          * be required.
 | |
|          */
 | |
|         context?: FC;
 | |
|         signal?: AbortSignal;
 | |
|         status?: Status<V>;
 | |
|     }
 | |
|     /**
 | |
|      * Options provided to {@link LRUCache#fetch} when the FC type is something
 | |
|      * other than `unknown`, `undefined`, or `void`
 | |
|      */
 | |
|     interface FetchOptionsWithContext<K, V, FC> extends FetchOptions<K, V, FC> {
 | |
|         context: FC;
 | |
|     }
 | |
|     /**
 | |
|      * Options provided to {@link LRUCache#fetch} when the FC type is
 | |
|      * `undefined` or `void`
 | |
|      */
 | |
|     interface FetchOptionsNoContext<K, V> extends FetchOptions<K, V, undefined> {
 | |
|         context?: undefined;
 | |
|     }
 | |
|     interface MemoOptions<K, V, FC = unknown> extends Pick<OptionsBase<K, V, FC>, 'allowStale' | 'updateAgeOnGet' | 'noDeleteOnStaleGet' | 'sizeCalculation' | 'ttl' | 'noDisposeOnSet' | 'noUpdateTTL' | 'noDeleteOnFetchRejection' | 'allowStaleOnFetchRejection' | 'ignoreFetchAbort' | 'allowStaleOnFetchAbort'> {
 | |
|         /**
 | |
|          * Set to true to force a re-load of the existing data, even if it
 | |
|          * is not yet stale.
 | |
|          */
 | |
|         forceRefresh?: boolean;
 | |
|         /**
 | |
|          * Context provided to the {@link OptionsBase.memoMethod} as
 | |
|          * the {@link MemoizerOptions.context} param.
 | |
|          *
 | |
|          * If the FC type is specified as unknown (the default),
 | |
|          * undefined or void, then this is optional.  Otherwise, it will
 | |
|          * be required.
 | |
|          */
 | |
|         context?: FC;
 | |
|         status?: Status<V>;
 | |
|     }
 | |
|     /**
 | |
|      * Options provided to {@link LRUCache#memo} when the FC type is something
 | |
|      * other than `unknown`, `undefined`, or `void`
 | |
|      */
 | |
|     interface MemoOptionsWithContext<K, V, FC> extends MemoOptions<K, V, FC> {
 | |
|         context: FC;
 | |
|     }
 | |
|     /**
 | |
|      * Options provided to {@link LRUCache#memo} when the FC type is
 | |
|      * `undefined` or `void`
 | |
|      */
 | |
|     interface MemoOptionsNoContext<K, V> extends MemoOptions<K, V, undefined> {
 | |
|         context?: undefined;
 | |
|     }
 | |
|     /**
 | |
|      * Options provided to the
 | |
|      * {@link OptionsBase.memoMethod} function.
 | |
|      */
 | |
|     interface MemoizerOptions<K, V, FC = unknown> {
 | |
|         options: MemoizerMemoOptions<K, V, FC>;
 | |
|         /**
 | |
|          * Object provided in the {@link MemoOptions.context} option to
 | |
|          * {@link LRUCache#memo}
 | |
|          */
 | |
|         context: FC;
 | |
|     }
 | |
|     /**
 | |
|      * options which override the options set in the LRUCache constructor
 | |
|      * when calling {@link LRUCache#memo}.
 | |
|      *
 | |
|      * This is the union of {@link GetOptions} and {@link SetOptions}, plus
 | |
|      * {@link MemoOptions.forceRefresh}, and
 | |
|      * {@link MemoerOptions.context}
 | |
|      *
 | |
|      * Any of these may be modified in the {@link OptionsBase.memoMethod}
 | |
|      * function, but the {@link GetOptions} fields will of course have no
 | |
|      * effect, as the {@link LRUCache#get} call already happened by the time
 | |
|      * the memoMethod is called.
 | |
|      */
 | |
|     interface MemoizerMemoOptions<K, V, FC = unknown> extends Pick<OptionsBase<K, V, FC>, 'allowStale' | 'updateAgeOnGet' | 'noDeleteOnStaleGet' | 'sizeCalculation' | 'ttl' | 'noDisposeOnSet' | 'noUpdateTTL'> {
 | |
|         status?: Status<V>;
 | |
|         size?: Size;
 | |
|         start?: Milliseconds;
 | |
|     }
 | |
|     /**
 | |
|      * Options that may be passed to the {@link LRUCache#has} method.
 | |
|      */
 | |
|     interface HasOptions<K, V, FC> extends Pick<OptionsBase<K, V, FC>, 'updateAgeOnHas'> {
 | |
|         status?: Status<V>;
 | |
|     }
 | |
|     /**
 | |
|      * Options that may be passed to the {@link LRUCache#get} method.
 | |
|      */
 | |
|     interface GetOptions<K, V, FC> extends Pick<OptionsBase<K, V, FC>, 'allowStale' | 'updateAgeOnGet' | 'noDeleteOnStaleGet'> {
 | |
|         status?: Status<V>;
 | |
|     }
 | |
|     /**
 | |
|      * Options that may be passed to the {@link LRUCache#peek} method.
 | |
|      */
 | |
|     interface PeekOptions<K, V, FC> extends Pick<OptionsBase<K, V, FC>, 'allowStale'> {
 | |
|     }
 | |
|     /**
 | |
|      * Options that may be passed to the {@link LRUCache#set} method.
 | |
|      */
 | |
|     interface SetOptions<K, V, FC> extends Pick<OptionsBase<K, V, FC>, 'sizeCalculation' | 'ttl' | 'noDisposeOnSet' | 'noUpdateTTL'> {
 | |
|         /**
 | |
|          * If size tracking is enabled, then setting an explicit size
 | |
|          * in the {@link LRUCache#set} call will prevent calling the
 | |
|          * {@link OptionsBase.sizeCalculation} function.
 | |
|          */
 | |
|         size?: Size;
 | |
|         /**
 | |
|          * If TTL tracking is enabled, then setting an explicit start
 | |
|          * time in the {@link LRUCache#set} call will override the
 | |
|          * default time from `performance.now()` or `Date.now()`.
 | |
|          *
 | |
|          * Note that it must be a valid value for whichever time-tracking
 | |
|          * method is in use.
 | |
|          */
 | |
|         start?: Milliseconds;
 | |
|         status?: Status<V>;
 | |
|     }
 | |
|     /**
 | |
|      * The type signature for the {@link OptionsBase.fetchMethod} option.
 | |
|      */
 | |
|     type Fetcher<K, V, FC = unknown> = (key: K, staleValue: V | undefined, options: FetcherOptions<K, V, FC>) => Promise<V | undefined | void> | V | undefined | void;
 | |
|     /**
 | |
|      * the type signature for the {@link OptionsBase.memoMethod} option.
 | |
|      */
 | |
|     type Memoizer<K, V, FC = unknown> = (key: K, staleValue: V | undefined, options: MemoizerOptions<K, V, FC>) => V;
 | |
|     /**
 | |
|      * Options which may be passed to the {@link LRUCache} constructor.
 | |
|      *
 | |
|      * Most of these may be overridden in the various options that use
 | |
|      * them.
 | |
|      *
 | |
|      * Despite all being technically optional, the constructor requires that
 | |
|      * a cache is at minimum limited by one or more of {@link OptionsBase.max},
 | |
|      * {@link OptionsBase.ttl}, or {@link OptionsBase.maxSize}.
 | |
|      *
 | |
|      * If {@link OptionsBase.ttl} is used alone, then it is strongly advised
 | |
|      * (and in fact required by the type definitions here) that the cache
 | |
|      * also set {@link OptionsBase.ttlAutopurge}, to prevent potentially
 | |
|      * unbounded storage.
 | |
|      *
 | |
|      * All options are also available on the {@link LRUCache} instance, making
 | |
|      * it safe to pass an LRUCache instance as the options argumemnt to
 | |
|      * make another empty cache of the same type.
 | |
|      *
 | |
|      * Some options are marked as read-only, because changing them after
 | |
|      * instantiation is not safe. Changing any of the other options will of
 | |
|      * course only have an effect on subsequent method calls.
 | |
|      */
 | |
|     interface OptionsBase<K, V, FC> {
 | |
|         /**
 | |
|          * The maximum number of items to store in the cache before evicting
 | |
|          * old entries. This is read-only on the {@link LRUCache} instance,
 | |
|          * and may not be overridden.
 | |
|          *
 | |
|          * If set, then storage space will be pre-allocated at construction
 | |
|          * time, and the cache will perform significantly faster.
 | |
|          *
 | |
|          * Note that significantly fewer items may be stored, if
 | |
|          * {@link OptionsBase.maxSize} and/or {@link OptionsBase.ttl} are also
 | |
|          * set.
 | |
|          *
 | |
|          * **It is strongly recommended to set a `max` to prevent unbounded growth
 | |
|          * of the cache.**
 | |
|          */
 | |
|         max?: Count;
 | |
|         /**
 | |
|          * Max time in milliseconds for items to live in cache before they are
 | |
|          * considered stale.  Note that stale items are NOT preemptively removed by
 | |
|          * default, and MAY live in the cache, contributing to its LRU max, long
 | |
|          * after they have expired, unless {@link OptionsBase.ttlAutopurge} is
 | |
|          * set.
 | |
|          *
 | |
|          * If set to `0` (the default value), then that means "do not track
 | |
|          * TTL", not "expire immediately".
 | |
|          *
 | |
|          * Also, as this cache is optimized for LRU/MRU operations, some of
 | |
|          * the staleness/TTL checks will reduce performance, as they will incur
 | |
|          * overhead by deleting items.
 | |
|          *
 | |
|          * This is not primarily a TTL cache, and does not make strong TTL
 | |
|          * guarantees. There is no pre-emptive pruning of expired items, but you
 | |
|          * _may_ set a TTL on the cache, and it will treat expired items as missing
 | |
|          * when they are fetched, and delete them.
 | |
|          *
 | |
|          * Optional, but must be a non-negative integer in ms if specified.
 | |
|          *
 | |
|          * This may be overridden by passing an options object to `cache.set()`.
 | |
|          *
 | |
|          * At least one of `max`, `maxSize`, or `TTL` is required. This must be a
 | |
|          * positive integer if set.
 | |
|          *
 | |
|          * Even if ttl tracking is enabled, **it is strongly recommended to set a
 | |
|          * `max` to prevent unbounded growth of the cache.**
 | |
|          *
 | |
|          * If ttl tracking is enabled, and `max` and `maxSize` are not set,
 | |
|          * and `ttlAutopurge` is not set, then a warning will be emitted
 | |
|          * cautioning about the potential for unbounded memory consumption.
 | |
|          * (The TypeScript definitions will also discourage this.)
 | |
|          */
 | |
|         ttl?: Milliseconds;
 | |
|         /**
 | |
|          * Minimum amount of time in ms in which to check for staleness.
 | |
|          * Defaults to 1, which means that the current time is checked
 | |
|          * at most once per millisecond.
 | |
|          *
 | |
|          * Set to 0 to check the current time every time staleness is tested.
 | |
|          * (This reduces performance, and is theoretically unnecessary.)
 | |
|          *
 | |
|          * Setting this to a higher value will improve performance somewhat
 | |
|          * while using ttl tracking, albeit at the expense of keeping stale
 | |
|          * items around a bit longer than their TTLs would indicate.
 | |
|          *
 | |
|          * @default 1
 | |
|          */
 | |
|         ttlResolution?: Milliseconds;
 | |
|         /**
 | |
|          * Preemptively remove stale items from the cache.
 | |
|          *
 | |
|          * Note that this may *significantly* degrade performance, especially if
 | |
|          * the cache is storing a large number of items. It is almost always best
 | |
|          * to just leave the stale items in the cache, and let them fall out as new
 | |
|          * items are added.
 | |
|          *
 | |
|          * Note that this means that {@link OptionsBase.allowStale} is a bit
 | |
|          * pointless, as stale items will be deleted almost as soon as they
 | |
|          * expire.
 | |
|          *
 | |
|          * Use with caution!
 | |
|          */
 | |
|         ttlAutopurge?: boolean;
 | |
|         /**
 | |
|          * When using time-expiring entries with `ttl`, setting this to `true` will
 | |
|          * make each item's age reset to 0 whenever it is retrieved from cache with
 | |
|          * {@link LRUCache#get}, causing it to not expire. (It can still fall out
 | |
|          * of cache based on recency of use, of course.)
 | |
|          *
 | |
|          * Has no effect if {@link OptionsBase.ttl} is not set.
 | |
|          *
 | |
|          * This may be overridden by passing an options object to `cache.get()`.
 | |
|          */
 | |
|         updateAgeOnGet?: boolean;
 | |
|         /**
 | |
|          * When using time-expiring entries with `ttl`, setting this to `true` will
 | |
|          * make each item's age reset to 0 whenever its presence in the cache is
 | |
|          * checked with {@link LRUCache#has}, causing it to not expire. (It can
 | |
|          * still fall out of cache based on recency of use, of course.)
 | |
|          *
 | |
|          * Has no effect if {@link OptionsBase.ttl} is not set.
 | |
|          */
 | |
|         updateAgeOnHas?: boolean;
 | |
|         /**
 | |
|          * Allow {@link LRUCache#get} and {@link LRUCache#fetch} calls to return
 | |
|          * stale data, if available.
 | |
|          *
 | |
|          * By default, if you set `ttl`, stale items will only be deleted from the
 | |
|          * cache when you `get(key)`. That is, it's not preemptively pruning items,
 | |
|          * unless {@link OptionsBase.ttlAutopurge} is set.
 | |
|          *
 | |
|          * If you set `allowStale:true`, it'll return the stale value *as well as*
 | |
|          * deleting it. If you don't set this, then it'll return `undefined` when
 | |
|          * you try to get a stale entry.
 | |
|          *
 | |
|          * Note that when a stale entry is fetched, _even if it is returned due to
 | |
|          * `allowStale` being set_, it is removed from the cache immediately. You
 | |
|          * can suppress this behavior by setting
 | |
|          * {@link OptionsBase.noDeleteOnStaleGet}, either in the constructor, or in
 | |
|          * the options provided to {@link LRUCache#get}.
 | |
|          *
 | |
|          * This may be overridden by passing an options object to `cache.get()`.
 | |
|          * The `cache.has()` method will always return `false` for stale items.
 | |
|          *
 | |
|          * Only relevant if a ttl is set.
 | |
|          */
 | |
|         allowStale?: boolean;
 | |
|         /**
 | |
|          * Function that is called on items when they are dropped from the
 | |
|          * cache, as `dispose(value, key, reason)`.
 | |
|          *
 | |
|          * This can be handy if you want to close file descriptors or do
 | |
|          * other cleanup tasks when items are no longer stored in the cache.
 | |
|          *
 | |
|          * **NOTE**: It is called _before_ the item has been fully removed
 | |
|          * from the cache, so if you want to put it right back in, you need
 | |
|          * to wait until the next tick. If you try to add it back in during
 | |
|          * the `dispose()` function call, it will break things in subtle and
 | |
|          * weird ways.
 | |
|          *
 | |
|          * Unlike several other options, this may _not_ be overridden by
 | |
|          * passing an option to `set()`, for performance reasons.
 | |
|          *
 | |
|          * The `reason` will be one of the following strings, corresponding
 | |
|          * to the reason for the item's deletion:
 | |
|          *
 | |
|          * - `evict` Item was evicted to make space for a new addition
 | |
|          * - `set` Item was overwritten by a new value
 | |
|          * - `expire` Item expired its TTL
 | |
|          * - `fetch` Item was deleted due to a failed or aborted fetch, or a
 | |
|          *   fetchMethod returning `undefined.
 | |
|          * - `delete` Item was removed by explicit `cache.delete(key)`,
 | |
|          *   `cache.clear()`, or `cache.set(key, undefined)`.
 | |
|          */
 | |
|         dispose?: Disposer<K, V>;
 | |
|         /**
 | |
|          * The same as {@link OptionsBase.dispose}, but called *after* the entry
 | |
|          * is completely removed and the cache is once again in a clean state.
 | |
|          *
 | |
|          * It is safe to add an item right back into the cache at this point.
 | |
|          * However, note that it is *very* easy to inadvertently create infinite
 | |
|          * recursion this way.
 | |
|          */
 | |
|         disposeAfter?: Disposer<K, V>;
 | |
|         /**
 | |
|          * Set to true to suppress calling the
 | |
|          * {@link OptionsBase.dispose} function if the entry key is
 | |
|          * still accessible within the cache.
 | |
|          *
 | |
|          * This may be overridden by passing an options object to
 | |
|          * {@link LRUCache#set}.
 | |
|          *
 | |
|          * Only relevant if `dispose` or `disposeAfter` are set.
 | |
|          */
 | |
|         noDisposeOnSet?: boolean;
 | |
|         /**
 | |
|          * Boolean flag to tell the cache to not update the TTL when setting a new
 | |
|          * value for an existing key (ie, when updating a value rather than
 | |
|          * inserting a new value).  Note that the TTL value is _always_ set (if
 | |
|          * provided) when adding a new entry into the cache.
 | |
|          *
 | |
|          * Has no effect if a {@link OptionsBase.ttl} is not set.
 | |
|          *
 | |
|          * May be passed as an option to {@link LRUCache#set}.
 | |
|          */
 | |
|         noUpdateTTL?: boolean;
 | |
|         /**
 | |
|          * Set to a positive integer to track the sizes of items added to the
 | |
|          * cache, and automatically evict items in order to stay below this size.
 | |
|          * Note that this may result in fewer than `max` items being stored.
 | |
|          *
 | |
|          * Attempting to add an item to the cache whose calculated size is greater
 | |
|          * that this amount will be a no-op. The item will not be cached, and no
 | |
|          * other items will be evicted.
 | |
|          *
 | |
|          * Optional, must be a positive integer if provided.
 | |
|          *
 | |
|          * Sets `maxEntrySize` to the same value, unless a different value is
 | |
|          * provided for `maxEntrySize`.
 | |
|          *
 | |
|          * At least one of `max`, `maxSize`, or `TTL` is required. This must be a
 | |
|          * positive integer if set.
 | |
|          *
 | |
|          * Even if size tracking is enabled, **it is strongly recommended to set a
 | |
|          * `max` to prevent unbounded growth of the cache.**
 | |
|          *
 | |
|          * Note also that size tracking can negatively impact performance,
 | |
|          * though for most cases, only minimally.
 | |
|          */
 | |
|         maxSize?: Size;
 | |
|         /**
 | |
|          * The maximum allowed size for any single item in the cache.
 | |
|          *
 | |
|          * If a larger item is passed to {@link LRUCache#set} or returned by a
 | |
|          * {@link OptionsBase.fetchMethod} or {@link OptionsBase.memoMethod}, then
 | |
|          * it will not be stored in the cache.
 | |
|          *
 | |
|          * Attempting to add an item whose calculated size is greater than
 | |
|          * this amount will not cache the item or evict any old items, but
 | |
|          * WILL delete an existing value if one is already present.
 | |
|          *
 | |
|          * Optional, must be a positive integer if provided. Defaults to
 | |
|          * the value of `maxSize` if provided.
 | |
|          */
 | |
|         maxEntrySize?: Size;
 | |
|         /**
 | |
|          * A function that returns a number indicating the item's size.
 | |
|          *
 | |
|          * Requires {@link OptionsBase.maxSize} to be set.
 | |
|          *
 | |
|          * If not provided, and {@link OptionsBase.maxSize} or
 | |
|          * {@link OptionsBase.maxEntrySize} are set, then all
 | |
|          * {@link LRUCache#set} calls **must** provide an explicit
 | |
|          * {@link SetOptions.size} or sizeCalculation param.
 | |
|          */
 | |
|         sizeCalculation?: SizeCalculator<K, V>;
 | |
|         /**
 | |
|          * Method that provides the implementation for {@link LRUCache#fetch}
 | |
|          *
 | |
|          * ```ts
 | |
|          * fetchMethod(key, staleValue, { signal, options, context })
 | |
|          * ```
 | |
|          *
 | |
|          * If `fetchMethod` is not provided, then `cache.fetch(key)` is equivalent
 | |
|          * to `Promise.resolve(cache.get(key))`.
 | |
|          *
 | |
|          * If at any time, `signal.aborted` is set to `true`, or if the
 | |
|          * `signal.onabort` method is called, or if it emits an `'abort'` event
 | |
|          * which you can listen to with `addEventListener`, then that means that
 | |
|          * the fetch should be abandoned. This may be passed along to async
 | |
|          * functions aware of AbortController/AbortSignal behavior.
 | |
|          *
 | |
|          * The `fetchMethod` should **only** return `undefined` or a Promise
 | |
|          * resolving to `undefined` if the AbortController signaled an `abort`
 | |
|          * event. In all other cases, it should return or resolve to a value
 | |
|          * suitable for adding to the cache.
 | |
|          *
 | |
|          * The `options` object is a union of the options that may be provided to
 | |
|          * `set()` and `get()`. If they are modified, then that will result in
 | |
|          * modifying the settings to `cache.set()` when the value is resolved, and
 | |
|          * in the case of
 | |
|          * {@link OptionsBase.noDeleteOnFetchRejection} and
 | |
|          * {@link OptionsBase.allowStaleOnFetchRejection}, the handling of
 | |
|          * `fetchMethod` failures.
 | |
|          *
 | |
|          * For example, a DNS cache may update the TTL based on the value returned
 | |
|          * from a remote DNS server by changing `options.ttl` in the `fetchMethod`.
 | |
|          */
 | |
|         fetchMethod?: Fetcher<K, V, FC>;
 | |
|         /**
 | |
|          * Method that provides the implementation for {@link LRUCache#memo}
 | |
|          */
 | |
|         memoMethod?: Memoizer<K, V, FC>;
 | |
|         /**
 | |
|          * Set to true to suppress the deletion of stale data when a
 | |
|          * {@link OptionsBase.fetchMethod} returns a rejected promise.
 | |
|          */
 | |
|         noDeleteOnFetchRejection?: boolean;
 | |
|         /**
 | |
|          * Do not delete stale items when they are retrieved with
 | |
|          * {@link LRUCache#get}.
 | |
|          *
 | |
|          * Note that the `get` return value will still be `undefined`
 | |
|          * unless {@link OptionsBase.allowStale} is true.
 | |
|          *
 | |
|          * When using time-expiring entries with `ttl`, by default stale
 | |
|          * items will be removed from the cache when the key is accessed
 | |
|          * with `cache.get()`.
 | |
|          *
 | |
|          * Setting this option will cause stale items to remain in the cache, until
 | |
|          * they are explicitly deleted with `cache.delete(key)`, or retrieved with
 | |
|          * `noDeleteOnStaleGet` set to `false`.
 | |
|          *
 | |
|          * This may be overridden by passing an options object to `cache.get()`.
 | |
|          *
 | |
|          * Only relevant if a ttl is used.
 | |
|          */
 | |
|         noDeleteOnStaleGet?: boolean;
 | |
|         /**
 | |
|          * Set to true to allow returning stale data when a
 | |
|          * {@link OptionsBase.fetchMethod} throws an error or returns a rejected
 | |
|          * promise.
 | |
|          *
 | |
|          * This differs from using {@link OptionsBase.allowStale} in that stale
 | |
|          * data will ONLY be returned in the case that the {@link LRUCache#fetch}
 | |
|          * fails, not any other times.
 | |
|          *
 | |
|          * If a `fetchMethod` fails, and there is no stale value available, the
 | |
|          * `fetch()` will resolve to `undefined`. Ie, all `fetchMethod` errors are
 | |
|          * suppressed.
 | |
|          *
 | |
|          * Implies `noDeleteOnFetchRejection`.
 | |
|          *
 | |
|          * This may be set in calls to `fetch()`, or defaulted on the constructor,
 | |
|          * or overridden by modifying the options object in the `fetchMethod`.
 | |
|          */
 | |
|         allowStaleOnFetchRejection?: boolean;
 | |
|         /**
 | |
|          * Set to true to return a stale value from the cache when the
 | |
|          * `AbortSignal` passed to the {@link OptionsBase.fetchMethod} dispatches
 | |
|          * an `'abort'` event, whether user-triggered, or due to internal cache
 | |
|          * behavior.
 | |
|          *
 | |
|          * Unless {@link OptionsBase.ignoreFetchAbort} is also set, the underlying
 | |
|          * {@link OptionsBase.fetchMethod} will still be considered canceled, and
 | |
|          * any value it returns will be ignored and not cached.
 | |
|          *
 | |
|          * Caveat: since fetches are aborted when a new value is explicitly
 | |
|          * set in the cache, this can lead to fetch returning a stale value,
 | |
|          * since that was the fallback value _at the moment the `fetch()` was
 | |
|          * initiated_, even though the new updated value is now present in
 | |
|          * the cache.
 | |
|          *
 | |
|          * For example:
 | |
|          *
 | |
|          * ```ts
 | |
|          * const cache = new LRUCache<string, any>({
 | |
|          *   ttl: 100,
 | |
|          *   fetchMethod: async (url, oldValue, { signal }) =>  {
 | |
|          *     const res = await fetch(url, { signal })
 | |
|          *     return await res.json()
 | |
|          *   }
 | |
|          * })
 | |
|          * cache.set('https://example.com/', { some: 'data' })
 | |
|          * // 100ms go by...
 | |
|          * const result = cache.fetch('https://example.com/')
 | |
|          * cache.set('https://example.com/', { other: 'thing' })
 | |
|          * console.log(await result) // { some: 'data' }
 | |
|          * console.log(cache.get('https://example.com/')) // { other: 'thing' }
 | |
|          * ```
 | |
|          */
 | |
|         allowStaleOnFetchAbort?: boolean;
 | |
|         /**
 | |
|          * Set to true to ignore the `abort` event emitted by the `AbortSignal`
 | |
|          * object passed to {@link OptionsBase.fetchMethod}, and still cache the
 | |
|          * resulting resolution value, as long as it is not `undefined`.
 | |
|          *
 | |
|          * When used on its own, this means aborted {@link LRUCache#fetch} calls
 | |
|          * are not immediately resolved or rejected when they are aborted, and
 | |
|          * instead take the full time to await.
 | |
|          *
 | |
|          * When used with {@link OptionsBase.allowStaleOnFetchAbort}, aborted
 | |
|          * {@link LRUCache#fetch} calls will resolve immediately to their stale
 | |
|          * cached value or `undefined`, and will continue to process and eventually
 | |
|          * update the cache when they resolve, as long as the resulting value is
 | |
|          * not `undefined`, thus supporting a "return stale on timeout while
 | |
|          * refreshing" mechanism by passing `AbortSignal.timeout(n)` as the signal.
 | |
|          *
 | |
|          * For example:
 | |
|          *
 | |
|          * ```ts
 | |
|          * const c = new LRUCache({
 | |
|          *   ttl: 100,
 | |
|          *   ignoreFetchAbort: true,
 | |
|          *   allowStaleOnFetchAbort: true,
 | |
|          *   fetchMethod: async (key, oldValue, { signal }) => {
 | |
|          *     // note: do NOT pass the signal to fetch()!
 | |
|          *     // let's say this fetch can take a long time.
 | |
|          *     const res = await fetch(`https://slow-backend-server/${key}`)
 | |
|          *     return await res.json()
 | |
|          *   },
 | |
|          * })
 | |
|          *
 | |
|          * // this will return the stale value after 100ms, while still
 | |
|          * // updating in the background for next time.
 | |
|          * const val = await c.fetch('key', { signal: AbortSignal.timeout(100) })
 | |
|          * ```
 | |
|          *
 | |
|          * **Note**: regardless of this setting, an `abort` event _is still
 | |
|          * emitted on the `AbortSignal` object_, so may result in invalid results
 | |
|          * when passed to other underlying APIs that use AbortSignals.
 | |
|          *
 | |
|          * This may be overridden in the {@link OptionsBase.fetchMethod} or the
 | |
|          * call to {@link LRUCache#fetch}.
 | |
|          */
 | |
|         ignoreFetchAbort?: boolean;
 | |
|     }
 | |
|     interface OptionsMaxLimit<K, V, FC> extends OptionsBase<K, V, FC> {
 | |
|         max: Count;
 | |
|     }
 | |
|     interface OptionsTTLLimit<K, V, FC> extends OptionsBase<K, V, FC> {
 | |
|         ttl: Milliseconds;
 | |
|         ttlAutopurge: boolean;
 | |
|     }
 | |
|     interface OptionsSizeLimit<K, V, FC> extends OptionsBase<K, V, FC> {
 | |
|         maxSize: Size;
 | |
|     }
 | |
|     /**
 | |
|      * The valid safe options for the {@link LRUCache} constructor
 | |
|      */
 | |
|     type Options<K, V, FC> = OptionsMaxLimit<K, V, FC> | OptionsSizeLimit<K, V, FC> | OptionsTTLLimit<K, V, FC>;
 | |
|     /**
 | |
|      * Entry objects used by {@link LRUCache#load} and {@link LRUCache#dump},
 | |
|      * and returned by {@link LRUCache#info}.
 | |
|      */
 | |
|     interface Entry<V> {
 | |
|         value: V;
 | |
|         ttl?: Milliseconds;
 | |
|         size?: Size;
 | |
|         start?: Milliseconds;
 | |
|     }
 | |
| }
 | |
| /**
 | |
|  * Default export, the thing you're using this module to get.
 | |
|  *
 | |
|  * The `K` and `V` types define the key and value types, respectively. The
 | |
|  * optional `FC` type defines the type of the `context` object passed to
 | |
|  * `cache.fetch()` and `cache.memo()`.
 | |
|  *
 | |
|  * Keys and values **must not** be `null` or `undefined`.
 | |
|  *
 | |
|  * All properties from the options object (with the exception of `max`,
 | |
|  * `maxSize`, `fetchMethod`, `memoMethod`, `dispose` and `disposeAfter`) are
 | |
|  * added as normal public members. (The listed options are read-only getters.)
 | |
|  *
 | |
|  * Changing any of these will alter the defaults for subsequent method calls.
 | |
|  */
 | |
| export declare class LRUCache<K extends {}, V extends {}, FC = unknown> implements Map<K, V> {
 | |
|     #private;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.ttl}
 | |
|      */
 | |
|     ttl: LRUCache.Milliseconds;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.ttlResolution}
 | |
|      */
 | |
|     ttlResolution: LRUCache.Milliseconds;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.ttlAutopurge}
 | |
|      */
 | |
|     ttlAutopurge: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.updateAgeOnGet}
 | |
|      */
 | |
|     updateAgeOnGet: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.updateAgeOnHas}
 | |
|      */
 | |
|     updateAgeOnHas: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.allowStale}
 | |
|      */
 | |
|     allowStale: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.noDisposeOnSet}
 | |
|      */
 | |
|     noDisposeOnSet: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.noUpdateTTL}
 | |
|      */
 | |
|     noUpdateTTL: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.maxEntrySize}
 | |
|      */
 | |
|     maxEntrySize: LRUCache.Size;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.sizeCalculation}
 | |
|      */
 | |
|     sizeCalculation?: LRUCache.SizeCalculator<K, V>;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.noDeleteOnFetchRejection}
 | |
|      */
 | |
|     noDeleteOnFetchRejection: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.noDeleteOnStaleGet}
 | |
|      */
 | |
|     noDeleteOnStaleGet: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.allowStaleOnFetchAbort}
 | |
|      */
 | |
|     allowStaleOnFetchAbort: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.allowStaleOnFetchRejection}
 | |
|      */
 | |
|     allowStaleOnFetchRejection: boolean;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.ignoreFetchAbort}
 | |
|      */
 | |
|     ignoreFetchAbort: boolean;
 | |
|     /**
 | |
|      * Do not call this method unless you need to inspect the
 | |
|      * inner workings of the cache.  If anything returned by this
 | |
|      * object is modified in any way, strange breakage may occur.
 | |
|      *
 | |
|      * These fields are private for a reason!
 | |
|      *
 | |
|      * @internal
 | |
|      */
 | |
|     static unsafeExposeInternals<K extends {}, V extends {}, FC extends unknown = unknown>(c: LRUCache<K, V, FC>): {
 | |
|         starts: ZeroArray | undefined;
 | |
|         ttls: ZeroArray | undefined;
 | |
|         sizes: ZeroArray | undefined;
 | |
|         keyMap: Map<K, number>;
 | |
|         keyList: (K | undefined)[];
 | |
|         valList: (V | BackgroundFetch<V> | undefined)[];
 | |
|         next: NumberArray;
 | |
|         prev: NumberArray;
 | |
|         readonly head: Index;
 | |
|         readonly tail: Index;
 | |
|         free: StackLike;
 | |
|         isBackgroundFetch: (p: any) => boolean;
 | |
|         backgroundFetch: (k: K, index: number | undefined, options: LRUCache.FetchOptions<K, V, FC>, context: any) => BackgroundFetch<V>;
 | |
|         moveToTail: (index: number) => void;
 | |
|         indexes: (options?: {
 | |
|             allowStale: boolean;
 | |
|         }) => Generator<Index, void, unknown>;
 | |
|         rindexes: (options?: {
 | |
|             allowStale: boolean;
 | |
|         }) => Generator<Index, void, unknown>;
 | |
|         isStale: (index: number | undefined) => boolean;
 | |
|     };
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.max} (read-only)
 | |
|      */
 | |
|     get max(): LRUCache.Count;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.maxSize} (read-only)
 | |
|      */
 | |
|     get maxSize(): LRUCache.Count;
 | |
|     /**
 | |
|      * The total computed size of items in the cache (read-only)
 | |
|      */
 | |
|     get calculatedSize(): LRUCache.Size;
 | |
|     /**
 | |
|      * The number of items stored in the cache (read-only)
 | |
|      */
 | |
|     get size(): LRUCache.Count;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.fetchMethod} (read-only)
 | |
|      */
 | |
|     get fetchMethod(): LRUCache.Fetcher<K, V, FC> | undefined;
 | |
|     get memoMethod(): LRUCache.Memoizer<K, V, FC> | undefined;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.dispose} (read-only)
 | |
|      */
 | |
|     get dispose(): LRUCache.Disposer<K, V> | undefined;
 | |
|     /**
 | |
|      * {@link LRUCache.OptionsBase.disposeAfter} (read-only)
 | |
|      */
 | |
|     get disposeAfter(): LRUCache.Disposer<K, V> | undefined;
 | |
|     constructor(options: LRUCache.Options<K, V, FC> | LRUCache<K, V, FC>);
 | |
|     /**
 | |
|      * Return the number of ms left in the item's TTL. If item is not in cache,
 | |
|      * returns `0`. Returns `Infinity` if item is in cache without a defined TTL.
 | |
|      */
 | |
|     getRemainingTTL(key: K): number;
 | |
|     /**
 | |
|      * Return a generator yielding `[key, value]` pairs,
 | |
|      * in order from most recently used to least recently used.
 | |
|      */
 | |
|     entries(): Generator<[K, V], void, unknown>;
 | |
|     /**
 | |
|      * Inverse order version of {@link LRUCache.entries}
 | |
|      *
 | |
|      * Return a generator yielding `[key, value]` pairs,
 | |
|      * in order from least recently used to most recently used.
 | |
|      */
 | |
|     rentries(): Generator<(K | V | BackgroundFetch<V> | undefined)[], void, unknown>;
 | |
|     /**
 | |
|      * Return a generator yielding the keys in the cache,
 | |
|      * in order from most recently used to least recently used.
 | |
|      */
 | |
|     keys(): Generator<K, void, unknown>;
 | |
|     /**
 | |
|      * Inverse order version of {@link LRUCache.keys}
 | |
|      *
 | |
|      * Return a generator yielding the keys in the cache,
 | |
|      * in order from least recently used to most recently used.
 | |
|      */
 | |
|     rkeys(): Generator<K, void, unknown>;
 | |
|     /**
 | |
|      * Return a generator yielding the values in the cache,
 | |
|      * in order from most recently used to least recently used.
 | |
|      */
 | |
|     values(): Generator<V, void, unknown>;
 | |
|     /**
 | |
|      * Inverse order version of {@link LRUCache.values}
 | |
|      *
 | |
|      * Return a generator yielding the values in the cache,
 | |
|      * in order from least recently used to most recently used.
 | |
|      */
 | |
|     rvalues(): Generator<V | BackgroundFetch<V> | undefined, void, unknown>;
 | |
|     /**
 | |
|      * Iterating over the cache itself yields the same results as
 | |
|      * {@link LRUCache.entries}
 | |
|      */
 | |
|     [Symbol.iterator](): Generator<[K, V], void, unknown>;
 | |
|     /**
 | |
|      * A String value that is used in the creation of the default string
 | |
|      * description of an object. Called by the built-in method
 | |
|      * `Object.prototype.toString`.
 | |
|      */
 | |
|     [Symbol.toStringTag]: string;
 | |
|     /**
 | |
|      * Find a value for which the supplied fn method returns a truthy value,
 | |
|      * similar to `Array.find()`. fn is called as `fn(value, key, cache)`.
 | |
|      */
 | |
|     find(fn: (v: V, k: K, self: LRUCache<K, V, FC>) => boolean, getOptions?: LRUCache.GetOptions<K, V, FC>): V | undefined;
 | |
|     /**
 | |
|      * Call the supplied function on each item in the cache, in order from most
 | |
|      * recently used to least recently used.
 | |
|      *
 | |
|      * `fn` is called as `fn(value, key, cache)`.
 | |
|      *
 | |
|      * If `thisp` is provided, function will be called in the `this`-context of
 | |
|      * the provided object, or the cache if no `thisp` object is provided.
 | |
|      *
 | |
|      * Does not update age or recenty of use, or iterate over stale values.
 | |
|      */
 | |
|     forEach(fn: (v: V, k: K, self: LRUCache<K, V, FC>) => any, thisp?: any): void;
 | |
|     /**
 | |
|      * The same as {@link LRUCache.forEach} but items are iterated over in
 | |
|      * reverse order.  (ie, less recently used items are iterated over first.)
 | |
|      */
 | |
|     rforEach(fn: (v: V, k: K, self: LRUCache<K, V, FC>) => any, thisp?: any): void;
 | |
|     /**
 | |
|      * Delete any stale entries. Returns true if anything was removed,
 | |
|      * false otherwise.
 | |
|      */
 | |
|     purgeStale(): boolean;
 | |
|     /**
 | |
|      * Get the extended info about a given entry, to get its value, size, and
 | |
|      * TTL info simultaneously. Returns `undefined` if the key is not present.
 | |
|      *
 | |
|      * Unlike {@link LRUCache#dump}, which is designed to be portable and survive
 | |
|      * serialization, the `start` value is always the current timestamp, and the
 | |
|      * `ttl` is a calculated remaining time to live (negative if expired).
 | |
|      *
 | |
|      * Always returns stale values, if their info is found in the cache, so be
 | |
|      * sure to check for expirations (ie, a negative {@link LRUCache.Entry#ttl})
 | |
|      * if relevant.
 | |
|      */
 | |
|     info(key: K): LRUCache.Entry<V> | undefined;
 | |
|     /**
 | |
|      * Return an array of [key, {@link LRUCache.Entry}] tuples which can be
 | |
|      * passed to {@link LRLUCache#load}.
 | |
|      *
 | |
|      * The `start` fields are calculated relative to a portable `Date.now()`
 | |
|      * timestamp, even if `performance.now()` is available.
 | |
|      *
 | |
|      * Stale entries are always included in the `dump`, even if
 | |
|      * {@link LRUCache.OptionsBase.allowStale} is false.
 | |
|      *
 | |
|      * Note: this returns an actual array, not a generator, so it can be more
 | |
|      * easily passed around.
 | |
|      */
 | |
|     dump(): [K, LRUCache.Entry<V>][];
 | |
|     /**
 | |
|      * Reset the cache and load in the items in entries in the order listed.
 | |
|      *
 | |
|      * The shape of the resulting cache may be different if the same options are
 | |
|      * not used in both caches.
 | |
|      *
 | |
|      * The `start` fields are assumed to be calculated relative to a portable
 | |
|      * `Date.now()` timestamp, even if `performance.now()` is available.
 | |
|      */
 | |
|     load(arr: [K, LRUCache.Entry<V>][]): void;
 | |
|     /**
 | |
|      * Add a value to the cache.
 | |
|      *
 | |
|      * Note: if `undefined` is specified as a value, this is an alias for
 | |
|      * {@link LRUCache#delete}
 | |
|      *
 | |
|      * Fields on the {@link LRUCache.SetOptions} options param will override
 | |
|      * their corresponding values in the constructor options for the scope
 | |
|      * of this single `set()` operation.
 | |
|      *
 | |
|      * If `start` is provided, then that will set the effective start
 | |
|      * time for the TTL calculation. Note that this must be a previous
 | |
|      * value of `performance.now()` if supported, or a previous value of
 | |
|      * `Date.now()` if not.
 | |
|      *
 | |
|      * Options object may also include `size`, which will prevent
 | |
|      * calling the `sizeCalculation` function and just use the specified
 | |
|      * number if it is a positive integer, and `noDisposeOnSet` which
 | |
|      * will prevent calling a `dispose` function in the case of
 | |
|      * overwrites.
 | |
|      *
 | |
|      * If the `size` (or return value of `sizeCalculation`) for a given
 | |
|      * entry is greater than `maxEntrySize`, then the item will not be
 | |
|      * added to the cache.
 | |
|      *
 | |
|      * Will update the recency of the entry.
 | |
|      *
 | |
|      * If the value is `undefined`, then this is an alias for
 | |
|      * `cache.delete(key)`. `undefined` is never stored in the cache.
 | |
|      */
 | |
|     set(k: K, v: V | BackgroundFetch<V> | undefined, setOptions?: LRUCache.SetOptions<K, V, FC>): this;
 | |
|     /**
 | |
|      * Evict the least recently used item, returning its value or
 | |
|      * `undefined` if cache is empty.
 | |
|      */
 | |
|     pop(): V | undefined;
 | |
|     /**
 | |
|      * Check if a key is in the cache, without updating the recency of use.
 | |
|      * Will return false if the item is stale, even though it is technically
 | |
|      * in the cache.
 | |
|      *
 | |
|      * Check if a key is in the cache, without updating the recency of
 | |
|      * use. Age is updated if {@link LRUCache.OptionsBase.updateAgeOnHas} is set
 | |
|      * to `true` in either the options or the constructor.
 | |
|      *
 | |
|      * Will return `false` if the item is stale, even though it is technically in
 | |
|      * the cache. The difference can be determined (if it matters) by using a
 | |
|      * `status` argument, and inspecting the `has` field.
 | |
|      *
 | |
|      * Will not update item age unless
 | |
|      * {@link LRUCache.OptionsBase.updateAgeOnHas} is set.
 | |
|      */
 | |
|     has(k: K, hasOptions?: LRUCache.HasOptions<K, V, FC>): boolean;
 | |
|     /**
 | |
|      * Like {@link LRUCache#get} but doesn't update recency or delete stale
 | |
|      * items.
 | |
|      *
 | |
|      * Returns `undefined` if the item is stale, unless
 | |
|      * {@link LRUCache.OptionsBase.allowStale} is set.
 | |
|      */
 | |
|     peek(k: K, peekOptions?: LRUCache.PeekOptions<K, V, FC>): V | undefined;
 | |
|     /**
 | |
|      * Make an asynchronous cached fetch using the
 | |
|      * {@link LRUCache.OptionsBase.fetchMethod} function.
 | |
|      *
 | |
|      * If the value is in the cache and not stale, then the returned
 | |
|      * Promise resolves to the value.
 | |
|      *
 | |
|      * If not in the cache, or beyond its TTL staleness, then
 | |
|      * `fetchMethod(key, staleValue, { options, signal, context })` is
 | |
|      * called, and the value returned will be added to the cache once
 | |
|      * resolved.
 | |
|      *
 | |
|      * If called with `allowStale`, and an asynchronous fetch is
 | |
|      * currently in progress to reload a stale value, then the former
 | |
|      * stale value will be returned.
 | |
|      *
 | |
|      * If called with `forceRefresh`, then the cached item will be
 | |
|      * re-fetched, even if it is not stale. However, if `allowStale` is also
 | |
|      * set, then the old value will still be returned. This is useful
 | |
|      * in cases where you want to force a reload of a cached value. If
 | |
|      * a background fetch is already in progress, then `forceRefresh`
 | |
|      * has no effect.
 | |
|      *
 | |
|      * If multiple fetches for the same key are issued, then they will all be
 | |
|      * coalesced into a single call to fetchMethod.
 | |
|      *
 | |
|      * Note that this means that handling options such as
 | |
|      * {@link LRUCache.OptionsBase.allowStaleOnFetchAbort},
 | |
|      * {@link LRUCache.FetchOptions.signal},
 | |
|      * and {@link LRUCache.OptionsBase.allowStaleOnFetchRejection} will be
 | |
|      * determined by the FIRST fetch() call for a given key.
 | |
|      *
 | |
|      * This is a known (fixable) shortcoming which will be addresed on when
 | |
|      * someone complains about it, as the fix would involve added complexity and
 | |
|      * may not be worth the costs for this edge case.
 | |
|      *
 | |
|      * If {@link LRUCache.OptionsBase.fetchMethod} is not specified, then this is
 | |
|      * effectively an alias for `Promise.resolve(cache.get(key))`.
 | |
|      *
 | |
|      * When the fetch method resolves to a value, if the fetch has not
 | |
|      * been aborted due to deletion, eviction, or being overwritten,
 | |
|      * then it is added to the cache using the options provided.
 | |
|      *
 | |
|      * If the key is evicted or deleted before the `fetchMethod`
 | |
|      * resolves, then the AbortSignal passed to the `fetchMethod` will
 | |
|      * receive an `abort` event, and the promise returned by `fetch()`
 | |
|      * will reject with the reason for the abort.
 | |
|      *
 | |
|      * If a `signal` is passed to the `fetch()` call, then aborting the
 | |
|      * signal will abort the fetch and cause the `fetch()` promise to
 | |
|      * reject with the reason provided.
 | |
|      *
 | |
|      * **Setting `context`**
 | |
|      *
 | |
|      * If an `FC` type is set to a type other than `unknown`, `void`, or
 | |
|      * `undefined` in the {@link LRUCache} constructor, then all
 | |
|      * calls to `cache.fetch()` _must_ provide a `context` option. If
 | |
|      * set to `undefined` or `void`, then calls to fetch _must not_
 | |
|      * provide a `context` option.
 | |
|      *
 | |
|      * The `context` param allows you to provide arbitrary data that
 | |
|      * might be relevant in the course of fetching the data. It is only
 | |
|      * relevant for the course of a single `fetch()` operation, and
 | |
|      * discarded afterwards.
 | |
|      *
 | |
|      * **Note: `fetch()` calls are inflight-unique**
 | |
|      *
 | |
|      * If you call `fetch()` multiple times with the same key value,
 | |
|      * then every call after the first will resolve on the same
 | |
|      * promise<sup>1</sup>,
 | |
|      * _even if they have different settings that would otherwise change
 | |
|      * the behavior of the fetch_, such as `noDeleteOnFetchRejection`
 | |
|      * or `ignoreFetchAbort`.
 | |
|      *
 | |
|      * In most cases, this is not a problem (in fact, only fetching
 | |
|      * something once is what you probably want, if you're caching in
 | |
|      * the first place). If you are changing the fetch() options
 | |
|      * dramatically between runs, there's a good chance that you might
 | |
|      * be trying to fit divergent semantics into a single object, and
 | |
|      * would be better off with multiple cache instances.
 | |
|      *
 | |
|      * **1**: Ie, they're not the "same Promise", but they resolve at
 | |
|      * the same time, because they're both waiting on the same
 | |
|      * underlying fetchMethod response.
 | |
|      */
 | |
|     fetch(k: K, fetchOptions: unknown extends FC ? LRUCache.FetchOptions<K, V, FC> : FC extends undefined | void ? LRUCache.FetchOptionsNoContext<K, V> : LRUCache.FetchOptionsWithContext<K, V, FC>): Promise<undefined | V>;
 | |
|     fetch(k: unknown extends FC ? K : FC extends undefined | void ? K : never, fetchOptions?: unknown extends FC ? LRUCache.FetchOptions<K, V, FC> : FC extends undefined | void ? LRUCache.FetchOptionsNoContext<K, V> : never): Promise<undefined | V>;
 | |
|     /**
 | |
|      * In some cases, `cache.fetch()` may resolve to `undefined`, either because
 | |
|      * a {@link LRUCache.OptionsBase#fetchMethod} was not provided (turning
 | |
|      * `cache.fetch(k)` into just an async wrapper around `cache.get(k)`) or
 | |
|      * because `ignoreFetchAbort` was specified (either to the constructor or
 | |
|      * in the {@link LRUCache.FetchOptions}). Also, the
 | |
|      * {@link OptionsBase.fetchMethod} may return `undefined` or `void`, making
 | |
|      * the test even more complicated.
 | |
|      *
 | |
|      * Because inferring the cases where `undefined` might be returned are so
 | |
|      * cumbersome, but testing for `undefined` can also be annoying, this method
 | |
|      * can be used, which will reject if `this.fetch()` resolves to undefined.
 | |
|      */
 | |
|     forceFetch(k: K, fetchOptions: unknown extends FC ? LRUCache.FetchOptions<K, V, FC> : FC extends undefined | void ? LRUCache.FetchOptionsNoContext<K, V> : LRUCache.FetchOptionsWithContext<K, V, FC>): Promise<V>;
 | |
|     forceFetch(k: unknown extends FC ? K : FC extends undefined | void ? K : never, fetchOptions?: unknown extends FC ? LRUCache.FetchOptions<K, V, FC> : FC extends undefined | void ? LRUCache.FetchOptionsNoContext<K, V> : never): Promise<V>;
 | |
|     /**
 | |
|      * If the key is found in the cache, then this is equivalent to
 | |
|      * {@link LRUCache#get}. If not, in the cache, then calculate the value using
 | |
|      * the {@link LRUCache.OptionsBase.memoMethod}, and add it to the cache.
 | |
|      *
 | |
|      * If an `FC` type is set to a type other than `unknown`, `void`, or
 | |
|      * `undefined` in the LRUCache constructor, then all calls to `cache.memo()`
 | |
|      * _must_ provide a `context` option. If set to `undefined` or `void`, then
 | |
|      * calls to memo _must not_ provide a `context` option.
 | |
|      *
 | |
|      * The `context` param allows you to provide arbitrary data that might be
 | |
|      * relevant in the course of fetching the data. It is only relevant for the
 | |
|      * course of a single `memo()` operation, and discarded afterwards.
 | |
|      */
 | |
|     memo(k: K, memoOptions: unknown extends FC ? LRUCache.MemoOptions<K, V, FC> : FC extends undefined | void ? LRUCache.MemoOptionsNoContext<K, V> : LRUCache.MemoOptionsWithContext<K, V, FC>): V;
 | |
|     memo(k: unknown extends FC ? K : FC extends undefined | void ? K : never, memoOptions?: unknown extends FC ? LRUCache.MemoOptions<K, V, FC> : FC extends undefined | void ? LRUCache.MemoOptionsNoContext<K, V> : never): V;
 | |
|     /**
 | |
|      * Return a value from the cache. Will update the recency of the cache
 | |
|      * entry found.
 | |
|      *
 | |
|      * If the key is not found, get() will return `undefined`.
 | |
|      */
 | |
|     get(k: K, getOptions?: LRUCache.GetOptions<K, V, FC>): V | undefined;
 | |
|     /**
 | |
|      * Deletes a key out of the cache.
 | |
|      *
 | |
|      * Returns true if the key was deleted, false otherwise.
 | |
|      */
 | |
|     delete(k: K): boolean;
 | |
|     /**
 | |
|      * Clear the cache entirely, throwing away all values.
 | |
|      */
 | |
|     clear(): void;
 | |
| }
 | |
| //# sourceMappingURL=index.d.ts.map
 |