mirror of
https://github.com/grafana/grafana.git
synced 2024-11-26 19:00:54 -06:00
Update dependency lru-cache to v10 (#71585)
* Update dependency lru-cache to v10 * use named import * remove outdated type definitions and actually use named import --------- Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com> Co-authored-by: Ashley Harrison <ashley.harrison@grafana.com>
This commit is contained in:
parent
cd1b53ead1
commit
1cacc78eda
@ -347,7 +347,7 @@
|
|||||||
"kbar": "0.1.0-beta.40",
|
"kbar": "0.1.0-beta.40",
|
||||||
"lodash": "4.17.21",
|
"lodash": "4.17.21",
|
||||||
"logfmt": "^1.3.2",
|
"logfmt": "^1.3.2",
|
||||||
"lru-cache": "7.17.0",
|
"lru-cache": "10.0.0",
|
||||||
"lru-memoize": "^1.1.0",
|
"lru-memoize": "^1.1.0",
|
||||||
"marked": "^4.3.0",
|
"marked": "^4.3.0",
|
||||||
"memoize-one": "6.0.0",
|
"memoize-one": "6.0.0",
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
import { chain, difference } from 'lodash';
|
import { chain, difference } from 'lodash';
|
||||||
import LRU from 'lru-cache';
|
import { LRUCache } from 'lru-cache';
|
||||||
import Prism, { Grammar } from 'prismjs';
|
import Prism, { Grammar } from 'prismjs';
|
||||||
|
|
||||||
import { dateTime, AbsoluteTimeRange, LanguageProvider, HistoryItem, AbstractQuery } from '@grafana/data';
|
import { dateTime, AbsoluteTimeRange, LanguageProvider, HistoryItem, AbstractQuery } from '@grafana/data';
|
||||||
@ -79,8 +79,8 @@ export default class LokiLanguageProvider extends LanguageProvider {
|
|||||||
* not account for different size of a response. If that is needed a `length` function can be added in the options.
|
* not account for different size of a response. If that is needed a `length` function can be added in the options.
|
||||||
* 10 as a max size is totally arbitrary right now.
|
* 10 as a max size is totally arbitrary right now.
|
||||||
*/
|
*/
|
||||||
private seriesCache = new LRU<string, Record<string, string[]>>({ max: 10 });
|
private seriesCache = new LRUCache<string, Record<string, string[]>>({ max: 10 });
|
||||||
private labelsCache = new LRU<string, string[]>({ max: 10 });
|
private labelsCache = new LRUCache<string, string[]>({ max: 10 });
|
||||||
|
|
||||||
constructor(datasource: LokiDatasource, initialValues?: any) {
|
constructor(datasource: LokiDatasource, initialValues?: any) {
|
||||||
super();
|
super();
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
import { cloneDeep, defaults } from 'lodash';
|
import { cloneDeep, defaults } from 'lodash';
|
||||||
import LRU from 'lru-cache';
|
import { LRUCache } from 'lru-cache';
|
||||||
import React from 'react';
|
import React from 'react';
|
||||||
import { forkJoin, lastValueFrom, merge, Observable, of, OperatorFunction, pipe, throwError } from 'rxjs';
|
import { forkJoin, lastValueFrom, merge, Observable, of, OperatorFunction, pipe, throwError } from 'rxjs';
|
||||||
import { catchError, filter, map, tap } from 'rxjs/operators';
|
import { catchError, filter, map, tap } from 'rxjs/operators';
|
||||||
@ -92,7 +92,7 @@ export class PrometheusDatasource
|
|||||||
access: 'direct' | 'proxy';
|
access: 'direct' | 'proxy';
|
||||||
basicAuth: any;
|
basicAuth: any;
|
||||||
withCredentials: any;
|
withCredentials: any;
|
||||||
metricsNameCache = new LRU<string, string[]>({ max: 10 });
|
metricsNameCache = new LRUCache<string, string[]>({ max: 10 });
|
||||||
interval: string;
|
interval: string;
|
||||||
queryTimeout: string | undefined;
|
queryTimeout: string | undefined;
|
||||||
httpMethod: string;
|
httpMethod: string;
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
import { chain, difference, once } from 'lodash';
|
import { chain, difference, once } from 'lodash';
|
||||||
import LRU from 'lru-cache';
|
import { LRUCache } from 'lru-cache';
|
||||||
import Prism from 'prismjs';
|
import Prism from 'prismjs';
|
||||||
import { Value } from 'slate';
|
import { Value } from 'slate';
|
||||||
|
|
||||||
@ -121,8 +121,8 @@ export default class PromQlLanguageProvider extends LanguageProvider {
|
|||||||
* not account for different size of a response. If that is needed a `length` function can be added in the options.
|
* not account for different size of a response. If that is needed a `length` function can be added in the options.
|
||||||
* 10 as a max size is totally arbitrary right now.
|
* 10 as a max size is totally arbitrary right now.
|
||||||
*/
|
*/
|
||||||
private labelsCache = new LRU<string, Record<string, string[]>>({ max: 10 });
|
private labelsCache = new LRUCache<string, Record<string, string[]>>({ max: 10 });
|
||||||
private labelValuesCache = new LRU<string, string[]>({ max: 10 });
|
private labelValuesCache = new LRUCache<string, string[]>({ max: 10 });
|
||||||
constructor(datasource: PrometheusDatasource, initialValues?: Partial<PromQlLanguageProvider>) {
|
constructor(datasource: PrometheusDatasource, initialValues?: Partial<PromQlLanguageProvider>) {
|
||||||
super();
|
super();
|
||||||
|
|
||||||
|
259
public/app/types/lru-cache.d.ts
vendored
259
public/app/types/lru-cache.d.ts
vendored
@ -1,259 +0,0 @@
|
|||||||
// Type definitions for lru-cache 7.1.0
|
|
||||||
// TypeScript Version: 4.5
|
|
||||||
declare class LRUCache<K, V> {
|
|
||||||
constructor(options?: LRUCache.Options<K, V>);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return total length of objects in cache taking into account `length` options function.
|
|
||||||
*/
|
|
||||||
readonly length: number;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return total quantity of objects currently in cache. Note,
|
|
||||||
* that `stale` (see options) items are returned as part of this item count.
|
|
||||||
*/
|
|
||||||
readonly itemCount: number;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Same as Options.allowStale.
|
|
||||||
*/
|
|
||||||
allowStale: boolean;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Same as Options.length.
|
|
||||||
*/
|
|
||||||
lengthCalculator(value: V): number;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Same as Options.max. Resizes the cache when the `max` changes.
|
|
||||||
*/
|
|
||||||
max: number;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Same as Options.maxAge. Resizes the cache when the `maxAge` changes.
|
|
||||||
*/
|
|
||||||
maxAge: number;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Will update the "recently used"-ness of the key. They do what you think.
|
|
||||||
* `maxAge` is optional and overrides the cache `maxAge` option if provided.
|
|
||||||
*/
|
|
||||||
set(key: K, value: V, options?: LRUCache.SetOptions<V>): boolean;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Will update the "recently used"-ness of the key. They do what you think.
|
|
||||||
* `maxAge` is optional and overrides the cache `maxAge` option if provided.
|
|
||||||
*
|
|
||||||
* If the key is not found, will return `undefined`.
|
|
||||||
*/
|
|
||||||
get(key: K): V | undefined;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Returns the key value (or `undefined` if not found) without updating
|
|
||||||
* the "recently used"-ness of the key.
|
|
||||||
*
|
|
||||||
* (If you find yourself using this a lot, you might be using the wrong
|
|
||||||
* sort of data structure, but there are some use cases where it's handy.)
|
|
||||||
*/
|
|
||||||
peek(key: K): V | undefined;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Check if a key is in the cache, without updating the recent-ness
|
|
||||||
* or deleting it for being stale.
|
|
||||||
*/
|
|
||||||
has(key: K): boolean;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Deletes a key out of the cache.
|
|
||||||
*/
|
|
||||||
del(key: K): void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Clear the cache entirely, throwing away all values.
|
|
||||||
*/
|
|
||||||
reset(): void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Manually iterates over the entire cache proactively pruning old entries.
|
|
||||||
*/
|
|
||||||
prune(): void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Just like `Array.prototype.forEach`. Iterates over all the keys in the cache,
|
|
||||||
* in order of recent-ness. (Ie, more recently used items are iterated over first.)
|
|
||||||
*/
|
|
||||||
forEach<T = this>(callbackFn: (this: T, value: V, key: K, cache: this) => void, thisArg?: T): void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The same as `cache.forEach(...)` but items are iterated over in reverse order.
|
|
||||||
* (ie, less recently used items are iterated over first.)
|
|
||||||
*/
|
|
||||||
forEach<T = this>(callbackFn: (this: T, value: V, key: K, cache: this) => void, thisArg?: T): void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return an array of the keys in the cache.
|
|
||||||
*/
|
|
||||||
keys(): K[];
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return an array of the values in the cache.
|
|
||||||
*/
|
|
||||||
values(): V[];
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return an array of the cache entries ready for serialization and usage with `destinationCache.load(arr)`.
|
|
||||||
*/
|
|
||||||
dump(): Array<LRUCache.Entry<K, V>>;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Loads another cache entries array, obtained with `sourceCache.dump()`,
|
|
||||||
* into the cache. The destination cache is reset before loading new entries
|
|
||||||
*
|
|
||||||
* @param cacheEntries Obtained from `sourceCache.dump()`
|
|
||||||
*/
|
|
||||||
load(cacheEntries: ReadonlyArray<LRUCache.Entry<K, V>>): void;
|
|
||||||
}
|
|
||||||
|
|
||||||
// eslint-disable-next-line no-redeclare
|
|
||||||
declare namespace LRUCache {
|
|
||||||
interface Options<K, V> {
|
|
||||||
/**
|
|
||||||
* @type {number | undefined}
|
|
||||||
* the number of most recently used items to keep.
|
|
||||||
* note that we may store fewer items than this if maxSize is hit.
|
|
||||||
*/
|
|
||||||
max?: number | undefined;
|
|
||||||
|
|
||||||
// if you wish to track item size, you must provide a maxSize
|
|
||||||
// note that we still will only keep up to max *actual items*,
|
|
||||||
// so size tracking may cause fewer than max items to be stored.
|
|
||||||
// At the extreme, a single item of maxSize size will cause everything
|
|
||||||
// else in the cache to be dropped when it is added. Use with caution!
|
|
||||||
// Note also that size tracking can negatively impact performance,
|
|
||||||
// though for most cases, only minimally.
|
|
||||||
maxSize?: number | undefined;
|
|
||||||
|
|
||||||
// buffers or other items where memory size depends on the object itself.
|
|
||||||
// also note that oversized items do NOT immediately get dropped from
|
|
||||||
// the cache, though they will cause faster turnover in the storage.
|
|
||||||
// Return an positive integer which is the size of the item,
|
|
||||||
// if a positive integer is not returned, will use 0 as the size.
|
|
||||||
sizeCalculation?: (value, key) => number;
|
|
||||||
|
|
||||||
// function to call when the item is removed from the cache
|
|
||||||
// Note that using this can negatively impact performance.
|
|
||||||
dispose?: (value, key) => void;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* By default, if you set a `dispose()` method, then it'll be called whenever
|
|
||||||
* a `set()` operation overwrites an existing key. If you set this option,
|
|
||||||
* `dispose()` will only be called when a key falls out of the cache,
|
|
||||||
* not when it is overwritten.
|
|
||||||
*/
|
|
||||||
noDisposeOnSet?: boolean | undefined;
|
|
||||||
|
|
||||||
// max time to live for items before they are considered stale
|
|
||||||
// note that stale items are NOT preemptively removed by default,
|
|
||||||
// and MAY live in the cache, contributing to its LRU max, long after
|
|
||||||
// they have expired.
|
|
||||||
// Also, as this cache is optimized for LRU/MRU operations, some of
|
|
||||||
// the staleness/TTL checks will reduce performance, as they will incur
|
|
||||||
// overhead by deleting items.
|
|
||||||
// Must be a positive integer in ms, defaults to 0, which means "no TTL"
|
|
||||||
ttl?: number;
|
|
||||||
|
|
||||||
// Minimum amount of time in ms in which to check for staleness.
|
|
||||||
// Defaults to 1, which means that the current time
|
|
||||||
// is checked at most once per millisecond.
|
|
||||||
// Set to 0 to check the current time every time staleness is tested.
|
|
||||||
// Note that setting this to a higher value
|
|
||||||
// will improve performance somewhat while using ttl tracking,
|
|
||||||
// albeit at the expense of keeping stale items
|
|
||||||
// around a bit longer than intended.
|
|
||||||
ttlResolution?: number;
|
|
||||||
|
|
||||||
// Preemptively remove stale items from the cache.
|
|
||||||
// Note that this may significantly degrade performance,
|
|
||||||
// especially if the cache is storing a large number of items.
|
|
||||||
// It is almost always best to just leave the stale items in the cache,
|
|
||||||
// and let them fall out as new items are added.
|
|
||||||
// Note that this means that allowStale is a bit pointless,
|
|
||||||
// as stale items will be deleted almost as soon as they expire.
|
|
||||||
// Use with caution!
|
|
||||||
ttlAutopurge?: boolean;
|
|
||||||
|
|
||||||
// By default, if you set ttl, it'll only delete stale items
|
|
||||||
// from the cache when you get(key).
|
|
||||||
// That is, it's not preemptively pruning items.
|
|
||||||
// If you set allowStale:true, it'll return the stale value
|
|
||||||
// as well as deleting it. If you don't set this,
|
|
||||||
// then it'll return undefined when you try to get a stale entry.
|
|
||||||
// Note that when a stale entry is fetched,
|
|
||||||
// even if it is returned due to allowStale being set,
|
|
||||||
// it is removed from the cache immediately.
|
|
||||||
// You can immediately put it back in the cache if you wish,
|
|
||||||
// thus resetting the TTL.
|
|
||||||
// This may be overridden by passing an options object to cache.get().
|
|
||||||
// The cache.has() method will always return false for stale items.
|
|
||||||
// Boolean, default false, only relevant if ttl is set.
|
|
||||||
allowStale?: boolean;
|
|
||||||
|
|
||||||
// When using time-expiring entries with ttl,
|
|
||||||
// setting this to true will make each item's
|
|
||||||
// age reset to 0 whenever it is retrieved from cache with get(),
|
|
||||||
// causing it to not expire.
|
|
||||||
// (It can still fall out of cache based on recency of use, of course.)
|
|
||||||
// This may be overridden by passing an options object to cache.get().
|
|
||||||
// Boolean, default false, only relevant if ttl is set.
|
|
||||||
updateAgeOnGet?: boolean;
|
|
||||||
|
|
||||||
// update the age of items on cache.has(), renewing their TTL
|
|
||||||
// boolean, default false
|
|
||||||
updateAgeOnHas?: boolean;
|
|
||||||
|
|
||||||
// update the "recently-used"-ness of items on cache.has()
|
|
||||||
// boolean, default false
|
|
||||||
updateRecencyOnHas?: boolean;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Function that is used to calculate the length of stored items.
|
|
||||||
* If you're storing strings or buffers, then you probably want to do
|
|
||||||
* something like `function(n, key){return n.length}`. The default
|
|
||||||
* is `function(){return 1}`, which is fine if you want to store
|
|
||||||
* `max` like-sized things. The item is passed as the first argument,
|
|
||||||
* and the key is passed as the second argument.
|
|
||||||
*/
|
|
||||||
length?(value: V, key?: K): number;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* By default, if you set a `maxAge`, it'll only actually pull stale items
|
|
||||||
* out of the cache when you `get(key)`. (That is, it's not pre-emptively
|
|
||||||
* doing a `setTimeout` or anything.) If you set `stale:true`, it'll return
|
|
||||||
* the stale value before deleting it. If you don't set this, then it'll
|
|
||||||
* return `undefined` when you try to get a stale entry,
|
|
||||||
* as if it had already been deleted.
|
|
||||||
*/
|
|
||||||
stale?: boolean | undefined;
|
|
||||||
}
|
|
||||||
|
|
||||||
interface SetOptions<V> {
|
|
||||||
ttl?: number;
|
|
||||||
// Will prevent calling the sizeCalculation function
|
|
||||||
// and just use the specified number if it is a positive integer
|
|
||||||
size?: number;
|
|
||||||
// Same as above
|
|
||||||
sizeCalculator?: (value: V) => number;
|
|
||||||
// Will prevent calling a dispose function in the case of overwrites
|
|
||||||
noDisposeOnSet?: boolean;
|
|
||||||
}
|
|
||||||
|
|
||||||
interface Entry<K, V> {
|
|
||||||
k: K;
|
|
||||||
v: V;
|
|
||||||
e: number;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
declare module 'lru-cache' {
|
|
||||||
export = LRUCache;
|
|
||||||
}
|
|
17
yarn.lock
17
yarn.lock
@ -19414,7 +19414,7 @@ __metadata:
|
|||||||
lint-staged: 13.2.3
|
lint-staged: 13.2.3
|
||||||
lodash: 4.17.21
|
lodash: 4.17.21
|
||||||
logfmt: ^1.3.2
|
logfmt: ^1.3.2
|
||||||
lru-cache: 7.17.0
|
lru-cache: 10.0.0
|
||||||
lru-memoize: ^1.1.0
|
lru-memoize: ^1.1.0
|
||||||
marked: ^4.3.0
|
marked: ^4.3.0
|
||||||
memoize-one: 6.0.0
|
memoize-one: 6.0.0
|
||||||
@ -23159,10 +23159,10 @@ __metadata:
|
|||||||
languageName: node
|
languageName: node
|
||||||
linkType: hard
|
linkType: hard
|
||||||
|
|
||||||
"lru-cache@npm:7.17.0":
|
"lru-cache@npm:10.0.0, lru-cache@npm:^9.1.1 || ^10.0.0":
|
||||||
version: 7.17.0
|
version: 10.0.0
|
||||||
resolution: "lru-cache@npm:7.17.0"
|
resolution: "lru-cache@npm:10.0.0"
|
||||||
checksum: 28c2a98ad313b8d61beac1f08257b6f0ca990e39d24a9bc831030b6e209447cfb11c6d9d1a774282189bfc9609d1dfd17ebe485228dd68f7b96b6b9b7740894e
|
checksum: 18f101675fe283bc09cda0ef1e3cc83781aeb8373b439f086f758d1d91b28730950db785999cd060d3c825a8571c03073e8c14512b6655af2188d623031baf50
|
||||||
languageName: node
|
languageName: node
|
||||||
linkType: hard
|
linkType: hard
|
||||||
|
|
||||||
@ -23191,13 +23191,6 @@ __metadata:
|
|||||||
languageName: node
|
languageName: node
|
||||||
linkType: hard
|
linkType: hard
|
||||||
|
|
||||||
"lru-cache@npm:^9.1.1 || ^10.0.0":
|
|
||||||
version: 10.0.0
|
|
||||||
resolution: "lru-cache@npm:10.0.0"
|
|
||||||
checksum: 18f101675fe283bc09cda0ef1e3cc83781aeb8373b439f086f758d1d91b28730950db785999cd060d3c825a8571c03073e8c14512b6655af2188d623031baf50
|
|
||||||
languageName: node
|
|
||||||
linkType: hard
|
|
||||||
|
|
||||||
"lru-memoize@npm:^1.1.0":
|
"lru-memoize@npm:^1.1.0":
|
||||||
version: 1.1.0
|
version: 1.1.0
|
||||||
resolution: "lru-memoize@npm:1.1.0"
|
resolution: "lru-memoize@npm:1.1.0"
|
||||||
|
Loading…
Reference in New Issue
Block a user