chore: format with Prettier
This commit is contained in:
@@ -85,7 +85,7 @@ export function createStreamOptimizedHeader(capacitySectors, descriptorSizeSecto
|
||||
// version
|
||||
headerBuffer.writeUInt32LE(3, 4)
|
||||
// newline, compression, markers
|
||||
const flags = 1 | 1 << 16 | 1 << 17
|
||||
const flags = 1 | (1 << 16) | (1 << 17)
|
||||
headerBuffer.writeUInt32LE(flags, 8)
|
||||
headerBuffer.writeBigUInt64LE(BigInt(capacitySectors), 12)
|
||||
const grainSizeSectors = 128
|
||||
@@ -99,8 +99,8 @@ export function createStreamOptimizedHeader(capacitySectors, descriptorSizeSecto
|
||||
headerBuffer.writeBigInt64LE(BigInt(grainDirectoryOffsetSectors), 56)
|
||||
const grainDirectoryEntries = Math.ceil(Math.ceil(capacitySectors / grainSizeSectors) / numGTEsPerGT)
|
||||
const grainTableEntries = grainDirectoryEntries * numGTEsPerGT
|
||||
const grainDirectorySizeSectors = Math.ceil(grainDirectoryEntries * 4 / SECTOR_SIZE)
|
||||
const grainTableSizeSectors = Math.ceil(grainTableEntries * 4 / SECTOR_SIZE)
|
||||
const grainDirectorySizeSectors = Math.ceil((grainDirectoryEntries * 4) / SECTOR_SIZE)
|
||||
const grainTableSizeSectors = Math.ceil((grainTableEntries * 4) / SECTOR_SIZE)
|
||||
const overheadSectors = 1 + descriptorSizeSectors + grainDirectorySizeSectors + grainTableSizeSectors
|
||||
headerBuffer.writeBigInt64LE(BigInt(overheadSectors), 64)
|
||||
// newline mangling detector
|
||||
@@ -112,6 +112,6 @@ export function createStreamOptimizedHeader(capacitySectors, descriptorSizeSecto
|
||||
grainDirectorySizeSectors,
|
||||
grainTableSizeSectors,
|
||||
grainDirectoryEntries,
|
||||
grainTableEntries
|
||||
grainTableEntries,
|
||||
}
|
||||
}
|
||||
|
||||
@@ -96,10 +96,15 @@ test('Can generate a small VMDK file', async () => {
|
||||
const blockSize = 1024 * 1024
|
||||
const b1 = Buffer.allocUnsafe(blockSize)
|
||||
const b2 = Buffer.allocUnsafe(blockSize)
|
||||
const blockGenerator = [{ lba: 0, block: b1 }, { lba: blockSize, block: b2 }]
|
||||
const blockGenerator = [
|
||||
{ lba: 0, block: b1 },
|
||||
{ lba: blockSize, block: b2 },
|
||||
]
|
||||
const fileName = 'result.vmdk'
|
||||
const geometry = { sectorsPerTrackCylinder: 63, heads: 16, cylinders: 10402 }
|
||||
const readStream = asyncIteratorToStream(await generateVmdkData(fileName, 2 * blockSize, blockSize, blockGenerator, geometry))
|
||||
const readStream = asyncIteratorToStream(
|
||||
await generateVmdkData(fileName, 2 * blockSize, blockSize, blockGenerator, geometry)
|
||||
)
|
||||
const pipe = readStream.pipe(createWriteStream(fileName))
|
||||
await fromEvent(pipe, 'finish')
|
||||
|
||||
@@ -112,7 +117,12 @@ test('Can generate a small VMDK file', async () => {
|
||||
const data = await readVmdkGrainTable(createFileAccessor(fileName))
|
||||
expect(bufferToArray(data.grainLogicalAddressList)).toEqual(expectedLBAs)
|
||||
const grainFileOffsetList = bufferToArray(data.grainFileOffsetList)
|
||||
const parser = new VMDKDirectParser(createReadStream(fileName), bufferToArray(data.grainLogicalAddressList), grainFileOffsetList, false)
|
||||
const parser = new VMDKDirectParser(
|
||||
createReadStream(fileName),
|
||||
bufferToArray(data.grainLogicalAddressList),
|
||||
grainFileOffsetList,
|
||||
false
|
||||
)
|
||||
await parser.readHeader()
|
||||
const resLbas = []
|
||||
const resBuffers = []
|
||||
|
||||
@@ -2227,7 +2227,8 @@ export const restoreBackup = (backup, sr, { generateNewMacAddresses = false, sta
|
||||
|
||||
export const deleteBackup = backup => _call('backupNg.deleteVmBackup', { id: resolveId(backup) })
|
||||
|
||||
export const deleteBackups = async backups => _call('backupNg.deleteVmBackups', { ids: backups.map(backup => resolveId(backup)) })
|
||||
export const deleteBackups = async backups =>
|
||||
_call('backupNg.deleteVmBackups', { ids: backups.map(backup => resolveId(backup)) })
|
||||
|
||||
export const createMetadataBackupJob = props =>
|
||||
_call('metadataBackup.createJob', props)
|
||||
|
||||
Reference in New Issue
Block a user