chore: format with Prettier

This commit is contained in:
Julien Fontanet
2022-01-28 15:55:02 +01:00
parent bac976144e
commit 14fcbb6b8b
3 changed files with 19 additions and 8 deletions

View File

@@ -85,7 +85,7 @@ export function createStreamOptimizedHeader(capacitySectors, descriptorSizeSecto
// version
headerBuffer.writeUInt32LE(3, 4)
// newline, compression, markers
const flags = 1 | 1 << 16 | 1 << 17
const flags = 1 | (1 << 16) | (1 << 17)
headerBuffer.writeUInt32LE(flags, 8)
headerBuffer.writeBigUInt64LE(BigInt(capacitySectors), 12)
const grainSizeSectors = 128
@@ -99,8 +99,8 @@ export function createStreamOptimizedHeader(capacitySectors, descriptorSizeSecto
headerBuffer.writeBigInt64LE(BigInt(grainDirectoryOffsetSectors), 56)
const grainDirectoryEntries = Math.ceil(Math.ceil(capacitySectors / grainSizeSectors) / numGTEsPerGT)
const grainTableEntries = grainDirectoryEntries * numGTEsPerGT
const grainDirectorySizeSectors = Math.ceil(grainDirectoryEntries * 4 / SECTOR_SIZE)
const grainTableSizeSectors = Math.ceil(grainTableEntries * 4 / SECTOR_SIZE)
const grainDirectorySizeSectors = Math.ceil((grainDirectoryEntries * 4) / SECTOR_SIZE)
const grainTableSizeSectors = Math.ceil((grainTableEntries * 4) / SECTOR_SIZE)
const overheadSectors = 1 + descriptorSizeSectors + grainDirectorySizeSectors + grainTableSizeSectors
headerBuffer.writeBigInt64LE(BigInt(overheadSectors), 64)
// newline mangling detector
@@ -112,6 +112,6 @@ export function createStreamOptimizedHeader(capacitySectors, descriptorSizeSecto
grainDirectorySizeSectors,
grainTableSizeSectors,
grainDirectoryEntries,
grainTableEntries
grainTableEntries,
}
}

View File

@@ -96,10 +96,15 @@ test('Can generate a small VMDK file', async () => {
const blockSize = 1024 * 1024
const b1 = Buffer.allocUnsafe(blockSize)
const b2 = Buffer.allocUnsafe(blockSize)
const blockGenerator = [{ lba: 0, block: b1 }, { lba: blockSize, block: b2 }]
const blockGenerator = [
{ lba: 0, block: b1 },
{ lba: blockSize, block: b2 },
]
const fileName = 'result.vmdk'
const geometry = { sectorsPerTrackCylinder: 63, heads: 16, cylinders: 10402 }
const readStream = asyncIteratorToStream(await generateVmdkData(fileName, 2 * blockSize, blockSize, blockGenerator, geometry))
const readStream = asyncIteratorToStream(
await generateVmdkData(fileName, 2 * blockSize, blockSize, blockGenerator, geometry)
)
const pipe = readStream.pipe(createWriteStream(fileName))
await fromEvent(pipe, 'finish')
@@ -112,7 +117,12 @@ test('Can generate a small VMDK file', async () => {
const data = await readVmdkGrainTable(createFileAccessor(fileName))
expect(bufferToArray(data.grainLogicalAddressList)).toEqual(expectedLBAs)
const grainFileOffsetList = bufferToArray(data.grainFileOffsetList)
const parser = new VMDKDirectParser(createReadStream(fileName), bufferToArray(data.grainLogicalAddressList), grainFileOffsetList, false)
const parser = new VMDKDirectParser(
createReadStream(fileName),
bufferToArray(data.grainLogicalAddressList),
grainFileOffsetList,
false
)
await parser.readHeader()
const resLbas = []
const resBuffers = []

View File

@@ -2227,7 +2227,8 @@ export const restoreBackup = (backup, sr, { generateNewMacAddresses = false, sta
export const deleteBackup = backup => _call('backupNg.deleteVmBackup', { id: resolveId(backup) })
export const deleteBackups = async backups => _call('backupNg.deleteVmBackups', { ids: backups.map(backup => resolveId(backup)) })
export const deleteBackups = async backups =>
_call('backupNg.deleteVmBackups', { ids: backups.map(backup => resolveId(backup)) })
export const createMetadataBackupJob = props =>
_call('metadataBackup.createJob', props)