apr144-bam
Version:
Parser for BAM and BAM index (bai) files
160 lines (142 loc) • 4.08 kB
text/typescript
import Long from 'long'
import Chunk from './chunk'
import VirtualOffset from './virtualOffset'
export function timeout(ms: number) {
return new Promise(resolve => setTimeout(resolve, ms))
}
export function longToNumber(long: Long) {
if (
long.greaterThan(Number.MAX_SAFE_INTEGER) ||
long.lessThan(Number.MIN_SAFE_INTEGER)
) {
throw new Error('integer overflow')
}
return long.toNumber()
}
/**
* Properly check if the given AbortSignal is aborted.
* Per the standard, if the signal reads as aborted,
* this function throws either a DOMException AbortError, or a regular error
* with a `code` attribute set to `ERR_ABORTED`.
*
* For convenience, passing `undefined` is a no-op
*
* @param {AbortSignal} [signal] an AbortSignal, or anything with an `aborted` attribute
* @returns nothing
*/
export function checkAbortSignal(signal?: AbortSignal) {
if (!signal) {
return
}
if (signal.aborted) {
// console.log('bam aborted!')
if (typeof DOMException === 'undefined') {
const e = new Error('aborted')
//@ts-ignore
e.code = 'ERR_ABORTED'
throw e
} else {
throw new DOMException('aborted', 'AbortError')
}
}
}
/**
* Skips to the next tick, then runs `checkAbortSignal`.
* Await this to inside an otherwise synchronous loop to
* provide a place to break when an abort signal is received.
* @param {AbortSignal} signal
*/
export async function abortBreakPoint(signal?: AbortSignal) {
await Promise.resolve()
checkAbortSignal(signal)
}
export function canMergeBlocks(chunk1: Chunk, chunk2: Chunk) {
return (
chunk2.minv.blockPosition - chunk1.maxv.blockPosition < 65000 &&
chunk2.maxv.blockPosition - chunk1.minv.blockPosition < 5000000
)
}
export interface BamOpts {
assemblyName?: string
viewAsPairs?: boolean
pairAcrossChr?: boolean
maxInsertSize?: number
signal?: AbortSignal
}
export interface BaseOpts {
assemblyName?: string
signal?: AbortSignal
}
export function makeOpts(obj: AbortSignal | BaseOpts = {}): BaseOpts {
return 'aborted' in obj ? ({ signal: obj } as BaseOpts) : obj
}
export function optimizeChunks(chunks: Chunk[], lowest?: VirtualOffset) {
const mergedChunks: Chunk[] = []
let lastChunk: Chunk | undefined
if (chunks.length === 0) {
return chunks
}
chunks.sort((c0, c1) => {
const dif = c0.minv.blockPosition - c1.minv.blockPosition
return dif === 0 ? c0.minv.dataPosition - c1.minv.dataPosition : dif
})
for (const chunk of chunks) {
if (!lowest || chunk.maxv.compareTo(lowest) > 0) {
if (lastChunk === undefined) {
mergedChunks.push(chunk)
lastChunk = chunk
} else {
if (canMergeBlocks(lastChunk, chunk)) {
if (chunk.maxv.compareTo(lastChunk.maxv) > 0) {
lastChunk.maxv = chunk.maxv
}
} else {
mergedChunks.push(chunk)
lastChunk = chunk
}
}
}
}
return mergedChunks
}
export function parsePseudoBin(bytes: Buffer, offset: number) {
const lineCount = longToNumber(
Long.fromBytesLE(
Array.prototype.slice.call(bytes, offset, offset + 8),
true,
),
)
return { lineCount }
}
export function findFirstData(
firstDataLine: VirtualOffset | undefined,
virtualOffset: VirtualOffset,
) {
return firstDataLine
? firstDataLine.compareTo(virtualOffset) > 0
? virtualOffset
: firstDataLine
: virtualOffset
}
export function parseNameBytes(
namesBytes: Buffer,
renameRefSeq: (arg: string) => string = s => s,
) {
let currRefId = 0
let currNameStart = 0
const refIdToName = []
const refNameToId: Record<string, number> = {}
for (let i = 0; i < namesBytes.length; i += 1) {
if (!namesBytes[i]) {
if (currNameStart < i) {
let refName = namesBytes.toString('utf8', currNameStart, i)
refName = renameRefSeq(refName)
refIdToName[currRefId] = refName
refNameToId[refName] = currRefId
}
currNameStart = i + 1
currRefId += 1
}
}
return { refNameToId, refIdToName }
}