1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176
|
# Author: Martin D. Smith
"""
Utils that support fringe DataSet features.
"""
from __future__ import absolute_import
import os
import tempfile
import logging
import json
import shutil
import datetime
import pysam
from pbcore.util.Process import backticks
log = logging.getLogger(__name__)
def getTimeStampedName(mType):
"""Generate a timestamped name using the given metatype 'mType' and the
current UTC time"""
mType = mType.lower()
mType = '_'.join(mType.split('.'))
time = datetime.datetime.utcnow().strftime("%y%m%d_%H%M%S%f")[:-3]
return "{m}-{t}".format(m=mType, t=time)
def getCreatedAt():
"""Generate a CreatedAt string using the current UTC time"""
return datetime.datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S%f")[:-6]
def which(exe):
if os.path.exists(exe) and os.access(exe, os.X_OK):
return exe
path = os.getenv('PATH')
for this_path in path.split(os.path.pathsep):
this_path = os.path.join(this_path, exe)
if os.path.exists(this_path) and os.access(this_path, os.X_OK):
return this_path
return None
def consolidateXml(indset, outbam, useTmp=True, cleanup=True):
log.debug("Consolidating to {o}".format(o=outbam))
final_free_space = disk_free(os.path.split(outbam)[0])
projected_size = sum(file_size(infn)
for infn in indset.toExternalFiles())
log.debug("Projected size: {p}".format(p=projected_size))
log.debug("In place free space: {f}".format(f=final_free_space))
# give it a 5% buffer
if final_free_space < (projected_size * 1.05):
raise RuntimeError("Not enough space available in {o} to consolidate, "
"{p} required, {a} available".format(
o=os.path.split(outbam)[0],
p=projected_size * 1.05,
a=final_free_space
))
# indset is a dataset, not a filename. So we need to write the dataset out
# to a file anyway. We'll use tmp for that, but we won't go so far as to
# copy the actual resources:
tmpout = tempfile.mkdtemp(suffix="consolidate-xml")
tmp_xml = os.path.join(tmpout,
"orig.{t}.xml".format(
t=indset.__class__.__name__.lower()))
tmp_free_space = disk_free(tmpout)
# need 2x for tmp in and out, plus 10% buffer
if useTmp:
log.debug("Tmp free space (need ~2x): {f}".format(f=tmp_free_space))
if (tmp_free_space > (projected_size * 2.1)) and useTmp:
log.debug("Using tmp storage: " + tmpout)
indset.copyTo(tmp_xml)
origOutBam = outbam
outbam = os.path.join(tmpout, "outfile.bam")
else:
log.debug("Using in place storage")
indset.write(tmp_xml)
useTmp = False
_pbmergeXML(tmp_xml, outbam)
if useTmp:
shutil.copy(outbam, origOutBam)
shutil.copy(outbam + ".pbi", origOutBam + ".pbi")
if cleanup:
shutil.rmtree(tmpout)
return outbam
def _tmpFiles(inFiles, tmpout=None):
tmpInFiles = []
if tmpout is None:
tmpout = tempfile.mkdtemp(suffix="consolidation-filtration")
for i, fname in enumerate(inFiles):
newfn = _infixFname(os.path.join(tmpout, os.path.basename(fname)), i)
shutil.copy(fname, newfn)
tmpInFiles.append(newfn)
return tmpInFiles
def disk_free(path):
if path == '':
path = os.getcwd()
space = os.statvfs(path)
return space.f_bavail * space.f_frsize
def file_size(path):
return os.stat(path).st_size
def _pbindexBam(fname):
cmd = "pbindex {i}".format(i=fname)
log.info(cmd)
o, r, m = backticks(cmd)
if r != 0:
raise RuntimeError(m)
return fname + ".pbi"
def _indexBam(fname):
pysam.samtools.index(fname, catch_stdout=False)
return fname + ".bai"
def _indexFasta(fname):
pysam.samtools.faidx(fname, catch_stdout=False)
return fname + ".fai"
def _pbmergeXML(indset, outbam):
cmd = "pbmerge -o {o} {i} ".format(i=indset,
o=outbam)
log.info(cmd)
o, r, m = backticks(cmd)
if r != 0:
raise RuntimeError("Pbmerge command failed: {c}\n Message: "
"{m}".format(c=cmd, m=m))
return outbam
def _infixFname(fname, infix):
prefix, extension = os.path.splitext(fname)
return prefix + str(infix) + extension
def _earlyInfixFname(fname, infix):
path, name = os.path.split(fname)
tokens = name.split('.')
tokens.insert(1, str(infix))
return os.path.join(path, '.'.join(tokens))
def _swapPath(dest, infile):
return os.path.join(dest, os.path.split(infile)[1])
def _fileCopy(dest, infile, uuid=None):
fn = _swapPath(dest, infile)
if os.path.exists(fn):
if uuid is None:
raise IOError("File name exists in destination: "
"{f}".format(f=infile))
subdir = os.path.join(dest, uuid)
if not os.path.exists(subdir):
os.mkdir(subdir)
fn = _swapPath(subdir, fn)
shutil.copyfile(infile, fn)
assert os.path.exists(fn)
return fn
# FIXME this is *not* numerically identical to the pbbam result!
# FIXME also it would be better to use numpy array ops, but the code below
# does not work for ndarray as written.
def hash_combine_zmw(zmw):
"""
Generate a unique hash for a ZMW, for use in downsampling filter.
"""
mask = 0xFFFF
upper = (zmw >> 16) & mask
lower = zmw & mask
result = 0
result ^= upper + 0x9e3779b9 + (result << 6) + (result >> 2)
result ^= lower + 0x9e3779b9 + (result << 6) + (result >> 2)
return result
def hash_combine_zmws(zmws):
return [hash_combine_zmw(zmw) for zmw in zmws]
|