Kaynağa Gözat

Initial import

Andrej Studen 2 ay önce
işleme
e9a122d0e6
2 değiştirilmiş dosya ile 235 ekleme ve 0 silme
  1. 6 0
      slicerScripts/config.json
  2. 229 0
      slicerScripts/parseDICOM.py

+ 6 - 0
slicerScripts/config.json

@@ -0,0 +1,6 @@
+{"project":"DCIS/Study",
+	"server":"onko-nix",
+	"schema":"study",
+	"query":"Imaging1",
+	"baseDir":"/home/studen/temp",
+"debug":0}

+ 229 - 0
slicerScripts/parseDICOM.py

@@ -0,0 +1,229 @@
+import DICOMLib
+import sys
+import json
+import numpy
+import zipfile
+import shutil
+import os
+
+def main(configFile=None):
+
+   
+
+   print('Imported!')
+   with open(configFile) as f:
+      config=json.load(f)
+
+   config.update(connectDB(config))
+   parseData(config,getMeanHeartDose)
+
+def connectDB(setup):
+   nixSuite=os.path.join(os.path.expanduser('~'),'software','src','nixsuite')
+   sys.path.append(os.path.join(nixSuite,'wrapper'))
+   import nixWrapper
+   nixWrapper.loadLibrary('labkeyInterface')
+   import labkeyInterface
+   import labkeyDatabaseBrowser
+   import labkeyFileBrowser
+
+   nixWrapper.loadLibrary('orthancInterface')
+   import orthancInterface
+   import orthancDatabaseBrowser
+   import orthancFileBrowser
+   importlib.reload(orthancFileBrowser)
+
+   net=labkeyInterface.labkeyInterface()
+   qfile='{}.json'.format(setup['server'])
+   fconfig=os.path.join(os.path.expanduser('~'),'.labkey',qfile)
+   net.init(fconfig)
+   net.getCSRF()
+
+   onet=orthancInterface.orthancInterface()
+   onet.init(fconfig)
+
+   return {"db":labkeyDatabaseBrowser.labkeyDB(net),
+      "fb":labkeyFileBrowser.labkeyFileBrowser(net),
+      "odb":orthancDatabaseBrowser.orthancDB(onet),
+      "ofb":orthancFileBrowser.orthancFileBrowser(onet)}
+
+
+#explicit template
+def _updateRow(config,r):
+   print(r)
+   return False
+
+def parseData(config,updateRow=_updateRow):
+#iterates over data appliying updateRow function to every row
+#updateRow is an implementation of a generic function 
+#with arguments
+#   def updateRow(config,r)
+#returning True if row needs to be updated on the server
+#and False otherwise
+#update values are stored in the r dict
+
+   db=config['db']
+    
+   qFilter=config.get('qFilter',[])
+   debug=config.get('debug',False)
+#get dataset 
+   ds=db.selectRows(config['project'],config['schema'],config['query'],qFilter)
+   rows=ds['rows']
+#shorten list in debug mode
+   if debug:
+      rows=rows[0:3]
+   for r in rows:
+#this could be a generic function either as parameter of config or an argument to parseData
+      update=updateRow(config,r)
+      #print(r)
+      if not update:
+         continue
+      db.modifyRows('update',config['project'],config['schema'],config['query'],[r])
+
+
+
+def getMeanHeartDose(config,r):
+#calculates mean heart dose
+#stores it as doseHeart to row r
+#return True if r needs to be updated on the server
+#and False if r is unchanged
+   sid=r['orthancStudyId']
+   if not sid:
+      print('No study for {}'.format(r['ParticipantId']))
+      return False
+   doseHeart=r['doseHeart']
+   if doseHeart:
+#no need to update
+      return False
+
+   #outDir=getDicomZip(config,sid)
+   outDir=getDicomInstances(config,sid)
+
+   nodes=loadVolumes(outDir)
+   msg=checkNodes(config,nodes)
+   if len(msg)>0:
+      r['comments']=msg
+   else:
+      r['doseHeart']=getMeanDose(nodes,'Heart')
+   clearDir(outDir)
+#needs updating
+   return True
+
+def loadVolumes(dataDir):
+   nodeNames=[]
+   with DICOMLib.DICOMUtils.TemporaryDICOMDatabase() as db:
+      DICOMLib.DICOMUtils.importDicom(dataDir, db)
+      patientUIDs = db.patients()
+      for patientUID in patientUIDs:
+         print(patientUID)
+         nodeNames.extend(DICOMLib.DICOMUtils.loadPatientByUID(patientUID))
+   #print(nodes)
+   nodes=[slicer.util.getNode(pattern=n) for n in nodeNames]
+   volumeNodes=[n for n in nodes if n.GetClassName()=='vtkMRMLScalarVolumeNode']
+   doseNodes=[n for n in volumeNodes if n.GetName().find('RTDOSE')>-1]
+   segmentationNodes=[n for n in nodes if n.GetClassName()=='vtkMRMLSegmentationNode']
+   nv=len(volumeNodes)
+   ns=len(segmentationNodes)
+   nd=len(doseNodes)
+   print(f'vol:{nv} seg:{ns} dose: {nd}')
+
+   return {'vol':volumeNodes,'dose':doseNodes,'seg':segmentationNodes}
+
+
+def checkNodes(config,nodes):
+   msg=''
+   nD=len(nodes['dose'])
+   if nD>1:
+      msg+=f'DOSE[{nD}]'
+   nS=len(nodes['seg'])
+   if nS>1:
+      if len(msg)>0:
+         msg+='/'
+      msg+=f'SEG[{nS}]'
+   return msg
+
+def getMeanDose(nodes,target):
+   segNode=nodes['seg'][0]
+   seg=segNode.GetSegmentation()
+
+   segmentIds=seg.GetSegmentIDs()
+#[seg.GetNthSegment(i) for i in range(seg.GetNumberOfSegments())]
+
+   targetSegmentIds=[s for s in segmentIds if seg.GetSegment(s).GetName()==target]
+   #labelmapVolumeNode = slicer.mrmlScene.AddNewNodeByClass("vtkMRMLLabelMapVolumeNode")
+   #export=slicer.vtkSlicerSegmentationsModuleLogic.ExportSegmentsToLabelmapNode
+   #export(segNode, targetSegmentIds, labelmapVolumeNode, nodes['dose'][0])
+   #nodes.update({'labelMap':labelmapVolumeNode})
+
+   doseNode=nodes['dose'][0]
+   doseArray = slicer.util.arrayFromVolume(doseNode)
+   segmentArray = slicer.util.arrayFromSegmentBinaryLabelmap(segNode, targetSegmentIds[0], doseNode)
+   doseVoxels = doseArray[segmentArray != 0]
+   print(numpy.mean(doseVoxels))
+#add a float() to avoid JSON complaining about float32 converion
+   return float(numpy.mean(doseVoxels))
+
+def getDicomInstances(config,sid):
+   odb=config['odb']
+   ofb=config['ofb']
+   sd=odb.getStudyData(sid)
+   series=sd['Series']
+   instances=[]
+   for s in series:
+      sed=odb.getSeriesData(s)
+      instances.extend(sed['Instances'])
+   #download instances one by one
+   baseDir=config.get('baseDir',os.path.join(os.path.expanduser('~'),'temp'))
+   outDir=os.path.join(baseDir,sid)
+   clearDir(outDir)
+   os.mkdir(outDir)
+   for oid in instances:
+      local=os.path.join(outDir,f'{oid}.dcm')
+      ofb.getInstance(oid,local)
+   return outDir
+
+
+def getDicomZip(config,sid):
+   ofb=config['ofb']
+   baseDir=config.get('baseDir',os.path.join(os.path.expanduser('~'),'temp'))
+   fname=f'{sid}.zip'
+   path=os.path.join(baseDir,fname)
+   if not os.path.isfile(path):
+      ofb.getZip('studies',sid,path,'archive')
+   print(f'Using {path}')
+   #unzip path
+   outDir=extractZip(config,path)
+   os.remove(path)
+   return outDir
+
+def extractZip(config,fname):
+#flattens the zip files in the baseDir/bname directory 
+#where bname is the basename of the file without the .zip suffix
+   fzip=zipfile.ZipFile(fname)
+   names=fzip.namelist()
+   bname=os.path.basename(fname)
+   bname=bname.replace('.zip','')
+   baseDir=config['baseDir']
+   outDir=os.path.join(baseDir,bname)
+   #clean
+   clearDir(outDir)
+   os.mkdir(outDir)
+   outnames=[os.path.join(outDir,f'out{i:03d}.dcm') for i in range(len(names))]
+   #extracts and renames (avoids *nix and win directory separator confusion)
+   for (member,out) in zip(names,outnames):
+      with fzip.open(member) as zf, open(out, 'wb') as f:
+         shutil.copyfileobj(zf, f)
+
+   return outDir
+
+def clearDir(outDir):
+   if os.path.isdir(outDir):
+      shutil.rmtree(outDir)
+
+if __name__=='__main__':
+   try:
+      main(sys.argv[1])
+   except IndexError:
+      main()
+   print('Succesful completion')
+   quit()
+