NASA Logo

NTRS

NTRS - NASA Technical Reports Server

Back to Results
Medics: Medical Decision Support System for Long-Duration Space ExplorationThe Autonomous Medical Operations (AMO) group at NASA Ames is developing a “medical decision support system” to enable astronauts on long-duration exploration missions to operate autonomously. The system will support clinical actions by providing medical interpretation advice and procedural recommendations during emergent care and clinical work performed by crew. The current state of development of the system, called MedICS (Medical Interpretation Classification and Segmentation) includes two separate aspects: a set of machine learning diagnostic models trained to analyze organ images and patient health records, and an interface to ultrasound diagnostic hardware and to medical repositories. Three sets of images of different organs and medical records were utilized for training machine learning models for various analyses, as follows: 1. Pneumothorax condition (collapsed lung). The trained model provides a positive or negative diagnosis of the condition. 2. Carotid artery occlusion. The trained model produces a diagnosis of 5 different occlusion levels (including “normal”). 3. Ocular retinal images. The model extracts optic disc pixels (image segmentation). This is a precursor step for advanced autonomous fundus clinical evaluation algorithms to be implemented in FY20. 4. Medical health records. The model produces a differential diagnosis for any particular individual, based on symptoms and other health and demographic information. A probability is calculated for each of 25 most common conditions. The same model provides the likelihood of survival. All results are provided with a confidence level. Item 1 images were provided by the US Army and were part of a data set for the clinical treatment of injured battlefield soldiers. This condition is relevant to possible space mishaps, due to pressure management issues. Item 2 images were provided by Houston Methodist Hospital, and item 3 health records were acquired from the MIT laboratory of computational physiology. The machine learning technology utilized is deep multilayer networks (Deep Learning), and new models will continue to be produced, as relevant data is made available and specific health needs of astronaut crews are identified. The interfacing aspects of the system include a GUI for running the different models, and retrieving and storing data, as well as support for integration with an augmented reality (AR) system deployed at JSC by Tietronix Software Inc. (HoloLens). The AR system provides guidance for the placement of an ultrasound transducer that captures images to be sent to the MedICS system for diagnosis. The image captured and the associated diagnosis appear in the technician’s AR visual display.
Document ID
20200000615
Acquisition Source
Ames Research Center
Document Type
Conference Paper
Authors
Colombano, Silvano
(NASA Ames Research Center Moffett Field, CA, United States)
Lindsey, Tony
(NASA Ames Research Center Moffett Field, CA, United States)
Mansouri-Samani, Masoud
(Stinger Ghaffarian Technologies Inc. (SGT Inc.) Moffett Field, CA, United States)
Lee, Rebecca
(Stinger Ghaffarian Technologies Inc. (SGT Inc.) Moffett Field, CA, United States)
Marker, Nimesh
(Stinger Ghaffarian Technologies Inc. (SGT Inc.) Moffett Field, CA, United States)
Date Acquired
January 31, 2020
Publication Date
January 27, 2020
Subject Category
Lunar And Planetary Science And Exploration
Report/Patent Number
ARC-E-DAA-TN77238
Meeting Information
Meeting: Human Research Program Investigators' Workshop (HRP IWS 2020)
Location: Galveston, TX
Country: United States
Start Date: January 27, 2020
End Date: January 30, 2020
Sponsors: NASA Headquarters
Funding Number(s)
CONTRACT_GRANT: NNA14AA60C
WBS: 354108.04.23.21
Distribution Limits
Public
Copyright
Public Use Permitted.
No Preview Available