This will solve this problem by building a system that can collect and simultaneously process multi-modal healthcare data, such as audio (e.g. speech, tone, sentiment), movement (e.g. gait), video (e.g. face recognition, pain estimation, sentiment analysis), and derived signals (e.g. speech-to-text natural language analysis), within the same capture and processing framework. This will revolutionize the way healthcare information is analyzed, opening up a new set of tools to effectively comprehend complex, interdependent medical data to derive actionable intelligence.