Voice / Text to Speech in/on OpenEHR

Hi everyone / anyone :slight_smile:

Very happy to join this forum having been interested and drawn towards OpenEzhR for a while, thanks to @Kanthan_Theivendran, Ian McNicoll, Alastair Allen and for helping me to better understand it and also an excellent 13 minute explainer video from @Sidharth_Ramesh actually which very much helped it all click into place for me.

We are looking at building a health app for GPs and lower tiered cities in India, helping improve transparency, efficiency and access to better quality healthcare and am exploring doing so on OpenEHR.

Iā€™d like the app to utilise speech to text / voice, but I have searched ā€œvoiceā€ and ā€œspeechā€ here in this forum and thereā€™s no mention of this that I can find.

Has anyone built UI/UX within the OpenEHR tooling/framework that handles this? Iā€™m not sure if this is viewed as an ā€œupstreamā€ factor and OpenEHR will handle any converted voice once itā€™s in text, but the UI and API considerations in terms of making this work well/fast/effectively from a UX standpoint is where I am comimg from, I guess. Can this be done using existing UI extensions / SDKs or would we need to develop this?

More generally, anyone who has or is looking at this, would love to hear from you, please :slight_smile:

Thanks!
Olly

2 Likes

@Olly_Cogan I saw in EHRCon24 @Sidharth_Ramesh showing a demo of this functionality. I also know some others are working on it but canā€™t disclose as it may be confidential so Iā€™ll let people jump in here to answer. Pretty cool functionality which would significantly reduce burden on providers.

I had my own little attempt to hook up an openEHR CDR to voice, albeit patient input not clinical

https://nhshackday.com/projects/23-cardiff/the_ballad_of_yann_and_alexa

Worth confirming though that openEHR itself is only about the data layer and definitely ā€˜downstreamā€™ of this kind of application, which is why you will not anything in official openEHR documentation.

There is nothing special about the openEHR API that would make it unsuitable to work with text/voice analysis which presumably , initially produces narrative.

Then we are into the challenge of generating structured input from narrative.

4 Likes

Hey @Olly_Cogan I can share a demo of the work weā€™ve been doing.
Weā€™ve got a generic voice ā†’ forms converter that takes any openEHR template as input. Thanks @Koray_Atalag for the mention.

Weā€™re currently experimenting with integrating external terminology referenced in openEHR templates. Eg: terminology://fhir.hl7.org/ValueSet/$expand?url=http://hl7.org/fhir/ValueSet/allergyintolerance-code in a DV_CODED_TEXT field as a refset should make the system only output terms from the terminology system.

I canā€™t share a video demo in public yet, but if you email (sidharth [at] medblocks [dot] com / dm me, I can show what weā€™ve been doing.

2 Likes

Hi all! We are setting a research study with @jpieraj to validate the population of openEHR templates from a clinical conversation in a real clinic using LLM and probably agentic RAG. It is early days so we will keep you updated on how it goes as we have some challenges such as Catalan/Spanish language use during the consultation, etcā€¦ Will contact you @Sidharth_Ramesh to see how we can collaborate on this!

4 Likes

A very important and interesting initiative - well done! I believe that Apperta is coordinating a similar kind of exercise in context of their efforts towards OpenEyes/openEHR harmonisation. I understand that we may have results demonstrated soon - perhaps at the January OpenEyes development group meeting.

4 Likes

@Olly_Cogan Thereā€™s a public demo of the speech to openEHR compositions running as a SMART app now available online: https://youtu.be/ShCzrHlzmHo?si=MXQVhoTo1yfuIJ6U&t=969

Go to around 16 mins and 10s to see it. @sebastian.iancu, @xabiermichelena and @Koray_Atalag have already seen :wink:

4 Likes