Stream: implementers
Topic: FHIR in Google's ML
Josh Mandel (Jan 29 2018 at 14:48):
Pre-print article from Friday: Scalable and accurate deep learning for electronic health records shows (among many other things) where Google's using FHIR in deepn learning for EHRs.
It's super exciting to see this work shared! Congrats @Eyal Oren and @Patrik Sundberg!
Travis Stenerson (Jan 29 2018 at 16:58):
This is amazing, is Tensorflow on FHIR next? I wish there were/hope there will be more details about the data pre-processing into FHIR. Particularly how they made or represented the clinical notes.
Eyal Oren (Jan 29 2018 at 21:59):
Hey Travis, yes, we're using TF on top of FHIR but opening that up publicly is a bit tricky (standard models don't work very well for this kind of data, for various reasons I can go into). We're planning to though. As for details on data-processing, there is a (somewhat short) description in the "methods" section, but we may publish a more extensive description. Re: notes .. we just read them left to right ;-) The sequence of tokens within each notes is maintained, as is the sequence of notes, and the relative timestamp of notes (as are all the relative timestamps of all data). Then we "just" train, with various architectures as described, including eg LSTMs.
Travis Stenerson (Feb 02 2018 at 06:39):
Not often you get an immediate reply from one of lead authors! Incredible work, congrats! Those tokens I guess were just key words encountered repeatedly? Did you use a terminology service to link synonyms? I read the short section in appendix A, but definitely will be reading the rest, especially about the architecture of the feed forward model used, as I tried and failed to use one on time series diabetes data. I have a bit of an idea of why this kind of data doesn't work well with standard models, but I'd still love to hear the reasons.
Alex Sullivan (Feb 02 2018 at 15:21):
"we did not harmonize elements to a standard terminology or ontology"
This makes me smile, there is so much room to grow from here
Eyal Oren (Mar 02 2018 at 20:41):
Not often you get an immediate reply from one of lead authors! Incredible work, congrats! Those tokens I guess were just key words encountered repeatedly? Did you use a terminology service to link synonyms? I read the short section in appendix A, but definitely will be reading the rest, especially about the architecture of the feed forward model used, as I tried and failed to use one on time series diabetes data. I have a bit of an idea of why this kind of data doesn't work well with standard models, but I'd still love to hear the reasons.
sorry for late reply, didn't see this previously. For noten, each word becomes a token, in sequence. Not terminology services are used, the model learns synonyms (google: embeddings and word2vec for more info)
Eyal Oren (Mar 02 2018 at 20:41):
"we did not harmonize elements to a standard terminology or ontology"
This makes me smile, there is so much room to grow from here
Thanks!
Last updated: Apr 12 2022 at 19:14 UTC