Hi Just thinking of use cases: Say I have a bunch of hand written doctor notes and medical test...

Hi Just thinking of use cases: Say I have a bunch of hand written doctor notes and medical test reports. I want to convert the same in the form of a proper data and then run ML models to find high risk patients. Will I be able to do that with this? Also, does it include the complete data lifecycle in terms of all the processes from data cleaning to feature engineering and final model selection and deployment. And is deployment of models in built in this or do I have to deploy the model seperately? Whats the process for the same? How much amount of data can it handle on a higher side?

ramavtar
| Deals Bought: 12 | Member Since:
    Waldemar_Cogniflow
    Waldemar_CogniflowFounder team
    | Member Since:

    Hi!

    Hand written doctor notes could be difficult to read with great precision because typically the quality of doctor's written text is pretty bad, but you can try some of them with our OCR model and see if it can get something readable. If you want to use that to later train a model using the extracted text from the OCR model, then it does not need to be perfect and typos could be accepted by the language model and live with that.

    For text data Cogniflow does some preprocessing and cleaning that you can customize (expanding the advanced option area) at the final step when creating an experiment.

    Model selection and deployment are addressed by Cogniflow and you do not have to worry about that. The platform trains different candidate models and hyper-parameters and it will recommend the one that gets the best optimization metric (typically F1 but you can change if you want). You always have the possibility to inspect all the candidate model and use any of them if you do not want to use the recommended one (maybe because you want to use a faster model at inference time no matter it was not the best at the accuracy level).

    Automatic deployment of all candidate models is executed after training so you can use them immediately, any of the trained models through our rest API.

    The amount of data depends of the experiment type, typically for text experiment you can train tens or hundreds thousands training examples, for images/audio web based solution will allow to train up to 10k - 15k images. If you require more resource you can reach us and let us know, we'll be happy to help you.

    Hope this can be helpful.

    Happy training!