Archive for the ‘Machine Learning’ Category

Multimodal Deep Learning – A Fusion of Multiple Modalities – NASSCOM Community

Multimodal Deep Learning and its Applications

As humans, our perception of the world is through our senses. We identify objects or anything through vision, sound, touch, and odor. Our way of processing this sensory information is multimodal. Modality refers to the way something is recognized, experienced, and recorded. Multimodal deep learning is an extensive research branch in Deep learning that works on the fusion of multimodal data.

The human brain consists of millions of neural networks that process multiple modalities from the external world. It could be recognizing a persons body movements, tone of voice, or even mimicking sounds. For AI to interpret Human Intelligence, we need a reasonable fusion of multimodal data and this is done through Multimodal Deep Learning.

Multimodal Machine Learning is developing computer algorithms that learn and predict using Multimodal datasets.

Multimodal Deep learning is a subset of the machine learning branch. With this technology, AI models are trained to identify relationships between multiple modalities such as images, videos, and texts and provide accurate predictions. From identifying the relevant link between datasets, Deep Learning models will be able to capture any place's environment and a person's emotional state.

If we say, Unimodal models that interpret only a single dataset have proven efficient in computer vision and Natural Language Processing. Unimodal models have limited capabilities; in certain tasks, these models failed to recognize humor, sarcasm, and hate speech. Whereas, Multimodal learning models can be referred to as a combination of unimodal models.

Multimodal deep learning includes modalities like visual, audio, and textual datasets. 3D visual and LiDAR data are slightly used multimodal data.

Multimodal Learning models work on the fusion of multiple unimodal neural networks.

First unimodal neural networks process the data separately and encode them, later, the encoded data is extracted and fused. Multimodal data fusion is an important process carried out using multiple fusion techniques. Finally, with the fusion of multimodal data, neural networks recognize and predict the outcome of the input key.

For example, in any video, there might be two unimodal models visual data and audio data. The perfect synchronization of both unimodal datasets provides simultaneous working of both models.

Fusing multimodal datasets improves the accuracy and robustness of Deep learning models, enhancing their performance in real-time scenarios.

Multimodal Deep learning has potential applications in computer vision algorithms. Here are some of its applications;

The research to reduce human efforts and develop machines matching with human intelligence is enormous. This requires multimodal datasets that can be combined using Machine Learning and Deep Learning models, paving the way for more advanced AI tools.

The recent surge in the popularity of AI tools has brought more additional investments in Artificial Intelligence and Machine Learning technology. This is a great time to grab job opportunities by learning and upskilling yourself in Artificial Intelligence and Machine Learning.

Continue reading here:
Multimodal Deep Learning - A Fusion of Multiple Modalities - NASSCOM Community

Prediction prolonged mechanical ventilation in trauma patients of … – Nature.com

Esteban, A. et al. Evolution of mortality over time in patients receiving mechanical ventilation. Am. J. Respir. Crit. Care Med. 188, 220230. https://doi.org/10.1164/rccm.201212-2169OC (2013).

Article PubMed Google Scholar

Divo, M. J., Murray, S., Cortopassi, F. & Celli, B. R. Prolonged mechanical ventilation in Massachusetts: The 2006 prevalence survey. Respir. Care 55, 16931698 (2010).

PubMed Google Scholar

Hsu, C. L. et al. Timing of tracheostomy as a determinant of weaning success in critically ill patients: A retrospective study. Crit. Care 9, R46-52. https://doi.org/10.1186/cc3018 (2005).

Article PubMed Google Scholar

Wang, C. H. et al. Predictive factors of in-hospital mortality in ventilated intensive care unit: A prospective cohort study. Medicine (Baltimore) 96, e9165. https://doi.org/10.1097/md.0000000000009165 (2017).

Article MathSciNet PubMed Google Scholar

Clark, P. A. & Lettieri, C. J. Clinical model for predicting prolonged mechanical ventilation. J. Crit. Care 28, 880.e881-880.e887 (2013).

Article Google Scholar

Sheikhbardsiri, H., Esamaeili Abdar, Z., Sheikhasadi, H., Ayoubi Mahani, S. & Sarani, A. Observance of patients rights in emergency department of educational hospitals in south-east Iran. Int. J. Hum. Rights Healthcare. 13, 435444 (2020).

Article Google Scholar

Parreco, J., Hidalgo, A., Parks, J. J., Kozol, R. & Rattan, R. Using artificial intelligence to predict prolonged mechanical ventilation and tracheostomy placement. J. Surg. Res. 228, 179187 (2018).

Article PubMed Google Scholar

Agle, S. C. et al. Early predictors of prolonged mechanical ventilation in major torso trauma patients who require resuscitation. Am. J. Surg. 192, 822827 (2006).

Article PubMed Google Scholar

Dimopoulou, I. et al. Prediction of prolonged ventilatory support in blunt thoracic trauma patients. Intensive Care Med. 29, 11011105 (2003).

Article PubMed Google Scholar

Figueroa-Casas, J. B. et al. Predictive models of prolonged mechanical ventilation yield moderate accuracy. J. Crit. Care 30, 502505 (2015).

Article PubMed Google Scholar

Davarani, E. R., Tavan, A., Amiri, H. & Sahebi, A. Response capability of hospitals to an incident caused by mass gatherings in southeast Iran. Injury 53, 17221726 (2022).

Article PubMed Google Scholar

Young, D., Harrison, D. A., Cuthbertson, B. H. & Rowan, K. Effect of early vs late tracheostomy placement on survival in patients receiving mechanical ventilation: The TracMan randomized trial. JAMA 309, 21212129. https://doi.org/10.1001/jama.2013.5154 (2013).

Article CAS PubMed Google Scholar

Gomes Silva, B. N., Andriolo, R. B., Saconato, H., Atallah, A. N. & Valente, O. Early versus late tracheostomy for critically ill patients. Cochrane Database Syst. Rev. 3 144 (2012).

Google Scholar

Rose, L. et al. Variation in definition of prolonged mechanical ventilation. Respir. Care 62, 13241332 (2017).

Article PubMed Google Scholar

Clark, P. A. & Lettieri, C. J. Clinical model for predicting prolonged mechanical ventilation. J. Crit. Care 28, 880-e881 (2013).

Article Google Scholar

Brook, A. D., Sherman, G., Malen, J. & Kollef, M. H. Early versus late tracheostomy in patients who require prolonged mechanical ventilation. Am. J. Crit. Care 9, 352 (2000).

Article CAS PubMed Google Scholar

Chang, Y.-C. et al. Ventilator dependence risk score for the prediction of prolonged mechanical ventilation in patients who survive sepsis/septic shock with respiratory failure. Sci. Rep. 8, 111 (2018).

ADS Google Scholar

Lone, N. I. & Walsh, T. S. Prolonged mechanical ventilation in critically ill patients: Epidemiology, outcomes and modelling the potential cost consequences of establishing a regional weaning unit. Crit. Care 15, 110 (2011).

Article Google Scholar

Dunn, H. et al. Mobilization of prolonged mechanical ventilation patients: An integrative review. Heart Lung 46, 221233. https://doi.org/10.1016/j.hrtlng.2017.04.033 (2017).

Article PubMed PubMed Central Google Scholar

Abujaber, A. et al. Using trauma registry data to predict prolonged mechanical ventilation in patients with traumatic brain injury: Machine learning approach. PLoS ONE 15, e0235231 (2020).

Article CAS PubMed PubMed Central Google Scholar

Zolbanin, H. M., Delen, D. & Zadeh, A. H. Predicting overall survivability in comorbidity of cancers: A data mining approach. Decis. Support Syst. 74, 150161 (2015).

Article Google Scholar

Shaikhina, T. et al. Decision tree and random forest models for outcome prediction in antibody incompatible kidney transplantation. Biomed. Signal Process. Control 52, 456462 (2019).

Article ADS Google Scholar

Archer, K. J. & Kimes, R. V. Empirical characterization of random forest variable importance measures. Comput. Stat. Data Anal. 52, 22492260 (2008).

Article MathSciNet MATH Google Scholar

Dag, A., Oztekin, A., Yucel, A., Bulur, S. & Megahed, F. M. Predicting heart transplantation outcomes through data analytics. Decis. Support Syst. 94, 4252 (2017).

Article Google Scholar

Cui, S., Wang, D., Wang, Y., Yu, P.-W. & Jin, Y. An improved support vector machine-based diabetic readmission prediction. Comput. Methods Programs Biomed. 166, 123135 (2018).

Article PubMed Google Scholar

Hale, A. T. et al. Machine-learning analysis outperforms conventional statistical models and CT classification systems in predicting 6-month outcomes in pediatric patients sustaining traumatic brain injury. Neurosurg. Focus 45, E2 (2018).

Article PubMed Google Scholar

Shi, H.-Y., Hwang, S.-L., Lee, K.-T. & Lin, C.-L. In-hospital mortality after traumatic brain injury surgery: A nationwide population-based comparison of mortality predictors used in artificial neural network and logistic regression models. J. Neurosurg. 118, 746752 (2013).

Article PubMed Google Scholar

Das, A. et al. Prediction of outcome in acute lower-gastrointestinal haemorrhage based on an artificial neural network: Internal and external validation of a predictive model. Lancet 362, 12611266 (2003).

Article PubMed Google Scholar

Han, J., Kamber, M. & Pei, J. Data Mining Concepts and Techniques 3rd edn. (University of Illinois at Urbana-Champaign Micheline Kamber Jian Pei Simon Fraser University, 2012).

MATH Google Scholar

Zolbanin, H. M., Delen, D. & Zadeh, A. H. Predicting overall survivability in comorbidity of cancers: A data mining approach. Decis Support Syst 74, 150161 (2015).

Article Google Scholar

Lakshmi, B. N., Indumathi, T. S. & Ravi, N. A study on C.5 decision tree classification algorithm for risk predictions during pregnancy. Procedia Technol. 24, 15421549 (2016).

Article Google Scholar

Rivers, E. et al. Early goal-directed therapy in the treatment of severe sepsis and septic shock. N. Engl. J. Med. 345, 13681377 (2001).

Article CAS PubMed Google Scholar

Weil, M. H. Functional Hemodynamic Monitoring 917 (Springer, 2005).

Book Google Scholar

Sevransky, J. Clinical assessment of hemodynamically unstable patients. Curr. Opin. Crit. Care 15, 234 (2009).

Article PubMed PubMed Central Google Scholar

Scheeren, T. W. L. et al. Current use of vasopressors in septic shock. Ann. Intensive Care 9, 112 (2019).

Article Google Scholar

Hidalgo, D. C., Patel, J., Masic, D., Park, D. & Rech, M. A. Delayed vasopressor initiation is associated with increased mortality in patients with septic shock. J. Crit. Care 55, 145148 (2020).

Article Google Scholar

Li, Y., Li, H. & Zhang, D. Timing of norepinephrine initiation in patients with septic shock: a systematic review and meta-analysis. Crit. Care 24, 19 (2020).

Article Google Scholar

Sellers, B. J., Davis, B. L., Larkin, P. W., Morris, S. E. & Saffle, J. R. Early prediction of prolonged ventilator dependence in thermally injured patients. J. Trauma 43, 899903 (1997).

Article CAS PubMed Google Scholar

Rachmale, S., Li, G., Wilson, G., Malinchoc, M. & Gajic, O. Practice of excessive FiO2 and effect on pulmonary outcomes in mechanically ventilated patients with acute lung injury. Respir. Care 57, 18871893 (2012).

Article PubMed Google Scholar

de Jonge, E. et al. Association between administered oxygen, arterial partial oxygen pressure and mortality in mechanically ventilated intensive care unit patients. Crit. Care 12, 18 (2008).

Article Google Scholar

Esan, A., Hess, D. R., Raoof, S., George, L. & Sessler, C. N. Severe hypoxemic respiratory failure: Part 1Ventilatory strategies. Chest 137, 12031216 (2010).

Article PubMed Google Scholar

Gajic, O. et al. Prediction of death and prolonged mechanical ventilation in acute lung injury. Crit. Care 11, 17 (2007).

Article Google Scholar

Seeley, E. et al. Predictors of mortality in acute lung injury during the era of lung protective ventilation. Thorax 63, 994998 (2008).

Article CAS PubMed Google Scholar

Nash, G., Blennerhassett, J. B. & Pontoppidan, H. Pulmonary lesions associated with oxygen therapy and artificial ventilation. Laval. Med. 276, 368374 (1967).

CAS Google Scholar

Ghauri, S. K., Javaeed, A., Mustafa, K. J. & Khan, A. S. Predictors of prolonged mechanical ventilation in patients admitted to intensive care units: A systematic review. Int. J. Health Sci. (Qassim) 13, 3138 (2019).

PubMed Google Scholar

Pu, L. et al. Weaning critically ill patients from mechanical ventilation: A prospective cohort study. J. Crit. Care 30, 862.e867813. https://doi.org/10.1016/j.jcrc.2015.04.001 (2015).

Article Google Scholar

Sellares, J. et al. Predictors of prolonged weaning and survival during ventilator weaning in a respiratory ICU. Intensive Care Med. 37, 775784. https://doi.org/10.1007/s00134-011-2179-3 (2011).

Article PubMed Google Scholar

Clark, P. A. & Lettieri, C. J. Clinical model for predicting prolonged mechanical ventilation. J. Crit. Care 28(880), e881-887. https://doi.org/10.1016/j.jcrc.2013.03.013 (2013).

Article Google Scholar

Clark, P. A., Inocencio, R. C. & Lettieri, C. J. I-TRACH: Validating a tool for predicting prolonged mechanical ventilation. J. Intensive Care Med. 33, 567573. https://doi.org/10.1177/0885066616679974 (2018).

Article PubMed Google Scholar

Rojek-Jarmua, A., Hombach, R. & Krzych, J. APACHE II score cannot predict successful weaning from prolonged mechanical ventilation. Chron. Respir. Dis. 14, 270275. https://doi.org/10.1177/1479972316687100 (2017).

Article PubMed PubMed Central Google Scholar

Go here to read the rest:
Prediction prolonged mechanical ventilation in trauma patients of ... - Nature.com

More ‘machine learning’ cameras to track truckies through NSW – Big Rigs

Transport for NSW (TfNSW) is installing a number of machine learning traffic counting and classifying cameras across the state.

Images of heavy vehicles are taken by the cameras, which then classify the type of vehicle in transit and the type of cargo being transported.

In a bulletin to industry TfNSW said the information collected helps to shape the future of freight, to better understand freight movements, improve road safety, and enable more efficient deliveries.

The cameras are not used for enforcement or monitoring people or private vehicles, said TfNSW.

Truckies can expect to see cameras installed at the following locations over the coming weeks:

According to thefact sheet on the cameras webpage, there is a radar sensor and camera on the unit that takes a picture of the heavy vehicles when certain criteria are met.

After the picture is taken, artificial intelligence within the unit can tell the difference between different types of heavy vehicles, for example, a container carrying heavy vehicle, B-double or semi-trailer.

The units are also able to track changes in load. If a shipping container truck entered a location carrying one container and left with two containers the platform contains a record of this change.

Aside from the above locations, TfNSw says similar cameras can also be found at:

For more information, visit the Machine Learning cameras webpagewhich includes a factsheet with details about what these units are and what they do.

If you have any questions or would like more information, you can contact the TfNSW project team at freight@transport.nsw.gov.au.

Follow this link:
More 'machine learning' cameras to track truckies through NSW - Big Rigs

How machine learning can help crack the IT security problem – VentureBeat

Join top executives in San Francisco on July 11-12, to hear how leaders are integrating and optimizing AI investments for success. Learn More

Less than a decade ago, the prevailing wisdom was that every business should undergo digital transformations to boost internal operations and improve client relationships. Next, they were being told that cloud workloads are the future and that elastic computer solutions enabled them to operate in an agile and more cost-effective manner, scaling up and down as needed.

While digital transformations and cloud migrations are undoubtedly smart decisions that all organizations should make (and those that havent yet, what are you doing!), security systems meant to protect such IT infrastructures havent been able to keep pace with threats capable of undermining them.

As internal business operations become increasingly digitized, boatloads more data are being produced. With data piling up, IT and cloud security systems come under increased pressure because more data leads to greater threats of security breaches.

In early 2022, a cyber extortion gang known as Lapsus$ went on a hacking spree, stealing source code and other valuable data from prominent companies, including Nvidia, Samsung, Microsoft and Ubisoft. The attackers had originally exploited the companies networks using phishing attacks, which led to a contractor being compromised, giving the hackers all the access the contractor had via Okta (an ID and authentication service). Source code and other files were then leaked online.

Transform 2023

Join us in San Francisco on July 11-12, where top executives will share how they have integrated and optimized AI investments for success and avoided common pitfalls.

This attack and numerous other data breaches target organizations of all types, ranging from large multinational corporations to small startups and growing firms. Unfortunately, in most organizations, there are simply too many data points for security engineers to locate, meaning current systems and methods to safeguard a network are fundamentally flawed.

Additionally, organizations are often overwhelmed by the various available tools to tackle these security challenges. Too many tools means organizations invest an exorbitant amount of time and energy not to mention resources in researching, purchasing and then integrating and running these tools. This puts added stress on executives and IT teams.

With so many moving parts, even the best security engineers are left helpless in trying to mitigate potential vulnerabilities in a network. Most organizations simply dont have the resources to make cybersecurity investments.

As a result, they are subject to a double-edged sword: Their business operations rely on the highest levels of security, but achieving that comes at a cost that most organizations simply cant afford.

A new approach to computer security is desperately needed to safeguard businesses and organizations sensitive data. The current standard approach comprises rules-based systems, usually with multiple tools to cover all bases. This practice leaves security analysts wasting time enabling and disabling rules and logging in and out of different systems in an attempt to establish what is and what isnt considered a threat.

The best option for organizations dealing with these ever-present pain points is to leverage machine learning (ML) algorithms. This way, algorithms can train a model based on behaviors, providing any business or organization a secure IT infrastructure. A tailored ML-based SaaS platform that operates efficiently and in a timely manner must be the priority of any organization or business seeking to revamp its security infrastructure.

Cloud-native application protection platforms (CNAPP), a security and compliance solution, can empower IT security teams to deploy and run secure cloud native applications in automated public cloud environments. CNAPPs can apply ML algorithms on cloud-based data to discover accounts with unusual permissions (one of the most common and undetected attack paths) and uncover potential threats including host and open source vulnerabilities.

ML can also knit together many anomalous data points to create rich stories of whats happening in a given network something that would take a human analyst days or weeks to uncover.

These platforms leverage ML through two primary practices. Cloud security posture management (CSPM) handles platform security by monitoring and delivering a full inventory to identify any deviations from customized security objectives and standard frameworks.

Cloud infrastructure entitlements management (CIEM) focuses on identity security by understanding all possible access to sensitive data through every identitys permission. On top of this, host and container vulnerabilities are also taken into account, meaning correct urgency can be applied to ongoing attacks. For example, anomalous behavior seen on a host with known vulnerabilities is far more pressing than on a host without known vulnerabilities.

Another ML-based SaaS option is to outsource the security operations center (SOC) and security incident and event management (SIEM) function to a third party and benefit from their ML algorithm. With dedicated security analysts investigating any and all threats, SaaS can use ML to handle critical security functions such as network monitoring, log management, single-sign on (SSO) and endpoint alerts, as well as access gateways.

SaaS ML platforms offer the most effective way to cover all the security bases. By applying ML to all behaviors, organizations can focus on their business objectives while algorithms pull all the necessary context and insights into a single security platform.

Running the complex ML algorithms to learn a baseline of what is normal in a given network and assessing risk is challenging even if an organization has the personnel to make it a reality. For the majority of organizations, using third-party platforms that have already built algorithms to be trained on data produces a more scalable and secure network infrastructure, doing so far more conveniently and effectively than home grown options.

Relying on a trusted third party to host a SaaS ML platform enables organizations to dedicate more time to internal needs, while the algorithms study the networks behavior to provide the highest levels of security.

When it comes to network security, relying on a trusted third party is no different than hiring a locksmith to repair the locks on your home. Most of us dont know how the locks on our homes work but we trust an outside expert to get the job done. Turning to third-party experts to run ML-algorithms enables businesses and organizations the flexibility and agility they need to operate in todays digital environment.

Maximizing this new approach to security allows all types of organizations to overcome their complex data problems without having to worry about the resources and tools needed to protect their network, providing unparalleled peace of mind.

Ganesh the Awesome (Steven Puddephatt) is a technical sales architect at GlobalDots.

Welcome to the VentureBeat community!

DataDecisionMakers is where experts, including the technical people doing data work, can share data-related insights and innovation.

If you want to read about cutting-edge ideas and up-to-date information, best practices, and the future of data and data tech, join us at DataDecisionMakers.

You might even considercontributing an articleof your own!

Read More From DataDecisionMakers

Visit link:
How machine learning can help crack the IT security problem - VentureBeat

AI and Machine Learning in Healthcare for the Clueless – Medscape

Recorded March 6, 2023. This transcript has been edited for clarity.

Robert A. Harrington, MD: Hi. This is Bob Harrington on theheart.org | Medscape Cardiology, and I'm here at the American College of Cardiology meetings in New Orleans, having a great time, by the way. It's really fun to be back live, in person, getting to see friends and colleagues, seeing live presentations, etc. If you've not been to a live meeting yet over the course of the past couple of years, please do start coming again, whether it's American College of Cardiology, American Heart Association, or European Society of Cardiology. It's fantastic.

Putting that aside, I've been learning many things at this meeting, particularly around machine learning, artificial intelligence (AI), and some of the advanced computational tools that people in the data-science space are using.

I'm fortunate to have an expert and, really, a rising thought leader in this field, Dr Jenine John. Jenine is a machine-learning research fellow at Brigham and Women's Hospital, working in Calum MacRae'sresearch group.

What she talked about on stage this morning is what do you have to know about this whole field. I thought we'd go through some of the basic concepts of data science, what machine learning is, what AI is, and what neural networks are.

How do we start to think about this? As practitioners, we're going to be faced with how to incorporate some of this into our practice. You're seeing machine-learning algorithms put into your clinical operations. You're starting to see ways that people are thinking about, for example, Can the machine read the echocardiogram as good as we can? What's appropriate for the machine? What's appropriate for us? What's the oversight of all of this?

We'll have a great conversation for the next 12-20 minutes and see what we can all learn together. Jenine, thank you for joining us here today.

Jenine John, MD: Thank you for having me.

Harrington: Before we get into the specifics of machine learning and what you need to know, give me a little bit of your story. You obviously did an internal medicine residency. You did a cardiology fellowship. Now, you're doing an advanced research fellowship. When did you get bitten by the bug to want to do data science, machine learning, etc.?

John: It was quite late, actually. After cardiology fellowship, I went to Brigham and Women's Hospital for a research fellowship. I started off doing epidemiology research, and I took classes at the public health school.

Harrington: The classic clinical researcher.

John: Exactly. That was great because I gained a foundation in epidemiology and biostatistics, which I believe is essential for anyone doing clinical research. In 2019, I was preparing to write a K grant, and for my third aim, I thought, Oh, I want to make this complex model that uses many variables. This thing called machine learning might be helpful. I basically just knew the term but didn't know much about it.

I talked to my program director who led me to Dr Rahul Deo and Dr Calum MacRae's group that's doing healthcare AI. Initially, I thought I would just collaborate with them.

Harrington: Have their expertise brought into your grant and help to elevate the whole grant? That's the typical thing to do.

John: Exactly. As I learned a bit more about machine learning, I realized that this is a skill set I should really try to develop. I moved full-time into that group and learned how to code and create machine-learning models specifically for cardiac imaging. Six months later, the pandemic hit, so everything took a shift again.

I believe it's a shift for the better because I was exposed to everything going on in digital health and healthcare startups. There was suddenly an interest in monitoring patients remotely and using tech more effectively. I also became interested in how we are applying AI to healthcare and how we can make sure that we do this well.

Harrington: There are a couple of things that I want to expand on. Maybe we'll start this way. Let's do the definitions. How would you define AI and its role in medicine? And then, talk about a subset of that. Define machine learning for the audience.

John: Artificial intelligence and machine learning, the two terms are used pretty much synonymously within healthcare, because when we talk about AI in healthcare, really, we're talking about machine learning. Some people use the term AI differently. They feel that it's only if a system is autonomously thinking independently that you can call it AI. For the purposes of healthcare, we pretty much use them synonymously.

Harrington: For what we're going to talk about today, we'll use them interchangeably.

John: Yes, exactly.

Harrington: Define machine learning.

John: Machine learning is when a machine uses data and learns from the data. It picks up patterns, and then, it can basically produce output based on those patterns.

Harrington: Give me an example that will resonate with a clinical audience. You're an imager, and much of the work so far has been in imaging.

John: Imaging is really where machine learning shines. For example, you can use machine learning on echocardiograms, and you can use it to pick up whether this patient has valvular disease or not. If you feed an AI model enough echocardiograms, it'll start to pick up the patterns and be able to tell whether this person has valvular disease or not.

Harrington: The group that you're working with has been very prominent in being able to say whether they have hypertrophic cardiomyopathy, valve disease, or amyloid infiltrative disease.

There are enough data there that the machine starts to recognize patterns.

John: Yes.

Harrington: You said that you were, at the Harvard School of Public Health, doing what I'll call classic clinical research training. I had the same training. I was a fellow 30-plus years ago in the Duke Databank for Cardiovascular Diseases, and it was about epidemiology and biostatistics and how to then apply those to the questions of clinical research.

You were doing very similar things, and you said something this morning in your presentation that stuck with me. You said you really need to understand these things before you make the leap into trying to understand machine learning. Expand on that a little bit.

John: I think that's so important because right now, what seems to happen is you have the people the data scientists and clinicians and they seem to be speaking different languages. We really need more collaboration and getting on the same page. When clinicians go into data science, I think the value is not in becoming pure data scientists and learning to create great machine-learning models. Rather, it's bringing that clinical thinking and that clinical research thinking, specifically, to data science. That's where epidemiology and biostatistics come in because you really need to understand those concepts so that you understand which questions you should be asking. Are you using the right dataset to ask those questions? Are there biases that could be present?

Harrington: Every week, as you know, we all pick up our journals, and there's a machine-learning paper in one of the big journals all the time. Some of the pushback you'll hear, whether it's on social media or in letters to the editors, is why did you use machine learning for this? Why couldn't you use classical logistic regression?

One of the speakers in your session, I thought, did a nice job of that. He said that often, standard conventional statistics are perfectly fine. Then there are some instances where the machine is really better, and imaging is a great example. Would you talk to the audience a little bit about that?

John: I see it more as a continuum. I think it's helpful to see it that way because right now, we see traditional biostatistics and machine learning as completely different. Really, it's a spectrum of tools. There are simple machine-learning methods where you can't really differentiate much from statistical methods, and there's a gray zone in the middle. For simpler data, such as tabular data, maybe.

Harrington: Give the audience an example of tabular data.

John: For example, if you have people who have had a myocardial infarction (MI), and then you have characteristics of those individuals, such as age, gender, and other factors, and you want to use those factors to predict who gets an MI, in that instance, traditional regression may be best. When you get to more complex data, that's where machine learning really shines. That's where it gets exciting because they are questions that we haven't been able to ask before with the methods that we have. Those are the questions that we want to start using machine learning to answer.

Harrington: We've all seen papers again over the past few years. The Mayo Group has published a series of these about information that you can derive from the EKG. You can derive, for example, potassium levels from the EKG. Not the extremes that we've all been taught, but subtle perturbations. I think I knew this, but I was still surprised to hear it when one of your co-speakers said that there are over 30,000 data points in the typical EKG.

There's no way you can use conventional statistics to understand that.

John: Exactly. One thing I was a little surprised to see is that machine learning does quite well with estimating the age of the individual on the EKG. If you show a cardiologist an EKG, we could get an approximate estimate, but we won't be as good as the machine. Modalities like EKG and echocardiogram, which have so many more data points, are where the machine can find patterns that even we can't figure out.

Harrington: The secret is to ingest a huge amount of data. One of the things that people will ask me is, "Well, why is this so hot now?" It's hot now for a couple of reasons, one of which is that there's an enormous amount of data available. Almost every piece of information can be put into zeros and ones. Then there's cloud computing, which allows the machine to ingest this enormous amount of information.

You're not going to tell the age of a person from a handful of EKGs. It's thousands to millions of EKGs that machines evaluated to get the age. Is that fair?

John: This is where we talk about big data because we need large amounts of data for the machine to learn how to interpret these patterns. It's one of the reasons I'm excited about AI because it's stimulating interest in multi-institution collaborations and sharing large datasets.

We're annotating, collecting, and organizing these large multi-institutional datasets that can be used for a variety of purposes. We can use the full range of analytic approaches, machine learning or not, to learn more about patients and how to care for them.

Harrington: I've heard both Calum and Rahul talk about how they can get echocardiograms, for example, from multiple institutions. As the machine gets better and better at reading and interpreting the echocardiograms or looking for patterns of valvular heart disease, they can even take a more limited imaging dataset and apply what they've learned from the larger expanded dataset, basically to improve the reading of that echocardiogram.

One of the things it's going to do, I think, is open up the opportunity for more people to contribute their data beyond the traditional academics.

John: Because so much data are needed for AI, there's a role for community centers and other institutions to contribute data so that we can make robust models that work not only in a few academic centers but also for the majority of the country.

Harrington: There are two more topics I want to cover. We've been, in some ways, talking about the hope of what we're going to use this for to make clinical medicine better. There's also what's been called the hype, the pitfalls, and the perils. Then I want to get into what do you need to know, particularly if you're a resident fellow, junior faculty member.

Let's do the perils and the hype. I hear from clinicians, particularly clinicians of my generation, that this is just a black box. How do I know it's right? People point to, for example, the Epic Sepsis Model, which failed miserably, with headlines all over the place. They worry about how they know whether it's right.

John: That's an extremely important question to ask. We're still in the early days of using AI and trying to figure out the pitfalls and how to avoid them. I think it's important to ask along the way, for each study, what is going on here. Is this a model that we can trust and rely on?

I also think that it's not inevitable that AI will transform healthcare just yet because we are so early on, and there is hype. There are some studies that aren't done well. We need more clinicians understanding machine learning and getting involved in these discussions so that we can lead the field and actually use the AI to transform healthcare.

Harrington: As you push algorithms into the healthcare setting, how do we evaluate them to make sure that the models are robust, that the data are representative, and that the algorithm is giving us, I'll call it, the right answer?

John: That's the tough part. I think one of the tools that's important is a prospective trial. Not only creating an algorithm and implementing right away but rather studying how it does. Is it actually working prospectively before implementing it?

We also need to understand that in healthcare, we can't necessarily accept the black box. We need explainability and interpretability, to get an understanding of the variables that are used, how they're being used within the algorithm, and how they're being applied.

One example that I think is important is that Optum created a machine-learning model to predict who was at risk for medical complications and high healthcare expenditures. The model did well, so they used the model to determine who should get additional resources to prevent these complications.

It turns out that African Americans were utilizing healthcare less, so their healthcare expenditure was lower. Because of that, the algorithm was saying these are not individuals who need additional resources.

Harrington: It's classic confounding.

John: There is algorithmic bias that can be an issue. That's why we need to look at this as clinical researchers and ask, "What's going on here? Are there biases?"

Harrington: One of the papers over the past couple of years came from one of our faculty members at Stanford, which looked at where the data are coming from for these models. It pointed out that there are many states in this country that contribute no data to the AI models.

That's part of what you're getting at, and that raises all sorts of equity questions. You're in Massachusetts. I'm in California. There is a large amount of data coming from those two states. From Mississippi and Louisiana, where we are now, much less data. How do we fix that?

John: I think we fix it by getting more clinicians involved. I've met so many passionate data scientists who want to contribute to patient care and make the world a better place, but they can't do it alone. They can't recruit health centers in Mississippi. We need clinicians and clinical researchers who will say, "I want to help with advancing healthcare, and I want to contribute data so that we can make this work." Currently, we have so many advances in some ways, but AI can open up so many new opportunities.

Harrington: There's a movement to assure that the algorithm is fair, right? That's the acronym that's being used to make sure that the data are representative of the populations that you're interested in and that you've eliminated the biases.

I'm always intrigued. When you talk to your friends in the tech world, they say, "Well, we do this all the time. We do A/B testing." They just constantly run through algorithms through A/B testing, which is a randomized study. How come we don't do more of that in healthcare?

John: I think it's complicated because we don't have the systems to do that effectively. If we had a system where patients come into the emergency room and we're using AI in that manner, then maybe we could start to incorporate some of these techniques that the tech industry uses. That's part of the issue. One is setting up systems to get the right data and enough data, and the other is how do we operationalize this so that we can effectively use AI within our systems and test it within our systems.

Harrington: As a longtime clinical researcher and clinical trialist, I've always asked why it is that clinical research is separate from the process of clinical care.

If we're going to effectively evaluate AI algorithms, for example, we've got to break down those barriers and bring research more into the care domain.

John: Yes. I love the concept of a learning health system and incorporating data and data collection into the clinical care of patients.

Harrington: Fundamentally, I believe that the clinicians make two types of decisions, one of which is that the answer is known. I always use the example of aspirin if you're having an ST-segment elevation MI. That's known. It shouldn't be on the physician to remember that. The system and the algorithms should enforce that. On the other hand, for much of what we do, the answer is not known, or it's uncertain.

Why don't we allow ongoing randomization to help us decide what is appropriate? We're not quite there yet, but I hope before the end of my career that we can push that closer together.

All right. Final topic for you. You talked this morning about what you need to know. Cardiology fellows and residents must approach you all the time and say, "Hey, I want to do what you do," or, "I don't want to do what you do because I don't want to learn to code, but I want to know how to use it down the road."

What do you tell students, residents, and fellows that they need to know?

John: I think all trainees and all clinicians, actually, should understand the fundamentals of AI because it is being used more and more in healthcare, and we need to be able to understand how to interpret the data that are coming out of AI models.

I recommend looking up topics as you go along. Something I see is clinicians avoid papers that involve AI because they feel they don't understand it. Just dive in and start reading these papers, because most likely, you will understand most of it. You can look up topics as you go along.

There's one course I recommend online. It's a free course through Coursera called AI in Healthcare Specialization. It's a course by Stanford, and it does a really good job of explaining concepts without getting into the details of the coding and the math.

Other than that, for people who want to get into the coding, first of all, don't be afraid to jump in. I recently talked to a friend who is a gastroenterologist, and she said, "I'd love to get into AI, but I don't think I'd be good at it." I asked, "Well, why not?" She said, "Because men tend to be good at coding."

I do not think that's true.

Harrington: I don't think that's true either.

John: It's interesting because we're all affected to some extent by the notions that society has instilled in us. Sometimes it takes effort to go beyond what you think is the right path or what you think is the traditional way of doing things, and ask, "What else is out there. What else can I learn?"

If you do want to get into coding, I would say that it's extremely important to join a group that specializes in healthcare AI because there are so many pitfalls that can happen. There are mistakes that could be made without you realizing it if you try to just learn things on your own without guidance.

Harrington: Like anything else, join an experienced research group that's going to impart to you the skills that you need to have.

John: Exactly.

Harrington: The question about women being less capable coders than men, we both say we don't believe that, and the data don't support that. It's interesting. At Stanford, for many years, the most popular major for undergraduate men has been computer science. In the past few years, it's also become the most popular major for undergrad women at Stanford.

We're starting to see, to your point, that maybe some of those attitudes are changing, and there'll be more role models like you to really help that next generation of fellows.

Final question. What do you want to do when you're finished?

John: My interests have changed, and now I'm veering away from academia and more toward the operational side of things. As I get into it, my feeling is that currently, the challenge is not so much creating the AI models but rather, as I said, setting up these systems so that we can get the right data and implement these models effectively. Now, I'm leaning more toward informatics and operations.

I think it's an evolving process. Medicine is changing quickly, and that's what I would say to trainees and other clinicians out there as well. Medicine is changing quickly, and I think there are many opportunities for clinicians who want to help make it happen in a responsible and impactful manner.

Harrington: And get proper training to do it.

John: Yes.

Harrington: Great. Jenine, thank you for joining us. I want to thank you, the listeners, for joining us in this discussion about data science, artificial intelligence, and machine learning.

My guest today on theheart.org | Medscape Cardiology has been Dr Jenine John, who is a research fellow at Brigham and Women's Hospital, specifically in the data science and machine learning realm.

Again, thank you for joining.

Robert A. Harrington, MD, is chair of medicine at Stanford University and former president of the American Heart Association. (The opinions expressed here are his and not those of the American Heart Association.) He cares deeply about the generation of evidence to guide clinical practice. He's also an over-the-top Boston Red Sox fan.

Follow Bob Harrington on Twitter

Follow theheart.org | Medscape Cardiology on Twitter

Follow Medscape on Facebook, Twitter, Instagram, and YouTube

See more here:
AI and Machine Learning in Healthcare for the Clueless - Medscape