Archive for the ‘Machine Learning’ Category
Medical content creation in the age of generative AI | Amazon Web Services – AWS Blog
Generative AI and transformer-based large language models (LLMs) have been in the top headlines recently. These models demonstrate impressive performance in question answering, text summarization, code, and text generation. Today, LLMs are being used in real settings by companies, including the heavily-regulated healthcare and life sciences industry (HCLS). The use cases can range from medical information extraction and clinical notes summarization to marketing content generation and medical-legal review automation (MLR process). In this post, we explore how LLMs can be used to design marketing content for disease awareness.
Marketing content is a key component in the communication strategy of HCLS companies. Its also a highly non-trivial balance exercise, because the technical content should be as accurate and precise as possible, yet engaging and empowering for the target audience. The main goal of the marketing content is to raise awareness about certain health conditions and disseminate knowledge of possible therapies among patients and healthcare providers. By accessing up-to-date and accurate information, healthcare providers can adapt their patients treatment in a more informed and knowledgeable way. However, medical content being highly sensitive, the generation process can be relatively slow (from days to weeks), and may go through numerous peer-review cycles, with thorough regulatory compliance and evaluation protocols.
Could LLMs, with their advanced text generation capabilities, help streamline this process by assisting brand managers and medical experts in their generation and review process?
To answer this question, the AWS Generative AI Innovation Center recently developed an AI assistant for medical content generation. The system is built upon Amazon Bedrock and leverages LLM capabilities to generate curated medical content for disease awareness. With this AI assistant, we can effectively reduce the overall generation time from weeks to hours, while giving the subject matter experts (SMEs) more control over the generation process. This is accomplished through anautomatedrevisionfunctionality, which allows the user to interact and send instructions and comments directly to the LLM via an interactive feedback loop. This is especially important since the revision of content is usually the main bottleneck in the process.
Since every piece of medical information can profoundly impact the well-being of patients, medical content generation comes with additional requirements and hinges upon the contents accuracy and precision. For this reason, our system has been augmented with additional guardrails for fact-checking and rules evaluation. The goal of these modules is to assess the factuality of the generated text and its alignment with pre-specified rules and regulations. With these additional features, you have more transparency and control over the underlying generative logic of the LLM.
This post walks you through the implementation details and design choices, focusing primarily on thecontent generationandrevision modules. Fact-checking and rules evaluation require special coverage and will be discussed in an upcoming post.
Image1:High-level overview of the AI-assistant and its different components
The overall architecture and the main steps in the content creation process are illustrated inImage 2.The solution has been designed using the following services:
Image 2: Content generation steps
The workflow is as follows:
To generate accurate medical content, the LLM is provided with a set of curated scientific data related to the disease in question, e.g. medical journals, articles, websites, etc. These articles are chosen by brand managers, medical experts and other SMEs with adequate medical expertise.
The input also consists of a brief, which describesthe general requirements and rules the generated content should adhere to (tone, style, target audience, number of words, etc.). In the traditional marketing content generation process, this brief is usually sent to content creation agencies.
It is also possible to integrate more elaborate rules or regulations, such as the HIPAA privacy guidelines for theprotection of health information privacy and security. Moreover, these rules can either be general and universally applicable or they can be more specific to certain cases. For example, some regulatory requirements may apply to some markets/regions or a particular disease. Our generative system allows a high degree of personalization so you can easily tailor and specialize the content to new settings, by simply adjusting the input data.
The content should be carefully adapted to the target audience, either patients or healthcare professionals. Indeed, the tone, style, and scientific complexity should be chosen depending on the readers familiarity with medical concepts.The content personalization is incredibly important for HCLS companies with a large geographical footprint, as it enables synergies and yields more efficiencies across regional teams.
From a system design perspective, we may need to process a large number of curated articles and scientific journals. This is especially true if the disease in question requires sophisticated medical knowledge or relies on more recent publications. Moreover, medical references contain a variety of information, structured in either plain text or more complex images, with embedded annotations and tables. To scale the system, it is important to seamlessly parse, extract, and store this information. For this purpose, we use Amazon Textract, a machine learning (ML) service for entity recognition and extraction.
Once the input data is processed, it is sent to the LLM as contextual information through API calls. With a context window as large as 200K tokens for Anthropic Claude 3, we can choose to either use the original scientific corpus, hence improving the quality of the generated content (though at the price of increased latency), or summarize the scientific references before using them in the generative pipeline.
Medical reference summarization is an essential step in the overall performance optimization and is achieved by leveraging LLM summarization capabilities. We use prompt engineering to send our summarization instructions to the LLM. Importantly, when performed, summarization should preserve as much articles metadata as possible, such as the title, authors, date, etc.
Image 3: A simplified version of the summarization prompt
To start the generative pipeline, the user can upload their input data to the UI. This will trigger the Textract and optionally, the summarization Lambda functions, which, upon completion, will write the processed data to an S3 bucket. Any subsequent Lambda function can read its input data directly from S3. By reading data from S3, we avoid throttling issues usually encountered with Websockets when dealing with large payloads.
Image 4: A high-level schematic of the content generation pipeline
Our solution relies primarily on prompt engineering to interact with Bedrock LLMs. All the inputs (articles, briefs and rules) are provided as parameters to the LLM via a LangChain PrompteTemplate object. We can guide the LLM further with few-shot examples illustrating, for instance, the citation styles. Fine-tuning in particular, Parameter-Efficient Fine-Tuning techniques can specialize the LLM further to the medical knowledge and will be explored at a later stage.
Image 5: A simplified schematic of the content generation prompt
Our pipeline is multilingual in the sense it can generate content in different languages. Claude 3, for example, has been trained on dozens of different languages besides English and can translate content between them. However, we recognize that in some cases, the complexity of the target language may require a specialized tool, in which case, we may resort to an additional translation step using Amazon Translate.
Image 6: Animation showing the generation of an article on Ehlers-Danlos syndrome, its causes, symptoms, and complications
Revision is an important capability in our solution because it enables you to further tune the generated content by iteratively prompting the LLM with feedback. Since the solution has been designed primarily as an assistant, these feedback loops allow our tool to seamlessly integrate with existing processes, hence effectively assisting SMEs in the design of accurate medical content. The user can, for instance, enforce a rule that has not been perfectly applied by the LLM in a previous version, or simply improve the clarity and accuracy of some sections. The revision can be applied to the whole text. Alternatively, the user can choose to correct individual paragraphs. In both cases, the revised version and the feedback are appended to a new prompt and sent to the LLM for processing.
Image 7: A simplified version of the content revision prompt
Upon submission of the instructions to the LLM, a Lambda function triggers a new content generation process with the updated prompt. To preserve the overall syntactic coherence, it is preferable to re-generate the whole article, keeping the other paragraphs untouched. However, one can improve the process by re-generating only those sections for which feedback has been provided. In this case, proper attention should be paid to the consistency of the text. This revision process can be applied recursively, by improving upon the previous versions, until the content is deemed satisfactory by the user.
Image 8: Animation showing the revision of the Ehlers-Danlos article. The user can ask, for example, for additional information
With the recent improvements in the quality of LLM-generated text, generative AI has become a transformative technology with the potential to streamline and optimize a wide range of processes and businesses.
Medical content generation for disease awareness is a key illustration of how LLMs can be leveraged to generate curated and high-quality marketing content in hours instead of weeks, hence yielding a substantial operational improvement andenabling more synergies between regional teams. Through its revision feature, our solution canbe seamlessly integrated with existing traditional processes, making it a genuine assistant tool empowering medical experts and brand managers.
Marketing content for disease awareness is also a landmark example of a highly regulated use case, where precision and accuracy of the generated content are critically important. To enable SMEs to detect and correct any possible hallucination and erroneous statements, we designed a factuality checking module with the purpose of detecting potential misalignment in the generated text with respect to source references.
Furthermore, our rule evaluation feature can help SMEs with the MLR process by automatically highlighting any inadequate implementation of rules or regulations. With these complementary guardrails, we ensure both scalability and robustness of our generative pipeline, and consequently, the safe and responsible deployment of AI in industrial and real-world settings.
Sarah Boufelja Y. is a Sr. Data Scientist with 8+ years of experience in Data Science and Machine Learning. In her role at the GenAII Center, she worked with key stakeholders to address their Business problems using the tools of machine learning and generative AI. Her expertise lies at the intersection of Machine Learning, Probability Theory and Optimal Transport.
Liza (Elizaveta) Zinovyeva is an Applied Scientist at AWS Generative AI Innovation Center and is based in Berlin. She helps customers across different industries to integrate Generative AI into their existing applications and workflows. She is passionate about AI/ML, finance and software security topics. In her spare time, she enjoys spending time with her family, sports, learning new technologies, and table quizzes.
Nikita Kozodoi is an Applied Scientist at the AWS Generative AI Innovation Center, where he builds and advances generative AI and ML solutions to solve real-world business problems for customers across industries. In his spare time, he loves playing beach volleyball.
Marion Eigneris a Generative AI Strategist who has led the launch of multiple Generative AI solutions. With expertise across enterprise transformation and product innovation, she specializes in empowering businesses to rapidly prototype, launch, and scale new products and services leveraging Generative AI.
Nuno Castro is a Sr. Applied Science Manager at AWS Generative AI Innovation Center. He leads Generative AI customer engagements, helping AWS customers find the most impactful use case from ideation, prototype through to production. Hes has 17 years experience in the field in industries such as finance, manufacturing, and travel, leading ML teams for 10 years.
Aiham Taleb, PhD, is an Applied Scientist at the Generative AI Innovation Center, working directly with AWS enterprise customers to leverage Gen AI across several high-impact use cases. Aiham has a PhD in unsupervised representation learning, and has industry experience that spans across various machine learning applications, including computer vision, natural language processing, and medical imaging.
See more here:
Medical content creation in the age of generative AI | Amazon Web Services - AWS Blog
The Machine Learning Guide for Predictive Accuracy: Interpolation and Extrapolation – Towards Data Science
class ModelFitterAndVisualizer: def __init__(self, X_train, y_train, y_truth, scaling=False, random_state=41): """ Initialize the ModelFitterAndVisualizer class with training and testing data.
Parameters: X_train (pd.DataFrame): Training data features y_train (pd.Series): Training data target y_truth (pd.Series): Ground truth for predictions scaling (bool): Flag to indicate if scaling should be applied random_state (int): Seed for random number generation """ self.X_train = X_train self.y_train = y_train self.y_truth = y_truth
self.initialize_models(random_state)
self.scaling = scaling
# Initialize models # ----------------------------------------------------------------- def initialize_models(self, random_state): """ Initialize the models to be used for fitting and prediction.
Parameters: random_state (int): Seed for random number generation """
# Define kernel for GPR kernel = 1.0 * RBF(length_scale=1.0) + WhiteKernel(noise_level=1.0)
# Define Ensemble Models Estimator # Decision Tree + Kernel Method estimators_rf_svr = [ ('rf', RandomForestRegressor(n_estimators=30, random_state=random_state)), ('svr', SVR(kernel='rbf')), ] estimators_rf_gpr = [ ('rf', RandomForestRegressor(n_estimators=30, random_state=random_state)), ('gpr', GaussianProcessRegressor(kernel=kernel, normalize_y=True, random_state=random_state)) ] # Decision Trees estimators_rf_xgb = [ ('rf', RandomForestRegressor(n_estimators=30, random_state=random_state)), ('xgb', xgb.XGBRegressor(random_state=random_state)), ]
self.models = [ SymbolicRegressor(random_state=random_state), SVR(kernel='rbf'), GaussianProcessRegressor(kernel=kernel, normalize_y=True, random_state=random_state), DecisionTreeRegressor(random_state=random_state), RandomForestRegressor(random_state=random_state), xgb.XGBRegressor(random_state=random_state), lgbm.LGBMRegressor(n_estimators=50, num_leaves=10, min_child_samples=3, random_state=random_state), VotingRegressor(estimators=estimators_rf_svr), StackingRegressor(estimators=estimators_rf_svr, final_estimator=RandomForestRegressor(random_state=random_state)), VotingRegressor(estimators=estimators_rf_gpr), StackingRegressor(estimators=estimators_rf_gpr, final_estimator=RandomForestRegressor(random_state=random_state)), VotingRegressor(estimators=estimators_rf_xgb), StackingRegressor(estimators=estimators_rf_xgb, final_estimator=RandomForestRegressor(random_state=random_state)), ]
# Define graph titles self.titles = [ "Ground Truth", "Training Points", "SymbolicRegressor", "SVR", "GPR", "DecisionTree", "RForest", "XGBoost", "LGBM", "Vote_rf_svr", "Stack_rf_svr__rf", "Vote_rf_gpr", "Stack_rf_gpr__rf", "Vote_rf_xgb", "Stack_rf_xgb__rf", ]
def fit_models(self): """ Fit the models to the training data.
Returns: self: Instance of the class with fitted models """ if self.scaling: scaler_X = MinMaxScaler() self.X_train_scaled = scaler_X.fit_transform(self.X_train) else: self.X_train_scaled = self.X_train.copy()
for model in self.models: model.fit(self.X_train_scaled, self.y_train) return self
def visualize_surface(self, x0, x1, width=400, height=500, num_panel_columns=5, vertical_spacing=0.06, horizontal_spacing=0, output=None, display=False, return_fig=False): """ Visualize the prediction surface for each model.
Parameters: x0 (np.ndarray): Meshgrid for feature 1 x1 (np.ndarray): Meshgrid for feature 2 width (int): Width of the plot height (int): Height of the plot output (str): File path to save the plot display (bool): Flag to display the plot """
num_plots = len(self.models) + 2 num_panel_rows = num_plots // num_panel_columns
whole_width = width * num_panel_columns whole_height = height * num_panel_rows
specs = [[{'type': 'surface'} for _ in range(num_panel_columns)] for _ in range(num_panel_rows)] fig = make_subplots(rows=num_panel_rows, cols=num_panel_columns, specs=specs, subplot_titles=self.titles, vertical_spacing=vertical_spacing, horizontal_spacing=horizontal_spacing)
for i, model in enumerate([None, None] + self.models): # Assign the subplot panels row = i // num_panel_columns + 1 col = i % num_panel_columns + 1
# Plot training points if i == 1: fig.add_trace(go.Scatter3d(x=self.X_train[:, 0], y=self.X_train[:, 1], z=self.y_train, mode='markers', marker=dict(size=2, color='darkslategray'), name='Training Data'), row=row, col=col)
surface = go.Surface(z=self.y_truth, x=x0, y=x1, showscale=False, opacity=.4) fig.add_trace(surface, row=row, col=col)
# Plot predicted surface for each model and ground truth else: y_pred = self.y_truth if model is None else model.predict(np.c_[x0.ravel(), x1.ravel()]).reshape(x0.shape) surface = go.Surface(z=y_pred, x=x0, y=x1, showscale=False) fig.add_trace(surface, row=row, col=col)
fig.update_scenes(dict( xaxis_title='x0', yaxis_title='x1', zaxis_title='y', ), row=row, col=col)
fig.update_layout(title='Model Predictions and Ground Truth', width=whole_width, height=whole_height)
# Change camera angle camera = dict( up=dict(x=0, y=0, z=1), center=dict(x=0, y=0, z=0), eye=dict(x=-1.25, y=-1.25, z=2) ) for i in range(num_plots): fig.update_layout(**{f'scene{i+1}_camera': camera})
if display: fig.show()
if output: fig.write_html(output)
if return_fig: return fig
Read this article:
The Machine Learning Guide for Predictive Accuracy: Interpolation and Extrapolation - Towards Data Science
Machine learning-based decision support model for selecting intra-arterial therapies for unresectable hepatocellular … – Nature.com
Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram L, Jemal A, et al. Global Cancer Statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2021;71:20949. https://doi.org/10.3322/caac.21660.
Article CAS PubMed Google Scholar
Marrero JA, Kulik LM, Sirlin CB, Zhu AX, Finn RS, Abecassis MM, et al. Diagnosis, staging, and management of hepatocellular Carcinoma: 2018 Practice Guidance by the American Association for the Study of Liver Diseases. Hepatology. 2018;68:72350. https://doi.org/10.1002/hep.29913.
Article PubMed Google Scholar
Villanueva A. Hepatocellular Carcinoma. N. Engl J Med. 2019;380:145062. https://doi.org/10.1056/NEJMra1713263.
Article CAS PubMed Google Scholar
Park JW, Chen M, Colombo M, Roberts LR, Schwartz M, Chen PJ, et al. Global patterns of hepatocellular carcinoma management from diagnosis to death: the BRIDGE Study. Liver Int. 2015;35:215566. https://doi.org/10.1111/liv.12818.
Article PubMed PubMed Central Google Scholar
Ghanaati H, Mohammadifard M, Mohammadifard M. A review of applying transarterial chemoembolization (TACE) method for management of hepatocellular carcinoma. J Fam Med Prim Care. 2021;10:355360. https://doi.org/10.4103/jfmpc.jfmpc_2347_20.
Article Google Scholar
Sidaway P. HAIC-FO improves outcomes in HCC. Nat Rev Clin Oncol. 2022;19:150. https://doi.org/10.1038/s41571-022-00599-0.
Article CAS PubMed Google Scholar
He M, Li Q, Zou R, Shen JX, Fang WQ, Tan GS, et al. Sorafenib plus hepatic arterial infusion of Oxaliplatin, Fluorouracil, and Leucovorin vs Sorafenib alone for hepatocellular carcinoma with portal vein invasion: a randomized clinical trial. JAMA Oncol. 2019;5:95360. https://doi.org/10.1001/jamaoncol.2019.0250.
Article PubMed PubMed Central Google Scholar
Zhang Z, Li C, Liao W, Huang Y, Wang Z A Combination of Sorafenib, an Immune Checkpoint Inhibitor, TACE and Stereotactic Body Radiation Therapy versus Sorafenib and TACE in Advanced Hepatocellular Carcinoma Accompanied by Portal Vein Tumor Thrombus. Cancers. 2022;14. https://doi.org/10.3390/cancers14153619.
Lencioni R, Llovet JM, Han G, Tak WY, Yang JM, Alfredo G, et al. Sorafenib or placebo plus TACE with doxorubicin-eluting beads for intermediate stage HCC: The SPACE trial. J Hepatol. 2016;64:10908. https://doi.org/10.1016/j.jhep.2016.01.012.
Article CAS PubMed Google Scholar
McGlynn KA, Petrick JL, El-Serag HB. Epidemiology of Hepatocellular Carcinoma. Hepatology. 2021;73:413. https://doi.org/10.1002/hep.31288.
Article CAS PubMed Google Scholar
An C, Zuo M, Li W, Chen Q, Wu P. Infiltrative Hepatocellular Carcinoma: Transcatheter arterial chemoembolization versus hepatic arterial infusion chemotherapy. Front Oncol. 2021;11:747496. https://doi.org/10.3389/fonc.2021.747496.
Article CAS PubMed PubMed Central Google Scholar
An C, Yao W, Zuo M, Li W, Chen Q, Wu P Pseudo-capsulated hepatocellular carcinoma: hepatic arterial infusion chemotherapy versus Transcatheter Arterial Chemoembolization. Acad Radiol. 2023. https://doi.org/10.1016/j.acra.2023.06.021.
Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Comput Struct Biotechnol J. 2015;13:817. https://doi.org/10.1016/j.csbj.2014.11.005.
Article CAS PubMed Google Scholar
An C, Yang H, Yu X, Han Z, Cheng Z, Liu F, et al. A machine learning model based on health records for predicting recurrence after microwave ablation of hepatocellular carcinoma. J Hepatocell Carcinoma. 2022;9:67184. https://doi.org/10.2147/JHC.S358197.
Article PubMed PubMed Central Google Scholar
Uche-Anya E, Anyane-Yeboa A, Berzin TM, Ghassemi M, May FP. Artificial intelligence in gastroenterology and hepatology: how to advance clinical practice while ensuring health equity. Gut. 2022;71:190915. https://doi.org/10.1136/gutjnl-2021-326271.
Article PubMed Google Scholar
Moons KGM, Altman DG, Reitsma JB, Ioannidis JPA, Macaskill P, Steyerberg EW. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. 2015;162:W173. https://doi.org/10.7326/M14-0698. Jan 6
Article PubMed Google Scholar
EASL Clinical Practice Guidelines: Management of hepatocellular carcinoma. J Hepatol. 2018;69:182-236. https://doi.org/10.1016/j.jhep.2018.03.019.
Cardella JF, Kundu S, Miller DL, Millward SF, Sacks D. Society of Interventional Radiology Clinical Practice Guidelines. J Vasc Inter Radio. 2009;20:S189191. https://doi.org/10.1016/j.jvir.2009.04.035.
Article Google Scholar
Liu W, Wei R, Chen J, Li Y, Pang H, Zhang W, et al. Prognosis prediction and risk stratification of transarterial chemoembolization or intraarterial chemotherapy for unresectable hepatocellular carcinoma based on machine learning. Eur Radiol. 2024 Jan 30. https://doi.org/10.1007/s00330-024-10581-2.
Article PubMed Google Scholar
Wang K, Tian J, Zheng C, Yang H, Ren J, Liu Y, et al. Interpretable prediction of 3-year all-cause mortality in patients with heart failure caused by coronary heart disease based on machine learning and SHAP. Comput Biol Med. 2021;137:104813. https://doi.org/10.1016/j.compbiomed.2021.104813.
Article PubMed Google Scholar
Ma M, Liu R, Wen C, Xu W, Xu Z, Wang S, et al. Predicting the molecular subtype of breast cancer and identifying interpretable imaging features using machine learning algorithms. Eur Radiol. 2022;32:165262. https://doi.org/10.1007/s00330-021-08271-4.
Article CAS PubMed Google Scholar
Li QJ, He MK, Chen HW, Fang WQ, Zhou WM, Liu X, et al. Hepatic arterial infusion of Oxaliplatin, Fluorouracil, and Leucovorin versus transarterial chemoembolization for large hepatocellular carcinoma: a randomized Phase III trial. J Clin Oncol. 2022;40:15060. https://doi.org/10.1200/JCO.21.00608.
Article CAS PubMed Google Scholar
Jin ZC, Zhong BY, Chen JJ, Zhu HD, Sun JH, Yin GW, et al. Real-world efficacy and safety of TACE plus camrelizumab and apatinib in patients with HCC (CHANCE2211): a propensity score matching study. Eur Radiol. 2023. https://doi.org/10.1007/s00330-023-09754-2.
Johnson PJ, Berhane S, Kagebayashi C, Shinji S, Mabel T, Helen LR, et al. Assessment of liver function in patients with hepatocellular carcinoma: a new evidence-based approach-the ALBI grade. J Clin Oncol. 2015;33:5508. https://doi.org/10.1200/JCO.2014.57.9151.
Article PubMed Google Scholar
Reig M, Forner A, Rimola J, Joana F, Marta B, ngeles G, et al. BCLC strategy for prognosis prediction and treatment recommendation: The 2022 update. J Hepatol. 2022;76:68193. https://doi.org/10.1016/j.jhep.2021.11.018.
Article PubMed Google Scholar
Song S, Bai M, Li X, Guo S, Yang W, Li C, et al. Early predictive value of circulating biomarkers for sorafenib in advanced hepatocellular carcinoma. Expert Rev Mol Diagn. 2022;22:36178. https://doi.org/10.1080/14737159.2022.2049248.
Article CAS PubMed Google Scholar
Hiraoka A, Ishimaru Y, Kawasaki H, Aibiki T, Okudaira T, Toshimori A, et al. Tumor Markers AFP, AFP-L3, and DCP in Hepatocellular Carcinoma Refractory to Transcatheter Arterial Chemoembolization. Oncology. 2015;89:16774. https://doi.org/10.1159/000381808.
Article CAS PubMed Google Scholar
Zhou H, Song T. Conversion therapy and maintenance therapy for primary hepatocellular carcinoma. Biosci Trends. 2021;15:15560. https://doi.org/10.5582/bst.2021.01091.
Article CAS PubMed Google Scholar
Fan J, Tang ZY, Yu YQ, Wu ZQ, Ma ZC, Zhou XD, et al. Improved survival with resection after transcatheter arterial chemoembolization (TACE) for unresectable hepatocellular carcinoma. Dig Surg. 1998;15:6748. https://doi.org/10.1159/000018676.
Article CAS PubMed Google Scholar
Shi F, Lian S, Mai Q, Mo ZQ, Zhuang WH, Cui W, et al. Microwave ablation after downstaging of hepatocellular carcinoma: outcome was similar to tumor within Milan criteria. Eur Radiol. 2020;30:245462. https://doi.org/10.1007/s00330-019-06604-y.
Article PubMed Google Scholar
Binnewies M, Roberts EW, Kersten K, Vincent C, Douglas FF, Miriam M, et al. Understanding the tumor immune microenvironment (TIME) for effective therapy. Nat Med. 2018;24:54150. https://doi.org/10.1038/s41591-018-0014-x.
Article CAS PubMed PubMed Central Google Scholar
Cao J, Su B, Peng R, Tang H, Tu DY, Tang YH, et al. Bioinformatics analysis of immune infiltrates and tripartite motif (TRIM) family genes in hepatocellular carcinoma. J Gastrointest Oncol. 2022;13:194258. https://doi.org/10.21037/jgo-22-619.
Article PubMed PubMed Central Google Scholar
Liu F, Liu D, Wang K, Xie XH, Su LY, Kuang M, et al. Deep learning radiomics based on contrast-enhanced ultrasound might optimize curative treatments for very-early or early-stage hepatocellular carcinoma patients. Liver Cancer. 2020;9:397413. https://doi.org/10.1159/000505694.
Article PubMed PubMed Central Google Scholar
Ding W, Wang Z, Liu FY, Cheng ZG, Yu XL, Han ZY, et al. A hybrid machine learning model based on semantic information can optimize treatment decision for nave single 3-5-cm HCC patients. Liver Cancer. 2022;11:25667. https://doi.org/10.1159/000522123.
Article CAS PubMed PubMed Central Google Scholar
See the rest here:
Machine learning-based decision support model for selecting intra-arterial therapies for unresectable hepatocellular ... - Nature.com
Here are 7 free AI classes you can take online from top tech firms, universities – Fortune
Almost a quarter of global jobs is expected to change within the next five years thanks to AI, and with only a small percentage of workers with skills in this field, the rush to learn the ins-and-outs of AI is ever more important.
ADVERTISEMENT
Join a global network of thought leaders and innovators. Understand the changes required across your organization to drive successful AI adoption.
AI is providing people with on-demand learning anywhere they are at any time of day on any day, says Jared Curham, a professor of work and organizational studies at MITs Sloan School of Management. Curhan recently launched two new AI-powered courses focused on the world of strategic negotiation and says that the technology is overall making education more accessible with personalized feedback and coaching.
While there are an increasing number of full-fledged AI degree programs, including within business schools, some students may be looking for a simpler or self-paced route. If youre interested in learning more about this in-demand field, several top tech firms and universities offer free online courses that serve as an introduction to AI technologies.
Amazon has more than 100 free and low-cost AI courses and learning resources available through AWS. Learners can obtain the basic skills in machine learning, generative AI, and foundational models. As a whole, the company has a commitment to provide free AI skills training to 2 million people by 2025.
The machine learning plan has nearly seven hours of free content in which individuals can learn the foundations of the technology, including relevant terminology, and decision-making processes. It also teaches users how to utilize Amazon SageMaker, the companys machine learning platform used by companies like AT&T and LG.
Google offers a beginner course for anyone who may be interested in how AI is being used in the real world. Google AI for Everyone, which is offered through online education platform edX, is a self-paced course that takes about four weeks to complete, assuming you dedicate two-to-three hours per week to the course. Participants learn about both AI and machine-learning principles and real-world applications of the technologies.
Google also covers what AI programming looks like and the process of teaching a computer how to learn. The course is taught by Laurence Moroney, who leads AI Advocacy at Google as part of the Google Research into Machine Intelligence (RMI) team. Nearly 12,000 people have enrolled in this free online course, according to edX.
If youre one of the 5.7 million people who has taken Harvard Universitys CS50 Introduction to Computer Science course through edX, then the universitys introductory AI class might be the best option for you. CS50, which is one of the most popular free online courses of all time, is a prerequisite for Harvards Introduction to Artificial Intelligence with Python course.
This seven-week course covers AI algorithms, game-playing engines, handwriting recognition, and machine translation. Students have to commit between 10 and 30 hours per week to complete the course, which includes hands-on projects and lectures. The course is taught by David J. Malan, a renowned computer scientist and Harvard professor.
IBM, which is recognized as a revolutionary leader in emerging technologies, offers an AI Foundations for Everyone specialization through Coursera. The specialization includes three courses:
The entire specialization takes about three months to complete, assuming you dedicate two hours per week to coursework. Students will learn the basics of what AI is, as well as its applications and ethical concerns. Theyll also hear from experts about starting a career in AI. The program is taught by Rav Ahuja and Antonio Cangiano, who work for IBMs Skills Network. Participants earn a certificate upon completion.
Intel has a goal to provide more than 30 million people with AI skills by 2030. As part of this commitment, the company provides dozens of free self-paced courses online on subjects such as deep learning for robotics, deep learning, and natural language processing.
Intel also has several AI Concepts educational pages that will walk you through definitions, real-world examples, tools, and resources for topics such as generative AI, AI inference, and transfer learning. Additionally, the company provides free on-demand webinars on more advanced AI use cases such as optimizing transformer models, optimizing AI workloads, and AI performance tuning.
As part of its Computational Social Science specialization through Coursera, the University of CaliforniaDavis offers a course focused on AI: Big Data, Artificial Intelligence, and Ethics. During this four-week course, participants learn about big data and its limitations, the history of artificial intelligence, and research ethics. The entire self-paced course takes about 12 total hours to complete.
The course is taught by Martin Hilbert, who is a professor at UC Davis and serves as a chair for computational social science. The course uses case studies to help participants learn AI concepts. More than 31,000 participants have completed this course, and those who do earn a certificate that can be shared on LinkedIn.
For someone who may be looking to break into AI or who wants to learn more about the applications of this new technology to different industries, the University of Pennsylvania offers a string of courses focused on artificial intelligence. The AI for Business specialization includes four courses:
These beginner courses take a total of about four months to complete and culminate in an applied learning project. Program participants complete peer-reviewed exercises to illustrate what theyve learned about data analytics, machine learning tools, and people management.The specialization is taught by eight UPenn professors from the Wharton School, a top-ranked business school by Fortune Education, and other professors from the university. The courses are offered through online education platform Coursera, and students can earn a certificate that can be displayed on their LinkedIn profile.
There is no one best course or program since AI is still so new. What ultimately matters is your curiosity to learn about AI, which you can do by working directly with prompt engineering or machine learning to gain hands-on skills.
You can certainly learn the foundations of AI in three monthsespecially if you already have a background in computer science. It is important to keep in mind that because AI is always changing and developing, you will need to keep up to date with the latest trends if you are looking to pursue a career focused on working with the technology.
Taking free AI courses on platforms such as Udemy or Codecademy is a great place to learn AI if youre a beginner. You can also learn AI by watching YouTube videos or reading through AI subreddits. The number of ways to learn AI are only growing, so there is ultimately no perfect path. Above all, just be curious, ask important questions, and dont be afraid to dive down rabbit holes
Check out all ofFortunesrankings of degree programs, and learn more about specific career paths.
Sydney Lake contributed to this piece.
Read the original:
Here are 7 free AI classes you can take online from top tech firms, universities - Fortune