Archive for the ‘Alphago’ Category

Putin and Xi seek to weaponize Artificial Intelligence against America – FOX Bangor/ABC 7 News and Stories

An open letter recently signed by Elon Musk, researchers from the Massachusetts Institute of Technology and Harvard University, and more than a thousand other prominent people set off alarm bells on advances in artificial intelligence (AI). The letter urged the worlds leading labs to hit the brakes on this powerful technology for six months because of the "profound risks to society and humanity."

A pause to consider the ramifications of this unpredictable new technology may have benefits. But our enemies will not wait while the U.S. engages in teleological discourse.

"By combining our wealth of research capacity and industrial capabilities, Russia and China can become world leaders in information technology, cybersecurity, and artificial intelligence (AI)," declared Russian President Vladimir Putin on March 21 during his meeting in Moscow with Chinese President Xi Jinping. The two authoritarian leaders vowed to usher in a new, anti-U.S. world order, and as their joint statement noted a "Deepening the Comprehensive Strategic Partnership of Coordination in the New Era," highlighted cooperation between Russia and China on AI.

AI is regarded as part of the fourth industrial revolution, which also includes the Internet of Things, genetic engineering, and quantum computing. Here is how Americas top adversaries, China and Russia, plan to weaponize this powerful tool against America.

CHINA WILL REQUIRE AI TO REFLECT SOCIALIST VALUES, NOT CHALLENGE SOCIAL ORDER

China codified its AI ambitions in the New Generation Artificial Intelligence Development Plan, which it adopted in July 2017. China had its AI awakening moment a year prior, according to Kaifu Li, ex-director of Google China.On March 19, 2016, Google DeepMinds artificial intelligence program AlphaGo defeated South Koreas Lee Sedol, the world champion in Go, the ancient Chinese game, in a highly anticipated match at the Four Seasons Hotel in Seouls Gwanghwamun district. Most South Korean TV networks were covering the event as 60 million Chinese tuned in and 100,000 English-speaking viewers watched YouTubes livestream. That a computer could beat the world champion shocked the Chinese. Sixteen months later, the Chinese Communist Party vowed that Beijing will lead the world of AI by 2030.

Chinas AI strategy centers on three primary goals: domestic surveillance, economic advancement and future warfare. The Chinese government is already using AI-driven software dubbed "one person, one file," that collects and stores vast amounts of data on its residents, in order to evaluate loyalty and risk to the regime. A giant network of surveillance cameras the Chinese authorities call "sharp eyes" tracks everyone continuously. Americans who travel to China, especially business executives and government officials, need to be aware of the risks associated with this blanket 24/7 monitoring.

When it comes to military applications, Chinas strategic ambitions for AI are what the CCP calls "intelligentized" and "informatized" warfare. Chinas Ministry of National Defense has established two research centers to execute this mission the Artificial Intelligence Research Center and the Unmanned Systems Research Center. The Peoples Liberation Armys (PLA) tasked its Academy of Military Science with ensuring that the PLAs warfighting doctrine is fully capitalized on disruptive technologies like AI and autonomous systems.

AFTER XI-PUTIN MEETING, TEAM BIDEN STILL DOESN'T GET WHAT'S JUST HAPPENED TO THE UNITED STATES

The United States is the primary target of Chinas AI-enabled warfare doctrine, as it is the only country that stands in the way of Chinas long-held policy goal of securing control over Taiwan. The CCP has decided that instead of following the track of U.S. military modernization, something Chinese military theorists view as linear trajectory, China will pursue "leapfrog development" of AI and autonomous technologies.

The PLA views AI technology as a "trump card" weapon that could be used in multiple ways to target perceived U.S. vulnerabilities, including U.S. battle networks and Americas way of war in general. An AI-enabled "swarming" tactic, for example is one of the approaches China could use to target and saturate the defenses of U.S. aircraft carriers.

AI swarming is a high-tech version of flooding U.S. airspace, in the run-up to an invasion of Taiwan, with hundreds of weaponized air balloons, of the kind that it recently flew across America. This would overwhelm the detection and defense capabilities of the U.S. North American Aerospace Defense Command (NORAD.) How many F-22s and $400,000 AIM-9X Sidewinder missiles would be needed to down them all?

The speed of Chinas progress in AI is of grave concern to the Pentagon and U.S. intelligence. In March, the U.S. Defense Intelligence Agency warned that China is "investing heavily in its AI and ML [machine learning] capabilities."

The 2023 Annual Threat Assessment by the Office of the Director of National Intelligence characterized Chinas AI and big data analytics capabilities as "rapidly expanding and improving," saying China is on track to "expand beyond domestic use." China is already an "AI peer in many areas and an AI leader in some applications," according to the 2021 Final Report by the U.S. National Security Commission on Artificial Intelligence. The report warned that "Chinas plans, resources, and progress should concern all Americans" and highlighted the importance of winning the "intensifying strategic competition" with China, which is determined to surpass the United States in the next few years.

CLICK HERE TO GET THE OPINION NEWSLETTER

Russia is lagging behind China and the U.S. in AI, but Moscow also seeks to become one of the world leaders in this novel technology. In 2017, Putin famously proclaimed "whichever country becomes the leader in artificial intelligence will become the ruler of the world."

In October 2019, Vladmir Putin approved Russia's "National Strategy for the Development of Artificial Intelligence to 2030" and directed his Cabinet to report annually about the progress of its implementation. Last year, Putin escalated his prioritization of AI. "Artificial intelligence technologies should be massively implemented in all industries in Russia this decade," he stated at the AI Journal Conference in Moscow in November 2022, urging Russia's researchers to "create breakthrough technologies of a new era." Russia's "place in the world, sovereignty, and security" depend on the results it achieves in AI, he said.

Russias AI strategy is primarily focused on robotics, robot-human interaction and counter-drone warfare. Russian military strategists believe that the expanding role of unmanned aerial vehicles (UAVs) in modern warfare necessitates the development of "UAV-killing UAV" systems. AI is also viewed by Russian strategists as a perfect technology to enable Moscows doctrine of "controlled chaos" as a way of deterring Washingtonfrom intervening in a conflict, such as the one in Ukraine. The doctrine envisions the targeting of the U.S. homeland with AI-enabled crippling cyber-attacks and spreading false information that could cause panic and disrupt the normal functioning of the society.

Russian doctrinal writings talk about "inspiring crisis" in an adversarys state by deploying AI-enabled cyber weapons and information operations in the run-up to a conflict. Using an "artificially maintained" crisis to trigger "aggravating factors such as dissatisfaction with existing government," would create a destabilizing effect on the opponent, pointing their focus inward and away from what Russia is doing, hypothesize Russian strategists.

As U.S. leaders make decisions regarding Americas pace of development in AI, they must remember that Russia and China are not only accelerating the speed of their AI research, they also plan to join forces to make critical gains in it. The goal is to create a new anti-U.S. world order, destabilize the U.S. from within, and defeat America on the battlefield if necessary. Now is not the time to cede our competitive advantage in AI to our top adversaries.

CLICK HERE TO READ MORE FROM REBEKAH KOFFLER

View original post here:
Putin and Xi seek to weaponize Artificial Intelligence against America - FOX Bangor/ABC 7 News and Stories

The Future of Generative Large Language Models and Potential … – JD Supra

[co-author: Geoffrey Vance, Perkins Coie LLP]

[Foreword by Geoffrey Vance: Although this article is technically co-authored by Jan and me, the vast majority of the technical discussion is Jans work. And thats the point. Lawyers arent supposed to be the Captains of the application of generative AI in the legal industry. A lawyers role is more like that of a First Officer, whose responsibility is to assist the primary pilot in getting to the ultimate destination. This article and all the exciting parts within it demonstrate how important data scientists and advanced technology are to the legal profession. The law firms and lawyers who dont reach that understanding fast will be left behind. Those who do will lead the future of the legal industry.]

Contents

Why is Human Language so Hard to Understand for Computer Programs?What are Large Language Models?The Zoo of Transformer Models: BERT and GPTChatGPTs LimitationsHow to Improve Large Language ModelsIntegrating LLM with existing Legal TechnologyUnderstand the Decisions; XAI

Human language is difficult for computer programs to understand because it is inherently complex, ambiguous, and context-dependent. Unlike computer languages, which are based on strict rules and syntax, human language is nuanced and can vary greatly based on the speaker, the situation, and the cultural context. As a result, building computer programs that can accurately understand and interpret human language is exceptionally complex and has been an ongoing challenge for artificial intelligence researchers since AI was first introduced. This is exactly the reason why it took so long for humans (in many of our lifetimes) to create reliable computer programs to deal with human language.

In addition, for many different reasons, early language models took shortcuts and none of them addressed all linguistic challenges. It was not until Google introduced the Transformer model in 2017 in the ground-breaking paper Attention is all you need that a full encoder-decoder model, using multiple layers of self-attention, resulted in a model capable of understanding almost all of the linguistic challenges. The model soon outperformed all other models on various linguistic tasks such as translation, Q&A, classification, text-analytics.

Before we dive into the specifics of large language models, lets first look at the basic definition. Large Language Models are artificial intelligence models that can generate human-like language based on a large amount of data they have been trained on. They use deep learning algorithms to analyze vast amounts of text, learning patterns and relationships between words, phrases, and concepts.

Some of the most well-known LLMs are the GPT series of models developed by OpenAI, BERT developed by Google, and T5 developed by Google Brain.

As encoder-decoder models such as the T5 model are very large and hard to train due to a lack of aligned training data, a variety of cut-down models (also called a zoo of transformer models) have been created. The two best known models are: BERT and GPT.

ChatGPT is an extension of GPT. It is based on the latest version of GPT (3.5) and has been fine-tuned for human-computer dialog using reinforcement learning. In addition, it is capable of sticking to human ethical values by using several additional mechanisms. These two capabilities are major achievements!

The core reason ChatGPT is so good is because transformers are the first computational models that take almost all linguistic phenomena seriously. Based on Googles transformers, OpenAI (with the help of Microsoft) has shaken up the world by introducing a model that can generate language that can no longer be distinguished from human language.

Much to our chagrin, ChatGPT is not the all-knowing General Artificial Intelligence most would like it to be. This is mainly due to the decoder-only architecture. ChatGPT is great for chatting, but one cannot control the factuality. This is due to the lack of an encoder mechanism. The longer the chats, the higher the odds that ChatGPT will get off-track or start hallucinating. Being a statistical process, this is a logical consequence: longer sequences are harder to control or predict than shorter ones.

Using ChatGPT on its own for anything else than just casual chit-chatting, is not wise. Using it for legal or medical advice without human validation of the factuality of such advice is just dangerous.

The AI research is aware of this, and there are a number of on-going approaches to improve todays models:

Currently, the Artificial Intelligence industry is working on all of the above improvements. In addition, one can also expect integrations with other forms of human perception: vision and speech. As you may not know, OpenAI is also the creator of Whisper, the state of the art Speech recognition for 100s of languages and DALL-E2, the well-known image generator, so adding speech to the mix is only a matter of time.

If you made it this far, you should by now understand that ChatGPT is not by itself a search engine, nor an eDiscovery data reviewer, a translator, knowledge base, or tool for legal analytics. But it can contribute to these functionalities.

Full-text search is one of the most important tools for legal professionals. It is an integral part of every piece of legal software, assisting lawyers in case law search, legal fact finding, document template search, among other tasks.

Todays typical workflow involved formulating a (Boolean) query, ranking results on some form or relevancy (jurisdiction, date, relevance, source, etc.), reviewing the results, and selecting the ones that matter. As the average query length on Google is only 1.2 words, we expect our search engine to find the most relevant hits with very little information. Defining the query can be hard and will always include human bias (the results one gets depends on the keywords used). What is more, reviewing the results of the search query can be time consuming, and one never knows what one misses. This is where Chatbots can help: by changing the search process into an AI-driven dialogue, we can change the whole search experience.

This is exactly what Microsoft does with the BING ChatGPT integration, but with a few risks in the current implementation:

As explained earlier, more focus on explaining where the results come from, the ability to eliminate information and a better understanding of the meaning of the text used to drive the dialogue is probably needed to get better results. Especially when we plan to use this for legal search, we need more transparency and understanding where the results come from.

Contract drafting is likely one of the most promising applications of textual generative artificial intelligence (AI) because contracts are typically highly structured documents that contain specific legal language, terms, and conditions. These documents are often lengthy, complex, and require a high degree of precision, making them time-consuming and expensive to produce.

Textual generative AI models can assist in the drafting of contracts by generating language that conforms to legal standards and meets specific requirements. By analyzing vast amounts of legal data and identifying patterns in legal language, these models can produce contract clauses and provisions that are consistent with legal norms and best practices.

Furthermore, AI-generated contract language can help ensure consistency and accuracy across multiple documents, reduce the risk of errors and omissions, and streamline the contract drafting process. This can save time and money for lawyers and businesses alike, while also reducing the potential for disputes and litigation.

But, here too, we need to do more vertical training, and probably more controlled text generation by understanding and incorporating the structure of legal documents in the text-generation process.

In all cases, it is important to note that AI-generated contract language should be reviewed by a qualified lawyer to ensure that it complies with applicable laws and regulations, and accurately reflects the parties intentions. While AI can assist in the drafting process, it cannot replace the expertise and judgment of a human lawyer.

We have serious doubts if generative Artificial Intelligence can be used as it is and provide help in providing meaningful legal advice. AI models lack the ability to provide personalized advice based on a clients specific circumstances, or to consider the ethical and moral dimensions of a legal issue. Legal advice requires a deep understanding of the law and the ability to apply legal principles to a particular situation. Text generation models do not have this knowledge. So, without additional frameworks capable of storing and understanding such knowledge, using models such as ChatGPT is a random walk in the court.

E-discovery is a process that involves the identification, collection, preservation, review, and production of electronically stored information (ESI) in the context of legal proceedings. While e-discovery often involves searching for specific information or documents, it is more accurately described as a sorting and classification process, rather than a search process.

The reason for this is that e-discovery involves the review and analysis of large volumes of data, often from a variety of sources and in different formats. ChatGPT is unable to handle the native formats this data is in.

The sorting and classification process in e-discovery is critical because it allows legal teams to identify and review relevant documents efficiently and accurately, while also complying with legal requirements for the preservation and production of ESI. Without this process, legal teams would be forced to manually review large volumes of data, which would be time-consuming, costly, and prone to error.

In summary, e-discovery is a sorting and classification process because it involves the review and analysis of large volumes of data, and the classification and organization of that data in a way that is relevant to the legal matter at hand. While searching for specific information is a part of eDiscovery, it is only one aspect of a larger process.

ChatGPT is neither a sorting, nor a text analytical or search tool. Models such as BERT or text-classification models based on word-embeddings or TF-IDF in combination with Support Vector Machines are better, faster, and better understood for Assisted Review and Active Learning.

Where Generative AI can help, is in the expansion of search queries. As we all know, humans are always biased. When humans define (Boolean) search queries, the search keywords chosen by human operators are subject to this bias. Generative AI can be very beneficial assisting users defining a search query and come up with keywords an end-user would not have thought of. This increases recall and limits human bias.

Legal documents can be lengthy and often contain boiler plate text. Summarization can provide a quick overview of the most important aspects of such a document. GPT is very good at summarization tasks. This can assist reviewers or project managers to get faster understanding of documents in eDiscovery.

As an AI language model, ChatGPT could be used to draft written responses to eDiscovery requests or provide suggested language for meet and confer sessions. However, it cannot provide personalized legal advice or make strategic decisions based on the specific circumstances of a case.

eDiscovery platforms enrich, filter, order and sort ESI into understandable structures. Such structures are used to generate reports. Reports can be in either structured formats (tables and graphs), or in the form of description in natural language. The latter can easily be generated from the ESI database by using generative AI to create a more human form of communication.

Here too, we can state that ChatGPT is not a text analytical or search tool. Straight forward search engines (using keyword, fuzzy and regular expression search), or advanced text-classification models such as BERT are better, faster and better understood for compliance monitoring and information governance purposes.

Nobody is more interested in explainable Artificial Intelligence (XAI) than DARPA, the Defense Advanced Research Projects Agency. Already in 2016, DARPA started an XAI program.

Ever since, DARPA has sponsored various research projects related to XAI, including the development of algorithms and models that can generate explanations for their decisions, the creation of benchmark datasets for testing XAI systems, and the exploration of new methods for evaluating the explainability and transparency of AI systems.

XAI is one of the hottest areas of research in the AI community. Without XAI, the application of artificial intelligence is unthinkable in areas such as finance, legal, medical or military.

XAI, refers to the development of AI systems that can provide clear and transparent explanations for their decision-making processes. Unlike traditional black-box AI systems, which are difficult or impossible to interpret, XAI systems aim to provide human-understandable explanations for their behavior.

XAI is not a single technology or approach, but rather a broad research area that includes various techniques and methods for achieving explainability in AI systems. Some approaches to XAI include rule-based systems, which use explicit rules to generate decisions that can be easily understood by humans; model-based systems, which use machine learning models that are designed to be interpretable and explainable; and hybrid systems, which combine multiple techniques to achieve a balance between accuracy and explainability.

The development of XAI is an active area of research, with many academic and industry researchers working to develop new techniques and tools for achieving transparency and explainability in AI systems. Ultimately, the goal of XAI is to promote the development of AI systems that are not only accurate and efficient, but also transparent and trustworthy, allowing humans to understand and control the decision-making processes of the AI system.

For legal applications, a full XAI framework is essential. Without XAI, there can also not be legal defensibility or trust.

Vaswani, Ashish, et al. Attention is all you need. Advances in neural information processing systems 30 (2017).

Devlin, Jacob, et al. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).

Tenney, Ian, Dipanjan Das, and Ellie Pavlick. BERT rediscovers the classical NLP pipeline. arXiv preprint arXiv:1905.05950 (2019).

Radford, Alec, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. Language Models are Unsupervised Multitask Learners. (2019). GPT-2.

Language Models are Few-Shot Learners, Tom B. Brown et al., arXiv:2005.14165, July 2020. GPT-3.

Ouyang, Long, et al. Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155 (2022).

Tegmark, Max (2017). Life 3.0 : being human in the age of artificial intelligence (First ed.). New York: Knopf.

Russell, Stuart (2017-08-31). Artificial intelligence: The future is superintelligent. Nature. 548 (7669): 520521. Bibcode:2017Natur.548..520R. doi:10.1038/548520a. ISSN 0028-0836.

Russell, Stuart, Human Compatible. 2019.

[1] Textual adversarial attacks are a type of cyber-attack that involves modifying or manipulating textual data in order to deceive or mislead machine learning models.

[View source.]

Read the rest here:
The Future of Generative Large Language Models and Potential ... - JD Supra

A Chatbot Beat the SAT. What Now? – The Atlantic

Last fall, when generative AI abruptly started turning out competent high-school- and college-level writing, some educators saw it as an opportunity. Perhaps it was time, at last, to dispose of the five-paragraph essay, among other bad teaching practices that have lingered for generations. Universities and colleges convened emergency town halls before winter terms began to discuss how large language models might reshape their work, for better and worse.

But just as quickly, most of those efforts evaporated into the reality of normal life. Educators and administrators have so many problems to address even before AI enters the picture; the prospect of utterly redesigning writing education and assessment felt impossible. Worthwhile, but maybe later. Then, with last weeks arrival of GPT-4, came another provocation. OpenAI, the company that created the new software, put out a paper touting its capacities. Among them: taking tests. AIs are no longer just producing passable five-paragraph essays. Now theyre excelling at the SAT, earning a score of 1410. Theyre getting passing grades on more than a dozen different AP exams. Theyre doing well enough on bar exams to be licensed as lawyers.

It would be nice if this news inspired educators, governments, certification agencies, and other groups to rethink what these tests really meanor even to reinvent them altogether. Alas, as was the case for rote-essay writing, whatever appetite for change the shock inspires might prove to be short-lived. GPT-4s achievements help reveal the underlying problem: Americans love standardized tests as much as we hate themand were unlikely to let them go even if doing so would be in our best interest.

Many of the initial responses to GPT-4s exam prowess were predictably immoderate: AI can keep up with human lawyers, or apply to Stanford, or make education useless. But why should it be startling in the slightest that software trained on the entire text of the internet performs well on standardized exams? AI can instantly run what amounts to an open-book test on any subject through statistical analysis and regression. Indeed, that anyone is surprised at all by this success suggests that people tend to get confused about what it means when computers prove effective at human activities.

Read: The college essay is dead

Back in the late 1990s, nobody thought a computer could ever beat a human at Go, the ancient Chinese game played with black and white stones. Chess had been mastered by supercomputers, but Go remainedat least in the hearts of its playersimmune to computation. They were wrong. Two decades later, DeepMinds AlphaGo was regularly beating Go masters. To accomplish this task, AlphaGo initially mimicked human players moves before running innumerable games against itself to find new strategies. The victory was construed by some as evidence that computers could overtake people at complex tasks previously thought to be uniquely human.

By rights, GPT-4s skill at the SAT should be taken as the opposite. Standardized tests feel inhuman from the start: You, a distinct individual, are forced to perform in a manner that can be judged by a machine, and then compared with that of many other individuals. Yet last weeks announcementof the 1410 score, the AP exams, and so ongave rise to an unease similar to that produced by AlphaGo.

Perhaps were anxious not that computers will strip us of humanity, but that machines will reveal the vanity of our human concerns. The experience of reasoning about your next set of moves in Go, as a human player doing so from the vantage point of human culture, cannot be replaced or reproduced by a Go-playing machineunless the only point of Go were to prove that Go can be mastered, rather than played. Such cultural values do exist: The designation of chess grand masters and Go 9-dan professionals suggests expertise in excess of mere performance in a folk game. The best players of chess and Go are sometimes seen as smart in a general sense, because they are good at a game that takes smarts of a certain sort. The same is true for AIs that play (and win) these games.

Read: A machine crushed us at Pokmon

Standardized tests occupy a similar cultural role. They were conceived to assess and communicate general performance on a subject such as math or reading. Whether and how they ever managed to do that is up for debate, but the accuracy and fairness of the exams became less important than their social function. To score a 1410 on the SAT says something about your capacities and prospectsmaybe you can get into Stanford. To pursue and then emerge victorious against a battery of AP tests suggests general ability warranting accelerated progress in college. (That victory doesnt necessarily provide that acceleration only emphasizes the seduction of its symbolism.) The bar exam measuresone hopessomeones subject-matter proficiency, but doesnt promise to ensure lawyerly effectiveness or even competence. To perform well on a standardized test indicates potential to perform well at some real future activity, but it has also come to have some value in itself, as a marker of success at taking tests.

That value was already being questioned, machine intelligence aside. Standardized tests have long been scrutinized for contributing to discrimination against minority and low-income students. The coronavirus pandemic, and its disruptions to educational opportunity, intensified those concerns. Many colleges and universities made the SAT and ACT optional for admissions. Graduate schools are giving up on the GRE, and aspiring law students may no longer have to take the LSAT in a couple of years.

GPT-4s purported prowess at these tests shows how little progress has been made at decoupling appearance from reality in the tests pursuit. Standardized tests might fairly assess human capacity, or they might do so unfairly, but either way, they hold an outsize role in Americans conception of themselves and their communities. Were nervous that tests might turn us into computers, but also that computers might reveal the conceit of valuing tests so much in the first place.

AI-based chess and Go computers didnt obsolesce play by people, but they did change human-training practices. Large language models may do the same for taking the SAT and other standardized exams, and evolve into a fancy form of test prep. In that case, they could end up helping those who would already have done well enough to score even higher. Or perhaps they will become the basis for a low-cost alternative that puts such training in the hands of everyonea reversal of examination inequity, and a democratization of vanity. No matter the case, the standardized tests will persist, only now the chatbots have to take them too.

Read more from the original source:
A Chatbot Beat the SAT. What Now? - The Atlantic

Exclusive: See the cover for Benjamn Labatut’s new novel, The … – Literary Hub

Literary Hub is pleased to reveal the cover for Benjamn Labatuts new novel, The MANIAC, which will be published by Penguin Press this fall. Heres how the publisher describes the book:

From one of contemporary literatures most exciting new voices, a haunting story centered on the Hungarian polymath John von Neumann, tracing the impact of his singular legacy on the dreams and nightmares of the twentieth century and the nascent age of AI

BenjamnLabatutsWhen We Cease to Understand the Worldelectrified a global readership. A Booker Prize and National Book Award finalist, and one of theNew York Times Ten Best Books of the Year, it explored the life and thought of a clutch of mathematicians and physicists who took science to strange and sometimes dangerous new realms. InThe MANIAC, Labatut has created a tour de force on an even grander scale.

A prodigy whose gifts terrified the people around him, John von Neumann transformed every field he touched, inventing game theory and the first programable computer, and pioneering AI, digital life, and cellular automata. Through a chorus of family members, friends, colleagues, and rivals, Labatut shows us the evolution of a mind unmatched and of a body of work that has unmoored the world in its wake.

The MANIACplaces von Neumann at the center of a literary triptych that begins with Paul Ehrenfest, an Austrian physicist and friend of Einstein, who fell into despair when he saw science and technology become tyrannical forces; it ends a hundred years later, in the showdown between the South Korean Go Master Lee Sedol and the AI program AlphaGo, an encounter embodying the central question of von Neumanns most ambitious unfinished project: the creation of a self-reproducing machine, an intelligence able to evolve beyond human understanding or control.

And heres the cover, which was designed by Darren Haggar:

The image on this cover was created by film director Bennett Miller, using OpenAIs DALL-E 2 software, the publisher told Lit Hub. He arrived at the final product by making extensive edits on variations of an image generated using the following prompt: a vintage photograph of huge plumes of smoke coming from an enormous UFO crashed in the desert.

The MANIAC will be published on October 3, 2023 by Penguin Press. You can preorder it here.

A show of Bennett Millers prints is currently on at the Gagosian Gallery through April 22, 2023.

Like Loading...

View post:
Exclusive: See the cover for Benjamn Labatut's new novel, The ... - Literary Hub

These companies are creating ChatGPT alternatives – Tech Monitor

Right now, major tech firms are clamouring to replicate the runaway success of ChatGPT, the generative AI chatbot developed by OpenAI using its GPT-3 large language model. Much like potential game-changers of the past, such as cloud-based Software as a Service (SaaS) platforms or blockchain technology (emphasis on potential), established companies and start-ups alike are going public with LLMs and ChatGPT alternatives in fear of being left behind.

While many of these will succeed some in the mass market and others with niche applications many more will likely fail as the market consolidates.

Who, then, are the companies in good stead to challenge OpenAI?

Googles LaMDA has attracted the most attention from mainstream observers of any LLM outside of GPT-3, but for not quite the same reasons.

Months before ChatGPT exploded into national headlines in late 2022, LaMDA was proving controversial after Google engineer Blake Lemoine was suspended for claiming falsely as became evident that it had developed sentience.

In reality, the LaMDA LLM operates similarly to its main competitor, except that it has fewer parameters at 137 billion compared to 175 billion for GPT-3.5, which was used to train ChatGPT.

LaMDA is also the bedrock of Googles chatbot competitor, named Bard, which the search giant is currently testing for search with select users. Bard had an inauspicious start, however, as it presented a factual error during a launch event.

Israel-based start-up AI21, while less well-known than its rival OpenAI, is a serious challenger in the market. The company created the Jurassic-2 large language model in 2021 with a similar number of parameters to GPT-3.5 178 billion compared to 175 billion and customisation capabilities.

March 2023 then saw the release of Jurassic-2, which focuses on optimised performance as opposed to size. According to AI21, the smallest version of Jurassic-2 outperforms even the largest version of its predecessor. It will also contain a grammatical correction API and text segmentation capabilities.

Users of AI21 studio can train their own versions of the LLM with as few as 50-100 training examples, which then become available for exclusive use.

AI21 also deployed Jurassic-1, and now Jurassic-2 to underpin its WordTune Spices chatbot, which distinguishes itself as a ChatGPT alternative by the use of live data retrieval and the citation of sources in its formulations. Given the risks of factual error and plagiarism associated with LLM chatbots, this is a significant advantage in an increasingly competitive field.

Founded by former OpenAI employees, Anthropic is fast making waves as a rival to its quasi-progenitor.

The generative AI company has launched its own large language model, Claude, whose ChatGPT alternative boasts what it calls constitutional AI. In effect, the model is designed to act according to programmed principles (i.e. its constitution) as opposed to ChatGPT, which is prohibited from answering certain controversial or dangerous queries.

Much like Microsofts investment in OpenAI, Google has invested $300m into Anthropic for a 10% stake in the company.

Baidu Chinas answer to Google is looking to combat its long-term struggles in the face of rival Tencent with its heavy investment in AI.

The team at Baidu has expanded its ERNIE 3.0 large language model into a new version called ERNIE 3.0 Titan. While its predecessor had just 10 billion parameters, Titans PaddlePaddle platform operates on 260 billion.

Titans creators claim that it is the largest dense pre-trained model so far and that it outperforms state-of-the-art models on natural language processing (NLP) tasks.

Hardware and software supplier Nvidia is currently core to the operation of ChatGPT, with an estimated 10,000 of the companys GPUs used to train the chatbot and a predicted 30,000 to be used in future.

This dynamic could be upended, however, as Nvidia CEO Jensen Huang announced in February 2023 that the firm plans to make its DGX AI supercomputer available via the cloud.

Already accessible through Oracle Cloud Infrastructure and Microsoft Azure, the AI supercomputer will have the capacity to allow customers to train their own large language models.

Nvidia has seen a financial boost as companies such as Google and Microsoft look to it for the GPUs necessary for training.

British AI company and Alphabet subsidiary Deepmind, famous for its AlphaGo program, is investing heavily in large language model research and development. Deepmind has iterated on multiple LLMs, including Gopher, Chinchilla and the RETRO system, which combines an LLM with an external database.

This experimentation is leading the way in more targeted and energy-efficient types of LLM Chinchilla has just 70 billion parameters, as opposed to others with double, triple or even more than that, yet outperforms the larger Gopher at certain tasks. Likewise for the 7.5 billion-parameter RETRO, whose external database allows it to outperform vastly larger models.

Not content to invest in the metaverse, Meta has also entered the LLM space with its LLaMA model. Mark Zuckerbergs company does not yet have a publicly available ChatGPT alternative but it is in development.

Unlike many others, the 65-billion parameter LLM has been made open source (upon request, crucially) with the intention of knowledge sharing and crowdsourcing bug fixes.

But just a week after it launched, a torrent for the LLM found its way to the wider internet via a 4Chan leak, prompting fears that such unfettered access could be used for phishing and other cybercrime activities.

Read this article:
These companies are creating ChatGPT alternatives - Tech Monitor