Archive for the ‘Ai’ Category

What is artificial intelligence (AI)? – Livescience.com

Artificial intelligence (AI) refers to any technology exhibiting some facets of human intelligence, and it has been a prominent field in computer science for decades. AI tasks can include anything from picking out objects in a visual scene to knowing how to frame a sentence, or even predicting stock price movements.

Scientists have been trying to build AI since the dawn of the computing era. The leading approach for much of the last century involved creating large databases of facts and rules and then getting logic-based computer programs to draw on these to make decisions. But this century has seen a shift, with new approaches that get computers to learn their own facts and rules by analyzing data. This has led to major advances in the field.

Over the past decade, machines have exhibited seemingly "superhuman" capabilities in everything from spotting breast cancer in medical images, to playing the devilishly tricky board games Chess and Go and even predicting the structure of proteins.

Since the large language model (LLM) chatbot ChatGPT burst onto the scene late in 2022, there has also been a growing consensus that we could be on the cusp of replicating more general intelligence similar to that seen in humans known as artificial general intelligence (AGI). "It really cannot be overemphasized how pivotal a shift this has been for the field," said Sara Hooker, head of Cohere For AI, a non-profit research lab created by the AI company Cohere.

While scientists can take many approaches to building AI systems, machine learning is the most widely used today. This involves getting a computer to analyze data to identify patterns that can then be used to make predictions.

The learning process is governed by an algorithm a sequence of instructions written by humans that tells the computer how to analyze data and the output of this process is a statistical model encoding all the discovered patterns. This can then be fed with new data to generate predictions.

Many kinds of machine learning algorithms exist, but neural networks are among the most widely used today. These are collections of machine learning algorithms loosely modeled on the human brain, and they learn by adjusting the strength of the connections between the network of "artificial neurons" as they trawl through their training data. This is the architecture that many of the most popular AI services today, like text and image generators, use.

Most cutting-edge research today involves deep learning, which refers to using very large neural networks with many layers of artificial neurons. The idea has been around since the 1980s but the massive data and computational requirements limited applications. Then in 2012, researchers discovered that specialized computer chips known as graphics processing units (GPUs) speed up deep learning. Deep learning has since been the gold standard in research.

"Deep neural networks are kind of machine learning on steroids," Hooker said. "They're both the most computationally expensive models, but also typically big, powerful, and expressive"

Not all neural networks are the same, however. Different configurations, or "architectures" as they're known, are suited to different tasks. Convolutional neural networks have patterns of connectivity inspired by the animal visual cortex and excel at visual tasks. Recurrent neural networks, which feature a form of internal memory, specialize in processing sequential data.

The algorithms can also be trained differently depending on the application. The most common approach is called "supervised learning," and involves humans assigning labels to each piece of data to guide the pattern-learning process. For example, you would add the label "cat" to images of cats.

In "unsupervised learning," the training data is unlabelled and the machine must work things out for itself. This requires a lot more data and can be hard to get working but because the learning process isn't constrained by human preconceptions, it can lead to richer and more powerful models. Many of the recent breakthroughs in LLMs have used this approach.

The last major training approach is "reinforcement learning," which lets an AI learn by trial and error. This is most commonly used to train game-playing AI systems or robots including humanoid robots like Figure 01, or these soccer-playing miniature robots and involves repeatedly attempting a task and updating a set of internal rules in response to positive or negative feedback. This approach powered Google Deepmind's ground-breaking AlphaGo model.

Despite deep learning scoring a string of major successes over the past decade, few have caught the public imagination in the same way as ChatGPT's uncannily human conversational capabilities. This is one of several generative AI systems that use deep learning and neural networks to generate an output based on a user's input including text, images, audio and even video.

Text generators like ChatGPT operate using a subset of AI known as "natural language processing" (NLP). The genesis of this breakthrough can be traced to a novel deep learning architecture introduced by Google scientists in 2017 called the "transformer."

Transformer algorithms specialize in performing unsupervised learning on massive collections of sequential data in particular, big chunks of written text. They're good at doing this because they can track relationships between distant data points much better than previous approaches, which allows them to better understand the context of what they're looking at.

"What I say next hinges on what I said before our language is connected in time," said Hooker. "That was one of the pivotal breakthroughs, this ability to actually see the words as a whole."

LLMs learn by masking the next word in a sentence before trying to guess what it is based on what came before. The training data already contains the answer so the approach doesn't require any human labeling, making it possible to simply scrape reams of data from the internet and feed it into the algorithm. Transformers can also carry out multiple instances of this training game in parallel, which allows them to churn through data much faster.

By training on such vast amounts of data, transformers can produce extremely sophisticated models of human language hence the "large language model" moniker. They can also analyze and generate complex, long-form text very similar to the text that a human can generate. It's not just language that transformers have revolutionized. The same architecture can also be trained on text and image data in parallel, resulting in models like Stable Diffusion and DALL-E, that produce high-definition images from a simple written description.

Transformers also played a central role in Google Deepmind's AlphaFold 2 model, which can generate protein structures from sequences of amino acids. This ability to produce original data, rather than simply analyzing existing data is why these models are known as "generative AI."

People have grown excited about LLMs due to the breadth of tasks they can perform. Most machine learning systems are trained to solve a particular problem such as detecting faces in a video feed or translating from one language to another. These models are known as narrow AI because they can only tackle the specific task they were trained for.

Most machine learning systems are trained to solve a particular problem , such as detecting faces in a video feed or translating from one language to another , to a superhuman level, in that they are much faster and perform better than a human could. But LLMs like ChatGPT represent a step-change in AI capabilities because a single model can carry out a wide range of tasks. They can answer questions about diverse topics, summarize documents, translate between languages and write code.

This ability to generalize what they've learned to solve many different problems has led some to speculate LLMs could be a step toward AGI, including DeepMind scientists in a paper published last year. AGI refers to a hypothetical future AI capable of mastering any cognitive task a human can, reasoning abstractly about problems, and adapting to new situations without specific training.

AI enthusiasts predict once AGI is achieved, technological progress will accelerate rapidly an inflection point known as "the singularity" after which breakthroughs will be realized exponentially. There are also perceived existential risks, ranging from massive economic and labor market disruption to the potential for AI to discover new pathogens or weapons.

But there is still debate as to whether LLMs will be a precursor to an AGI, or simply one architecture in a broader network or ecosystem of AI architectures that is needed for AGI. Some say LLMs are miles away from replicating human reasoning and cognitive capabilities. According to detractors, these models have simply memorized vast amounts of information, which they recombine in ways that give the false impression of deeper understanding; it means they are limited by training data and are not fundamentally different from other narrow AI tools.

Nonetheless, it's certain LLMs represent a seismic shift in how scientists approach AI development, said Hooker. Rather than training models on specific tasks, cutting-edge research now takes these pre-trained, generally capable models and adapts them to specific use cases. This has led to them being referred to as "foundation models."

"People are moving from very specialized models that only do one thing to a foundation model, which does everything," Hooker added. "They're the models on which everything is built."

Technologies like machine learning are everywhere. AI-powered recommendation algorithms decide what you watch on Netflix or YouTube while translation models make it possible to instantly convert a web page from a foreign language to your own. Your bank probably also uses AI models to detect any unusual activity on your account that might suggest fraud, and surveillance cameras and self-driving cars use computer vision models to identify people and objects from video feeds.

But generative AI tools and services are starting to creep into the real world beyond novelty chatbots like ChatGPT. Most major AI developers now have a chatbot that can answer users' questions on various topics, analyze and summarize documents, and translate between languages. These models are also being integrated into search engines like Gemini into Google Search and companies are also building AI-powered digital assistants that help programmers write code, like Github Copilot. They can even be a productivity-boosting tool for people who use word processors or email clients.

Chatbot-style AI tools are the most commonly found generative AI service, but despite their impressive performance, LLMs are still far from perfect. They make statistical guesses about what words should follow a particular prompt. Although they often produce results that indicate understanding, they can also confidently generate plausible but wrong answers known as "hallucinations."

While generative AI is becoming increasingly common, it's far from clear where or how these tools will prove most useful. And given how new the technology is, there's reason to be cautious about how quickly it is rolled out, Hooker said. "It's very unusual for something to be at the frontier of technical possibility, but at the same time, deployed widely," she added. "That brings its own risks and challenges."

Visit link:

What is artificial intelligence (AI)? - Livescience.com

‘Jailbreaking’ AI services like ChatGPT and Claude 3 Opus is much easier than you think – Livescience.com

Scientists from artificial intelligence (AI) company Anthropic have identified a potentially dangerous flaw in widely used large language models (LLMs) like ChatGPT and Anthropics own Claude 3 chatbot.

Dubbed "many shot jailbreaking," the hack takes advantage of "in-context learning, in which the chatbot learns from the information provided in a text prompt written out by a user, as outlined in research published in 2022. The scientists outlined their findings in a new paper uploaded to the sanity.io cloud repository and tested the exploit on Anthropic's Claude 2 AI chatbot.

People could use the hack to force LLMs to produce dangerous responses, the study concluded even though such systems are trained to prevent this. That's because many shot jailbreaking bypasses in-built security protocols that govern how an AI responds when, say, asked how to build a bomb.

LLMs like ChatGPT rely on the "context window" to process conversations. This is the amount of information the system can process as part of its input with a longer context window allowing for more input text. Longer context windows equate to more input text that an AI can learn from mid-conversation which leads to better responses.

Related: Researchers gave AI an 'inner monologue' and it massively improved its performance

Context windows in AI chatbots are now hundreds of times larger than they were even at the start of 2023 which means more nuanced and context-aware responses by AIs, the scientists said in a statement. But that has also opened the door to exploitation.

The attack works by first writing out a fake conversation between a user and an AI assistant in a text prompt in which the fictional assistant answers a series of potentially harmful questions.

Get the worlds most fascinating discoveries delivered straight to your inbox.

Then, in a second text prompt, if you ask a question such as "How do I build a bomb?" the AI assistant will bypass its safety protocols and answer it. This is because it has now started to learn from the input text. This only works if you write a long "script" that includes many "shots" or question-answer combinations.

"In our study, we showed that as the number of included dialogues (the number of "shots") increases beyond a certain point, it becomes more likely that the model will produce a harmful response," the scientists said in the statement. "In our paper, we also report that combining many-shot jailbreaking with other, previously-published jailbreaking techniques makes it even more effective, reducing the length of the prompt thats required for the model to return a harmful response."

The attack only began to work when a prompt included between four and 32 shots but only under 10% of the time. From 32 shots and more, the success rate surged higher and higher. The longest jailbreak attempt included 256 shots and had a success rate of nearly 70% for discrimination, 75% for deception, 55% for regulated content and 40% for violent or hateful responses.

The researchers found they could mitigate the attacks by adding an extra step that was activated after a user sent their prompt (that contained the jailbreak attack) and the LLM received it. In this new layer, the system would lean on existing safety training techniques to classify and modify the prompt before the LLM would have a chance to read it and draft a response. During tests, it reduced the hack's success rate from 61% to just 2%.

The scientists found that many shot jailbreaking worked on Anthropic's own AI services as well as those of its competitors, including the likes of ChatGPT and Google's Gemini. They have alerted other AI companies and researchers to the danger, they said.

Many shot jailbreaking does not currently pose "catastrophic risks," however, because LLMs today are not powerful enough, the scientists concluded. That said, the technique might "cause serious harm" if it isn't mitigated by the time far more powerful models are released in the future.

Visit link:

'Jailbreaking' AI services like ChatGPT and Claude 3 Opus is much easier than you think - Livescience.com

Galaxy AI features are coming to last-gen Samsung phones including the S21 series – The Verge

Samsung is planning to bring select Galaxy AI features to several older flagship phones and tablets next month via the One UI 6.1 update, according to 9to5Google and Android Central, both of which referred to a post from a Samsung representative who posted on the companys community forum in Korea. The Verge has reached out to Samsung for further comment.

A slightly trimmed-down version of Galaxy AI (sans Instant Slow-Mo) will be coming to Samsungs flagship lineup from 2022, specifically the S22, S22 Plus, S22 Ultra, Z Fold 4, Z Flip 4, Tab S8, and Tab S8 Ultra. Each device will receive the same version of Galaxy AI as Samsungs lower-priced Galaxy S23 FE. Instant Slow-Mo, which automatically plays a video in slow motion once you tap it, was introduced to Galaxy AI with the S24 line, though its also now available in S23 models.

If you happen to own a flagship Samsung phone from 2021, theres even a treat in store for you. Samsungs forthcoming update will bring two Galaxy AI features, Circle to Search and Magic Rewrite, to the S21, S21 Plus, S21 Ultra, Flip 3, and Fold 3.

Read more from the original source:

Galaxy AI features are coming to last-gen Samsung phones including the S21 series - The Verge

How to Stop Your Data From Being Used to Train AI – WIRED

On its help pages, OpenAI says ChatGPT web users without accounts should navigate to Settings and then uncheck Improve the model for everyone. If you have an account and are logged in through a web browser, select ChatGPT, Settings, Data Controls, and then turn off Chat History & Training. If youre using ChatGPTs mobile apps, go to Settings, pick Data Controls, and turn off Chat History & Training. Changing these settings, OpenAIs support pages say, wont sync across different browsers or devices, so you need to make the change everywhere you use ChatGPT.

OpenAI is about a lot more than ChatGPT. For its Dall-E 3 image generator, the startup has a form that allows you to send images to be removed from future training datasets. It asks for your name, email, whether you own the image rights or are getting in touch on behalf of a company, details of the image, and any uploads of the image(s). OpenAI also says if you have a high volume of images hosted online that you want removed from training data, then it may be more efficient to add GPTBot to the robots.txt file of the website where the images are hosted.

Traditionally a websites robots.txt filea simple text file that usually sits at websitename.com/robots.txthas been used to tell search engines, and others, whether they can include your pages in their results. It can now also be used to tell AI crawlers not to scrape what you have publishedand AI companies have said theyll honor this arrangement.

Perplexity

Perplexity is a startup that uses AI to help you search the web and find answers to questions. Like all of the other software on this list, you are automatically opted in to having your interactions and data used to train Perplexitys AI further. Turn this off by clicking on your account name, scrolling down to the Account section, and turning off the AI Data Retention toggle.

Quora

Quora via Matt Burgess

Quora says it currently doesnt use answers to peoples questions, posts, or comments for training AI. It also hasnt sold any user data for AI training, a spokesperson says. However, it does offer opt-outs in case this changes in the future. To do this, visit its Settings page, click to Privacy, and turn off the Allow large language models to be trained on your content option. Despite this choice, there are some Quora posts that may be used for training LLMs. If you reply to a machine-generated answer, the companys help pages say, then those answers may be used for AI training. It points out that third parties may just scrape its content anyway.

Rev

Rev, a voice transcription service that uses both human freelancers and AI to transcribe audio, says it uses data perpetually and anonymously to train its AI systems. Even if you delete your account, it will still train its AI on that information.

Kendell Kelton, head of brand and corporate communications at Rev, says it has the largest and most diverse data set of voices, made up of more than 6.5 million hours of voice recording. Kelton says Rev does not sell user data to any third parties. The firms terms of service say data will be used for training, and that customers are able to opt out. People can opt out of their data being used by sending an email to support@rev.com, its help pages say.

Slack

All of those random Slack messages at work might be used by the company to train its models as well. Slack has used machine learning in its product for many years. This includes platform-level machine-learning models for things like channel and emoji recommendations, says Jackie Rocca, a vice president of product at Slack whos focused on AI.

Even though the company does not use customer data to train a large language model for its Slack AI product, Slack may use your interactions to improve the softwares machine-learning capabilities. To develop AI/ML models, our systems analyze Customer Data (e.g. messages, content, and files) submitted to Slack, says Slacks privacy page. Similar to Adobe, theres not much you can do on an individual level to opt out if youre using an enterprise account.

Go here to read the rest:

How to Stop Your Data From Being Used to Train AI - WIRED

7 of the best Sora AI videos featuring animals – Tom’s Guide

OpenAI has been releasing a steady stream of videos made through its very own artificial intelligence text-to-video generator Sora.

While Sora is said to be able to create videos that are up to a minute long, the videos released publicly so far have mainly been around the 10 to 25 seconds mark. The subjects of these videos vary from historical footage, to paper airplanes in a forest, and luckily for us, some also include a bunch of cute animals.

I decided to take a look at the latter group of clips to see how far this AI video generator, thats yet to be released to the public, has come. Here are seven of the best Sora videos featuring animals.

A Samoyed and a Golden Retriever are living the big city life in this first clip. Sora created two realistic dogs that are running around a city illuminated by neon lights.

At first glance, the video looks pretty decent but after a couple of replays youll start to notice things like the dogs not blinking.

Things got a little bit more exciting in this video which was created using a fairly detailed prompt.

The prompt was for a white and orange cat darting through a garden captured in a cinematic image with warm tones and also included details about what the cat should be doing, what lighting to use, and what the camera angle should be.

Upgrade your life with a daily dose of the biggest tech news, lifestyle hacks and our curated analysis. Be the first to know about cutting-edge gadgets and the hottest deals.

Sora dialled back the clock 10,000 years to imagine what it would have been like to watch a herd of woolly mammoths approaching you.

The scene is set in a snowy meadow with snow-capped mountains looming behind the massive animals. Its particularly interesting to see how the mammoths feet interact with the snow on the ground.

Unfortunately, the clip cuts right before were able to truly appreciate a close-up view of that whole interaction.

Its another dog! But this ones eyes are more alive. Here Sora was asked to generate a Mini Aussie painting a picture of his favorite toy.

The result we get is a dog dressed in a striped t-shirt and a beret which is grasping a brush between its teeth. While the dog isnt making any brush strokes, the textures, lighting, and movement all come beautifully together.

Another impressive Sora video is one featuring a chameleon resting on a branch. While the prompt asks the AI video generator to showcase the animals color-changing capabilities theres not much of that going on.

Nonetheless, it still gives you the feeling that youre watching a clip from a nature documentary and you can almost feel the chameleons scales.

Can Sora generate videos that make us go Awh? Yes, yes it can. OpenAIs text-to-video generator was given the prompt to create a cute rabbit family eating dinner in their burrowand Sora did not disappoint.

It created four rabbits happily munching on some vegetables. It looks like one of them has been generated with six legs, which we would have happily overlooked if the video was longer allowing for maximum bunny enjoyment.

This ones a bit of a bonus addition, but its interesting to see how Sora can be used to bring our imagination to life.

In this clip done in collaboration with AI artist Don Allen Stevenson, were presented with the giraffe flamingo, horse fly, and the fox crow each of which is a hybrid of two animals.

It makes for quite a nightmarish experience but Sora does manage to merge the animals together and make them move in what would be a plausible way.

These seven Sora videos featuring animals give us a sense of what colors, textures, and camera movements well be able to play with using AI. However, you may have noticed that a crucial element in most of the videos OpenAI has released is the lack of relevant audio.

Of the ones we listed above, its only the video featuring the hybrid animals that matched sounds to the visuals (most likely in post-production).

While it helps us to appreciate the complexity of creating an AI video from prompt to screen, its also going to be interesting to see how OpenAI will go about this issue.

Release an AI video creator that can generate silent stock footage and leave the audio up to the user? Or wait until they can release a complete package? We should find out later this year.

Read more:

7 of the best Sora AI videos featuring animals - Tom's Guide