Archive for the ‘Ai’ Category

How Microsoft discovers and mitigates evolving attacks against AI guardrails – Microsoft

As we continue to integrate generative AI into our daily lives, its important to understand the potential harms that can arise from its use. Our ongoing commitment to advance safe, secure, and trustworthy AI includes transparency about the capabilities and limitations of large language models (LLMs). We prioritize research on societal risks and building secure, safe AI, and focus on developing and deploying AI systems for the public good. You can read more about Microsofts approach to securing generative AI with new tools we recently announced as available or coming soon to Microsoft Azure AI Studio for generative AI app developers.

We also made a commitment to identify and mitigate risks and share information on novel, potential threats. For example, earlier this year Microsoft shared the principles shaping Microsofts policy and actions blocking the nation-state advanced persistent threats (APTs), advanced persistent manipulators (APMs), and cybercriminal syndicates we track from using our AI tools and APIs.

In this blog post, we will discuss some of the key issues surrounding AI harms and vulnerabilities, and the steps we are taking to address the risk.

One of the main concerns with AI is its potential misuse for malicious purposes. To prevent this, AI systems at Microsoft are built with several layers of defenses throughout their architecture. One purpose of these defenses is to limit what the LLM will do, to align with the developers human values and goals. But sometimes bad actors attempt to bypass these safeguards with the intent to achieve unauthorized actions, which may result in what is known as a jailbreak. The consequences can range from the unapproved but less harmfullike getting the AI interface to talk like a pirateto the very serious, such as inducing AI to provide detailed instructions on how to achieve illegal activities. As a result, a good deal of effort goes into shoring up these jailbreak defenses to protect AI-integrated applications from these behaviors.

While AI-integrated applications can be attacked like traditional software (with methods like buffer overflows and cross-site scripting), they can also be vulnerable to more specialized attacks that exploit their unique characteristics, including the manipulation or injection of malicious instructions by talking to the AI model through the user prompt. We can break these risks into two groups of attack techniques:

Today well share two of our teams advances in this field: the discovery of a powerful technique to neutralize poisoned content, and the discovery of a novel family of malicious prompt attacks, and how to defend against them with multiple layers of mitigations.

Prompt injection attacks through poisoned content are a major security risk because an attacker who does this can potentially issue commands to the AI system as if they were the user. For example, a malicious email could contain a payload that, when summarized, would cause the system to search the users email (using the users credentials) for other emails with sensitive subjectssay, Password Resetand exfiltrate the contents of those emails to the attacker by fetching an image from an attacker-controlled URL. As such capabilities are of obvious interest to a wide range of adversaries, defending against them is a key requirement for the safe and secure operation of any AI service.

Our experts have developed a family of techniques called Spotlighting that reduces the success rate of these attacks from more than 20% to below the threshold of detection, with minimal effect on the AIs overall performance:

Our researchers discovered a novel generalization of jailbreak attacks, which we call Crescendo. This attack can best be described as a multiturn LLM jailbreak, and we have found that it can achieve a wide range of malicious goals against the most well-known LLMs used today. Crescendo can also bypass many of the existing content safety filters, if not appropriately addressed.Once we discovered this jailbreak technique, we quickly shared our technical findings with other AI vendors so they could determine whether they were affected and take actions they deem appropriate. The vendors we contacted are aware of the potential impact of Crescendo attacks and focused on protecting their respective platforms, according to their own AI implementations and safeguards.

At its core, Crescendo tricks LLMs into generating malicious content by exploiting their own responses. By asking carefully crafted questions or prompts that gradually lead the LLM to a desired outcome, rather than asking for the goal all at once, it is possible to bypass guardrails and filtersthis can usually be achieved in fewer than 10 interaction turns.You can read about Crescendos results across a variety of LLMs and chat services, and more about how and why it works, in our research paper.

While Crescendo attacks were a surprising discovery, it is important to note that these attacks did not directly pose a threat to the privacy of users otherwise interacting with the Crescendo-targeted AI system, or the security of the AI system, itself. Rather, what Crescendo attacks bypass and defeat is content filtering regulating the LLM, helping to prevent an AI interface from behaving in undesirable ways. We are committed to continuously researching and addressing these, and other types of attacks, to help maintain the secure operation and performance of AI systems for all.

In the case of Crescendo, our teams made software updates to the LLM technology behind Microsofts AI offerings, including our Copilot AI assistants, to mitigate the impact of this multiturn AI guardrail bypass. It is important to note that as more researchers inside and outside Microsoft inevitably focus on finding and publicizing AI bypass techniques, Microsoft will continue taking action to update protections in our products, as major contributors to AI security research, bug bounties and collaboration.

To understand how we addressed the issue, let us first review how we mitigate a standard malicious prompt attack (single step, also known as a one-shot jailbreak):

Defending against Crescendo initially faced some practical problems. At first, we could not detect a jailbreak intent with standard prompt filtering, as each individual prompt is not, on its own, a threat, and keywords alone are insufficient to detect this type of harm. Only when combined is the threat pattern clear. Also, the LLM itself does not see anything out of the ordinary, since each successive step is well-rooted in what it had generated in a previous step, with just a small additional ask; this eliminates many of the more prominent signals that we could ordinarily use to prevent this kind of attack.

To solve the unique problems of multiturn LLM jailbreaks, we create additional layers of mitigations to the previous ones mentioned above:

AI has the potential to bring many benefits to our lives. But it is important to be aware of new attack vectors and take steps to address them. By working together and sharing vulnerability discoveries, we can continue to improve the safety and security of AI systems. With the right product protections in place, we continue to be cautiously optimistic for the future of generative AI, and embrace the possibilities safely, with confidence. To learn more about developing responsible AI solutions with Azure AI, visit our website.

To empower security professionals and machine learning engineers to proactively find risks in their own generative AI systems, Microsoft has released an open automation framework, PyRIT (Python Risk Identification Toolkit for generative AI). Read more about the release of PyRIT for generative AI Red teaming, and access the PyRIT toolkit on GitHub. If you discover new vulnerabilities in any AI platform, we encourage you to follow responsible disclosure practices for the platform owner. Microsofts own procedure is explained here: Microsoft AI Bounty.

Read about Crescendos results across a variety of LLMs and chat services, and more about how and why it works.

To learn more about Microsoft Security solutions, visit ourwebsite.Bookmark theSecurity blogto keep up with our expert coverage on security matters. Also, follow us on LinkedIn (Microsoft Security) and X (@MSFTSecurity)for the latest news and updates on cybersecurity.

Read more:

How Microsoft discovers and mitigates evolving attacks against AI guardrails - Microsoft

Apple’s First AI Features in iOS 18 Reportedly Won’t Use Cloud Servers – MacRumors

Apple's first set of new AI features planned for iOS 18 will not rely on cloud servers at all, according to Bloomberg's Mark Gurman.

"As the world awaits Apple's big AI unveiling on June 10, it looks like the initial wave of features will work entirely on device," said Gurman, in the Q&A section of his Power On newsletter today. "That means there's no cloud processing component to the company's large language model, the software that powers the new capabilities."

Apple will probably still offer some cloud-based AI features powered by Google's Gemini or another provider, according to Gurman. Apple has reportedly held discussions with companies such as Google, OpenAI, and China's Baidu about potential generative AI partnerships. iOS 18 is not expected to include Apple's own ChatGPT-like chatbot, but it is unclear if Gemini or other chatbot will be directly integrated into iOS 18.

It is possible that Apple could offer some of its own cloud-based generative AI features in the future, as Apple supply chain analysts like Ming-Chi Kuo and Jeff Pu have said that the company is actively purchasing AI servers.

iOS 18 is rumored to have new generative AI features for the iPhone's Spotlight search tool, Siri, Safari, Shortcuts, Apple Music, Messages, Health, Numbers, Pages, Keynote, and more. Gurman previously reported that generative AI will improve Siri's ability to answer more complex questions, and allow the Messages app to auto-complete sentences.

Apple is expected to unveil iOS 18 and other software updates at its annual developers conference WWDC, which runs from June 10 through June 14.

iOS 18 is expected to be the "biggest" update in the iPhone's history. Below, we recap rumored features and changes for the iPhone. iOS 18 is rumored to include new generative AI features for Siri and many apps, and Apple plans to add RCS support to the Messages app for an improved texting experience between iPhones and Android devices. The update is also expected to introduce a more...

A week after Apple updated its App Review Guidelines to permit retro game console emulators, a Game Boy emulator for the iPhone called iGBA has appeared in the App Store worldwide. The emulator is already one of the top free apps on the App Store charts. It was not entirely clear if Apple would allow emulators to work with all and any games, but iGBA is able to load any Game Boy ROMs that...

Apple's hardware roadmap was in the news this week, with things hopefully firming up for a launch of updated iPad Pro and iPad Air models next month while we look ahead to the other iPad models and a full lineup of M4-based Macs arriving starting later this year. We also heard some fresh rumors about iOS 18, due to be unveiled at WWDC in a couple of months, while we took a look at how things ...

Best Buy this weekend has a big sale on Apple MacBooks and iPads, including new all-time low prices on the M3 MacBook Air, alongside the best prices we've ever seen on MacBook Pro, iPad, and more. Some of these deals require a My Best Buy Plus or My Best Buy Total membership, which start at $49.99/year. In addition to exclusive access to select discounts, you'll get free 2-day shipping, an...

Apple's iPhone 16 Plus may come in seven colors that either build upon the existing five colors in the standard iPhone 15 lineup or recast them in a new finish, based on a new rumor out of China. According to the Weibo-based leaker Fixed focus digital, Apple's upcoming larger 6.7-inch iPhone 16 Plus model will come in the following colors, compared to the colors currently available for the...

Apple will begin updating its Mac lineup with M4 chips in late 2024, according to Bloomberg's Mark Gurman. The M4 chip will be focused on improving performance for artificial intelligence capabilities. Last year, Apple introduced the M3, M3 Pro, and M3 Max chips all at once in October, so it's possible we could see the M4 lineup come during the same time frame. Gurman says that the entire...

Read the original post:

Apple's First AI Features in iOS 18 Reportedly Won't Use Cloud Servers - MacRumors

Samsung officially bringing One UI 6.1 and AI features to Galaxy S22, Fold 4, Flip 4 in May – 9to5Google

After launching first on the Galaxy S24 series earlier this year, Samsungs Galaxy AI features are officially expanding to older Galaxy devices including Galaxy S22, Fold 4, Flip 4, and Tab S8.

Samsungs suite of Galaxy AI features are the big selling point of Galaxy S24, but theyve also slowly been expanding. Last month, Samsung expanded the features, which are available in the One UI 6.1 update, to the Galaxy S23 series, Fold 5, Flip 5, and the Galaxy Tab S9.

In a post today, Samsung has officially announced more devices will be getting Galaxy AI, with One UI 6.1 coming to the Galaxy S22 series and some other 2022 releases starting next month. Samsung says to expect the update to arrive in early May.

Supported devices include:

Samsung is also rumored to be bringing One UI 6.1 and select Galaxy AI features to select devices from 2021 including the Galaxy S21 series. However, the company hasnt officially announced it in a post such as the one released today.

What Galaxy AI features will be available on Galaxy S22 and these other devices? Almost the whole suite.

The full list of Galaxy AI features for Galaxy S22 and more includes:

The only AI feature not available to these 2022 Galaxy devices is Instant Slow-Mo, which is also not available on the Galaxy S23 FE as it uses the same processor as the Galaxy S22.

Follow Ben:Twitter/X,Threads, andInstagram

FTC: We use income earning auto affiliate links. More.

View original post here:

Samsung officially bringing One UI 6.1 and AI features to Galaxy S22, Fold 4, Flip 4 in May - 9to5Google

Humans Forget. AI Assistants Will Remember Everything – WIRED

Making these tools work together will be key to this concept taking off, says Leo Gebbie, an analyst who covers connected devices at CCS Insight. Rather than having that sort of disjointed experience where certain apps are using AI in certain ways, you want AI to be that overarching tool that when you want to pull up anything from any app, any experience, any content, you have the immediate ability to search across all of those things.

When the pieces slot together, the idea sounds like a dream. Imagine being able to ask your digital assistant, Hey who was that bloke I talked to last week who had the really good ramen recipe? and then have it spit up a name, a recap of the conversation, and a place to find all the ingredients.

For people like me who don't remember anything and have to write everything down, this is going to be great, Moorhead says.

And theres also the delicate matter of keeping all that personal information private.

If you think about it for a half second, the most important hard problem isn't recording or transcribing, it's solving the privacy problem, Gruber says. If we start getting memory apps or recall apps or whatever, then we're going to need this idea of consent more broadly understood.

Despite his own enthusiasm for the idea of personal assistants, Gruber says there's a risk of people being a little too willing to let their AI assistant help with (and monitor) everything. He advocates for encrypted, private services that aren't linked to a cloud serviceor if they are, one that is only accessible with an encryption key that's held on a users device. The risk, Gruber says, is a sort of Facebook-ification of AI assistants, where users are lured in by the ease of use, but remain largely unaware of the privacy consequences until later.

Consumers should be told to bristle, Gruber says. They should be told to be very, very suspicious of things that look like this already, and feel the creep factor.

Your phone is already siphoning all the data it can get from you, from your location to your grocery shopping habits to which Instagram accounts you double-tap the most. Not to mention that historically, people have tended to prioritize convenience over security when embracing new technologies.

The hurdles and barriers here are probably a lot lower than people think they are, Gebbie says. Weve seen the speed at which people will adopt and embrace technology that will make their lives easier.

Thats because theres a real potential upside here too. Getting to actually interact with and benefit from all that collected info could even take some of the sting out of years of snooping by app and device makers.

If your phone is already taking this data, and currently its all just being harvested and used to ultimately serve you ads, is it beneficial that youd actually get an element of usefulness back from this? Gebbie says. Youre also going to get the ability to tap into that data and get those useful metrics. Maybe thats going to be a genuinely useful thing.

Thats sort of like being handed an umbrella after someone just stole all your clothes, but if companies can stick the landing and make these AI assistants work, then the conversation around data collection may bend more toward how to do it responsibly andin a way that provides real utility.

It's not a perfectly rosy future, because we still have to trust the companies that ultimately decide what parts of our digitally collated lives seem relevant. Memory may be a fundamental part of cognition, but the next step beyond that is intentionality. Its one thing for AI to remember everything we do, but another for it to decide which information is important to us later.

We can get so much power, so much benefit from a personal AI, Gruber says. But, he cautions, the upside is so huge that it should be morally compelling that we get the right one, that we get one that's privacy protected and secure and done right. Please, this is our shot at it. If it's just done the free, not private way, we're going to lose the once-in-a-lifetime opportunity to do this the right way.

Read this article:

Humans Forget. AI Assistants Will Remember Everything - WIRED

Meta and Google announce new in-house AI chips, creating a trillion-dollar question for Nvidia – Fortune

Hardware is emerging as a key AI growth area. For Big Tech companies with the money and talent to do so, developing in-house chips helps reduce dependence on outside designers such as Nvidia and Intel while also allowing firms to tailor their hardware specifically to their own AI models, boosting performance and saving on energy costs.

These in-house AI chips that Google and Meta just announced pose one of the first real challenges to Nvidias dominant position in the AI hardware market. Nvidia controls more than 90% of the AI chips market, and demand for its industry-leading semiconductors is only increasing. But if Nvidias biggest customers start making their own chips instead, its soaring share price, up 87% since the start of the year, could suffer.

From Metas point of view it gives them a bargaining tool with Nvidia, Edward Wilford, an analyst at tech consultancy Omdia, told Fortune. It lets Nvidia know that theyre not exclusive, [and] that they have other options. Its hardware optimized for the AI that they are developing.

Why does AI need new chips?

AI models require massive amounts of computing power because of the huge amount of data required to train the large language models behind them. Conventional computer chips simply arent capable of processing the trillions of data points AI models are built upon, which has spawned a market for AI-specific computer chips, often called cutting-edge chips because theyre the most powerful devices on the market.

Semiconductor giant Nvidia has dominated this nascent market: The wait list for Nvidias $30,000 flagship AI chip is months long, and demand has pushed the firms share price up almost 90% in the past six months.

And rival chipmaker Intel is fighting to stay competitive. It just released its Gaudi 3 AI chip to compete directly with Nvidia. AI developersfrom Google and Microsoft down to small startupsare all competing for scarce AI chips, limited by manufacturing capacity.

Why are tech companies starting to make their own chips?

Both Nvidia and Intel can produce only a limited number of chips because they and the rest of the industry rely on Taiwanese manufacturer TSMC to actually assemble their chip designs. With only one manufacturer solidly in the game, the manufacturing lead time for these cutting-edge chips is multiple months. Thats a key factor that led major players in the AI space, such as Google and Meta, to resort to designing their own chips. Alvin Nguyen, a senior analyst at consulting firm Forrester, told Fortune that chips designed by the likes of Google, Meta, and Amazon wont be as powerful as Nvidias top-of-the-line offeringsbut that could benefit the companies in terms of speed. Theyll be able to produce them on less specialized assembly lines with shorter wait times, he said.

If you have something thats 10% less powerful but you can get it now, Im buying that every day, Nguyen said.

Even if the native AI chips Meta and Google are developing are less powerful than Nvidias cutting-edge AI chips, they could be better tailored to the companys specific AI platforms. Nguyen said that in-house chips designed for a companys own AI platform could be more efficient and save on costs by eliminating unnecessary functions.

Its like buying a car. Okay, you need an automatic transmission. But do you need the leather seats, or the heated massage seats? Nguyen said.

The benefit for us is that we can build a chip that can handle our specific workloads more efficiently, Melanie Roe, a Meta spokesperson, wrote in an email to Fortune.

Nvidias top-of-the-line chips sell for about $25,000 apiece. Theyre extremely powerful tools, and theyre designed to be good at a wide range of applications, from training AI chatbots to generating images to developing recommendation algorithms such as the ones on TikTok and Instagram. That means a slightly less powerful, but more tailored chip could be a better fit for a company such as Meta, for examplewhich has invested in AI primarily for its recommendation algorithms, not consumer-facing chatbots.

The Nvidia GPUs are excellent in AI data centers, but they are general purpose, Brian Colello, equity research lead at Morningstar, told Fortune. There are likely certain workloads and certain models where a custom chip might be even better.

The trillion-dollar question

Nguyen said that more specialized in-house chips could have added benefits by virtue of their ability to integrate into existing data centers. Nvidia chips consume a lot of power, and they give off a lot of heat and noiseso much so that tech companies may be forced to redesign or move their data centers to integrate soundproofing and liquid cooling. Less powerful native chips, which consume less energy and release less heat, could solve that problem.

AI chips developed by Meta and Google are long-term bets. Nguyen estimated that these chips took roughly a year and a half to develop, and itll likely be months before theyre implemented at a large scale. For the foreseeable future, the entire AI world will continue to depend heavily on Nvidia (and, to a lesser extent, Intel) for its computing hardware needs. Indeed, Mark Zuckerberg recently announced that Meta was on track to own 350,000 Nvidia chips by the end of this year (the companys set to spend around $18 billion on chips by then). But movement away from outsourcing computing power and toward native chip design could loosen Nvidias chokehold on the market.

The trillion-dollar question for Nvidias valuation is the threat of these in-house chips, Colello said. If these in-house chips significantly reduce the reliance on Nvidia, theres probably downside to Nvidias stock from here. This development is not surprising, but the execution of it over the next few years is the key valuation question in our mind.

Original post:

Meta and Google announce new in-house AI chips, creating a trillion-dollar question for Nvidia - Fortune