Archive for the ‘Ai’ Category

OpenAI Unveils A.I. That Instantly Generates Eye-Popping Videos – The New York Times

In April, a New York start-up called Runway AI unveiled technology that let people generate videos, like a cow at a birthday party or a dog chatting on a smartphone, simply by typing a sentence into a box on a computer screen.

The four-second videos were blurry, choppy, distorted and disturbing. But they were a clear sign that artificial intelligence technologies would generate increasingly convincing videos in the months and years to come.

Just 10 months later, the San Francisco start-up OpenAI has unveiled a similar system that creates videos that look as if they were lifted from a Hollywood movie. A demonstration included short videos created in minutes of woolly mammoths trotting through a snowy meadow, a monster gazing at a melting candle and a Tokyo street scene seemingly shot by a camera swooping across the city.

OpenAI, the company behind the ChatGPT chatbot and the still-image generator DALL-E, is among the many companies racing to improve this kind of instant video generator, including start-ups like Runway and tech giants like Google and Meta, the owner of Facebook and Instagram. The technology could speed the work of seasoned moviemakers, while replacing less experienced digital artists entirely.

It could also become a quick and inexpensive way of creating online disinformation, making it even harder to tell whats real on the internet.

We are having trouble retrieving the article content.

Please enable JavaScript in your browser settings.

Thank you for your patience while we verify access. If you are in Reader mode please exit andlog intoyour Times account, orsubscribefor all of The Times.

Thank you for your patience while we verify access.

Already a subscriber?Log in.

Want all of The Times?Subscribe.

See the original post here:

OpenAI Unveils A.I. That Instantly Generates Eye-Popping Videos - The New York Times

Install open-source AI in a commercial robot and it’ll clean your room – Big Think

Using just open-source AIs, researchers got a commercial robot to find and move objects around a room it had never entered before. The bot isnt perfect, but it suggests we might not be as far from sharing our homes with domestic robots as experts previously believed.

Just completely impossible:Demo videos of robotscleaning kitchens,making snacks, anddoing other choresmight have you hoping your days of loading the dishwasher are numbered, but AI experts predict were stilla decade awayfrom handing even a fraction of our chores over to bots.

There is a very pervasive feeling in the [robotics] community that homes are difficult, robots are difficult, and combining homes and robots is just completely impossible, Mahi Shafiullah, a PhD student at NYU Courant,told MIT Technology Review.

Simply tell the robot what to pick and where to drop it in natural language, and it will do it.

Open-source, off-the-shelf:A major holdup in the home robot revolution is the fact that building a robot that could work inanyoneshome is a lot harder than training one to work in a controlled lab environment.

A new study co-led by Shafiullah and involving researchers from NYU and AI at Meta suggests we might be closer to domestic robots than we think, though.

Using only open-source software, they modified a commercially available robot so that it could move objects around a room it had never entered before on demand. They call the system OK-Robot, and detail the work in apapershared on the preprint server arXiv.

Simply tell the robot what to pick and where to drop it in natural language, and it will do it,tweetedLerrel Pinto, who co-led the study along with Shafiullah.

How it works:Thebot at the core of the OK-Robot system is calledStretch(you can buy one for just $19,950, plus shipping and taxes). Stretch has a wheeled base, a vertical pole, and a robotic arm that can slide up and down the pole. At the end of the arm is a gripper that allows the bot to grasp objects.

To turn the robot into something humans can talk to, the team equipped it with vision-language models (VLMs) AIs trained to understand both images and words as well as pre-trained navigation and grasping models.

They then created a 3D video of a room using the iPhone app Record3D and shared it with the robot that process took about six minutes. After that, they could give the robot a text command to move an object in the room to a new location, and it would locate the object and move it.

They tested OK-Robot in 10 rooms. In each room, they choose 10-20 objects that could fit in the robots gripper and told it to move them (one at a time) to another part of the room (Move the soda can to the box, Move the Takis on the desk to the nightstand, etc.).

Overall, the robot had a 58.5% success rate at completing the tasks. But in rooms that were less cluttered, its success rate was much higher: 82.4%.

Looking ahead:Even though OK-Robot can only do one thing (and doesnt always do it right), the fact that it relies on off-the-shelf models and doesnt require any special training to work in a new environment just a video of the room is pretty remarkable.

The next step for the team will beopen sourcing their codeso that others can build off of what theyve started and potentially help get domestic robots doing our chores sooner than predicted.

I think once people start believing home robots are possible, a lot more work will start happening in this space, said Shafiullah.

This article was originally published by our sister site, Freethink.

More here:

Install open-source AI in a commercial robot and it'll clean your room - Big Think

There’s AI, and Then There’s AGI: What You Need to Know to Tell the Difference – CNET

Imagine an AI that doesn't just answer questions like ChatGPT, but can make your morning coffee, do the dishes and care for your elderly parent while you're at work.

It's the future first envisioned by The Jetsons in 1962, and thanks to developments in AI, it finally seems feasible within the next decade.

But the implications extend far beyond an in-home Jarvis. That's why tech titans like Meta CEO Mark Zuckerberg want to take AI to this next level. Last month, he told The Verge his new goal is to build artificial general intelligence, or AGI. That puts him in the same league as ChatGPT-maker OpenAI and Google's DeepMind.

While Zuckerberg wants AGI to build into products to further connect with users, OpenAI and DeepMind have talked about the potential of AGI to benefit humanity. Regardless of their motivations, it's a big leap from the current state of AI, which is dominated by generative AI and chatbots. The latter have so far dazzled us with their writing skills, creative chops and seemingly endless answers (even if their responses aren't always accurate).

There is no standard definition for AGI, which leaves a lot open to interpretation and opinion. But it is safe to say AGI is closer to humanlike intelligence and encompasses a greater range of skills than most existing AIs. And it will have a profound impact on us.

But it has a long way to go before it fully emulates the human brain - not to mention the ability to make its own decisions. And so the current state of AGI could best be described as the Schrodinger's cat of AI: It simultaneously is and is not humanlike.

If you're wondering what all the fuss is about with AGI, this explainer is for you. Here's what you need to know.

Let's start with a term we've heard a lot in the last year: artificial intelligence. It's a branch of computer science thatsimulates aspects of human intelligence in machines.

Per Mark Riedl, professor in the Georgia Tech School of Interactive Computing and associate director of the Georgia Tech Machine Learning Center, AI is "the pursuit of algorithms and systems that emulate behaviors we think of as requiring intelligence."

That includes specific tasks like driving a car, planning a birthday party or writing code jobs that are already performed to a degree today by self-driving cars and more modest driving-assist features, or by assistants like ChatGPT if you give them the right prompt.

"These are things that we think that humans excel at and require cognition," Riedl added. "So any system that emulates those sorts of behaviors or automates those sorts of tasks can be considered artificial intelligence."

OpenAI's Dall-E 3 generative AI can create fanciful images like this spiky elecric guitar in front of a psychedelic green background. It uses GPT text processing to pump up your text prompts for more vivid, detailed results.

When an AI can perform a single task very well like, say, playing chess it's considered narrow intelligence. IBM's Watson, the question-answering AI that triumphed on Jeopardy in 2011, is perhaps the best-known example. Deep Blue, another IBM AI, was the chess-playing virtuoso that beat grandmaster Garry Kasparov in 1997.

But the thing about narrow intelligence is it can only do that one thing.

"It's not going to be able to play golf and it's not going to be able to drive a car," said Chirag Shah, a professor at the University of Washington. But Watson and Deep Blue can probably beat you at Jeopardy and chess, respectively.

Artificial general intelligence, on the other hand, is broader and harder to define.

AGI means a machine can do many things humans do or possibly all the things we do. It depends who you ask.

Human beings are the ultimate general intelligence because we are capable of doing so much: talking, driving, problem solving, writing and more.

Theoretically, an AGI would be able to perform these tasks indistinguishable from what Georgios-Alex Dimakis, a professor of engineering at the University of Texas, called "an extremely intelligent human."

But beyond the ability to match human proficiency, there is no consensus about what achievements merit the label. For some, the ability to perform a task as well as a person is in and of itself a sign of AGI. For others, AGI will only exist when it can do everything humans can do with their minds. And then there are those who believe it's somewhere in between.

Zuckerberg illustrated this fluidity in his interview with The Verge. "You can quibble about if general intelligence is akin to human-level intelligence, or is it like human-plus, or is it some far-future superintelligence," he said. "But to me, the important part is actually the breadth of it, which is that intelligence has all these different capabilities where you have to be able to reason and have intuition."

But the key is AGI is broad where AI is narrow.

The timeline for AGI is also up for debate.

Some say it's already here, or close. Others say it may never happen. Still more peg the estimate at five to 10 years DeepMind CEO Demis Hassabis is in this camp while yet others say it will be decades.

"My personal view is, no, it doesn't exist," Shah said.

He pointed to a March 2023 research paper from Microsoft, which referred to "sparks of AGI." The researchers said some of the conversations with recent large language models like GPT-4 are "starting to show that it actually understands things in a deeper way than simply answering questions," Shah said.

That means "you can actually have a free-form conversation with it like you would have with a human being," he added. What's more, the latest versions of chatbots like Google's Gemini and ChatGPT are capable of responding to more complex queries.

This ability does indeed point to AGI, if you accept the looser definition.

LLMs are a type of AI, fed content like books and news stories to first understand and then generate their own output text. LLMs are behind all the generative AI chatbots we know (and love?), like ChatGPT, Gemini, Microsoft Bing and Claude.ai.

What's interesting about LLMs is they aren't limited to one specific task. They can write poetry and plan vacations and even pass the bar exam, which means they can perform multiple tasks, another sign of AGI.

Then again, they are still prone to hallucinations, which occur when an LLM generates outputs that are incorrect or illogical. They are also subject to reasoning errors and gullibility and even provide different answers to the same question.

Hence the similarity to Schrodinger's cat, which in the thought experiment was simultaneously dead and alive until someone opened the box it was in to check.

This is perhaps the $100,000 question and another one that is hard to answer definitively.

If an AGI learns how to perform multiple household duties, we may finally have a Jetsons moment. There's also the potential for at-home assistants who understand you like a friend or family member and who can take care of you, which Shah said has huge potential for elder care.

And AGI will continue to influence the job market as it becomes capable of more and more tasks. That means more existing jobs are at risk, but the good news is new jobs will be created and opportunities will remain.

The short answer is no.

For starters, the ability to perform multiple tasks, as an AGI would, does not imply consciousness or self-will. And even if an AI had self-determination, the number of steps required to decide to wipe out humanity and then make progress toward that goal is too many to be realistically possible.

"There's a lot of things that I would say are not hard evidence or proof, but are working against that narrative [of robots killing us all someday]," Riedl said.

He also pointed to the issue of planning, which he defined as "thinking ahead into your own future to decide what to do to solve a problem that you've never solved before."

LLMs are trained on historical data and are very good at using old information like itineraries to address new problems, like how to plan a vacation.

But other problems require thinking about the future.

"How does an AI system think ahead and plan how to eliminate its adversaries when there is no historical information about that ever happening?" Riedl asked. "You would require planning and look ahead and hypotheticals that don't exist yet there's this big black hole of capabilities that humans can do that AI is just really, really bad at."

Dimakis, too, believes sentient robots killing us all has "a very low probability."

A much bigger risk is this technology ending up closed off within one or two big tech companies instead of being open like it is at universities.

"Having a monopoly or an oligopoly of one or two companies that are the only ones who have these new AI systems will be very bad for the economy because you'd have a huge concentration of technologies being built on top of these AI foundation models," Dimakis said. "And that is to me one of the biggest risks to consider in the immediate future."

AGI should not be confused with artificial super intelligence, which is an AI capable of making its own decisions. In other words, it is self-aware, or sentient. This is the AI many people fear now.

"You can think about any of these sci-fi stories and movies where you have robots and they have AI that are planning and thinking on their own," Shah said. "They're able to do things without being directed and can assume control completely on their own without any supervision."

But the good news is ASI is much further away than AGI. And so there's time to implement guardrails and guide or hinder its development.

That being said, Thorsten Joachims, a professor of computer science at Cornell, believes we will hold AI systems to higher standards than we hold ourselves and this may ultimately help us address some of society's shortcomings.

For example, humans commit crimes.

"We would never put up with it if an AI system did that," he said.

Joachims also pointed to decision-making, particularly in courts of law. Even well-educated and experienced professionals like judges pass down vastly different sentences for similar cases.

He believes we won't tolerate this kind of inconsistency in AI either. These higher standards will inform how AI systems are built and, in the end, they may not even look all that human.

In fact, AGI may ultimately help us solve problems we've long struggled with, like curing cancer. And even if that's the only thing a particular AI can do, that alone would be revolutionary.

"Maybe it cannot pass the Turing test" a standard method for assessing a computer's ability to pass as human "so maybe we wouldn't even consider it intelligent in any way, but certainly it would save billions of lives," said Adam Klivans, a professor of computer science at the University of Texas and director of the National Science Foundation's AI Institute for Foundations of Machine Learning. "It would be incredible."

In other words, AI can help us solve problems without fully mimicking human intelligence.

"These are not so much exactly AGI in the sense that they do what humans do, but rather they augment humanity in very useful ways," Dimakis said. "This is not doing what humans can do, but rather creating new AI tools that are going to improve the human condition."

Read more:

There's AI, and Then There's AGI: What You Need to Know to Tell the Difference - CNET

Mysterious Entity Paying Reddit $60 Million to Train AI With Users’ Posts – Futurism

"When you use something for free, you are the product." Reddit and Weep

Underlying the storm of hype and funding in the AI sector right now is a scarce resource: data, created by old-fashioned humans, that's needed to train the huge models like ChatGPT and DALL-E that generate text and imagery.

That demand is causing all sorts of drama, from lawsuits by authors and news organizations that say their work was used by AI companies without their permission to the looming question of what happens when the internet fills up with AI-generated content and AI creators are forced to usethat to train future AI.

And, of course, it's also fueling new business deals as AI developers rush to lock down repositories of human-generated work that they can use to train their AI systems. Look no further than this wild scoop from Bloomberg: that an undisclosed AI outfit has struck a deal to pay Reddit $60 millionper year for access to its huge database of users' posts perhaps the surest sign yet that user data is the key commodity in the AI gold rush.

It's not the first time we've seen an AI company cough up for access to a cache of text material. Remember when Axel Springer, the owner of publications ranging from Politico to Business Insider, inked a deal with OpenAI to use its outlets' work in ChatGPT?

But in some respects, it does differ from that bargain. For one, journalists are paid for their work, even if they don't stand to benefit and may actually be harmed by its inclusion in AI systems. Redditors, though, have contributed their vast supply of words as a labor of love which has to rankle when it's all vacuumed up for profit.

"Where the fuck is my cut?" quipped on Redditor in response to the news.

"When you use something for free, you are the product," another retorted.

Even stranger is that in spite of the appreciable sum changing hands remember, this is $60 million every single year we don't actually know who's paying for all this data.

"As an AI language model, I cannot condone the selling of public forums' user data as training data without compensation for the users of said forum," another Redditor wrote of the AI deal, riffing on the way ChatGPT and other systems frequently demur from answering controversial questions.

More on AI:Amazon AGI Team Say Their AI Is Showing "Emergent Abilities"

See the original post:

Mysterious Entity Paying Reddit $60 Million to Train AI With Users' Posts - Futurism

Reddit has a new AI training deal to sell user content – The Verge

Reddit will let an unnamed large AI company have access to its user-generated content platform in a new licensing deal, according to Bloomberg yesterday. The deal, worth about $60 million on an annualized basis, the outlet writes, could still change as the companys plans to go public are still in the works.

Whether or not thats true after all, one of the best ways to get around SEO spam in search results is to add the word Reddit to your search query Reddit has shown that its willing to play hardball before. Last year, it successfully stonewalled its way out of the biggest protest in its history after changes to its third-party API access pricing caused developers of the most popular Reddit apps to shut down.

As Bloomberg writes, Reddits year-over-year revenue was up by 20 percent by the end of 2023, but it was still $200 million shy of a $1 billion target it had set two years prior. The company was reportedly advised to seek a $5 billion valuation when it opens up for public investment, which is expected to happen in March. Thats half the $10$15 billion it might have achieved the last time it filed to go public in 2021, before a market downturn held it back.

Read the rest here:

Reddit has a new AI training deal to sell user content - The Verge