Archive for the ‘Alphago’ Category

Opinion: Alpha Phi Alpha develops leaders and promotes brotherhood – The San Diego Union-Tribune

Mitchell was initiated into Alpha Phi Alpha Fraternity Inc. by way of Beta Chapter on the campus of Howard University in 1995, and is president of the local alumni San Diego chapter, Zeta Sigma Lambda Chapter. He lives in La Jolla.

Alpha Phi Alpha Fraternity Inc. was founded on Dec. 4, 1906, on the campus of Cornell University by seven men. Henry Arthur Callis, Charles Henry Chapman, Eugene Kinckle Jones, George Biddle Kelley, Nathaniel Allison Murray, Robert Harold Ogle and Vertner Woodson Tandy dared to be pioneers in an uncharted field of student life.

We provide this platform for community commentary free of charge. Thank you to all the Union-Tribune subscribers whose support makes our journalism possible. If you are not a subscriber, please consider becoming one today.

Alpha Phi Alpha members at the 2020 MLK Parade. The green shirts were handed out to volunteers.

(Courtesy photo)

Our founders, known as the Jewels, went on to become a medical doctor, an educator, an executive secretary of the National Urban League, a civil engineer, an instructor, a secretary attached to the U.S. Senate Appropriations Committee and an architect. These men recognized the need for a strong bond of brotherhood among African Americans. Their success in establishing a fraternity created the framework for the creation of other African American Greek letter organizations.

Alpha Phi Alpha Fraternity Inc. develops leaders, promoting brotherhood and academic excellence while providing service and advocacy for our communities. Part of our legacy is our membership. One of our most famous members is Dr. Martin Luther King Jr., who was initiated in 1952 as a graduate student at Boston University working on his doctorate in systematic theology with an interest in philosophy and ethics.

As we celebrate his birthday each January, we are reminded that Dr. Kings journey was not alone his leadership abilities were developed by members of our fraternity, and brotherhood was promoted as he took the first of many steps on his journey to provide service and advocacy to our American community.

Locally, through collaboration with Zeta Sigma Lambda Foundation, we celebrate Martin Luther King Jr. in San Diego with an annual parade on Harbor Drive. Each parade has a theme, and is full of dazzling floats, bands, drill teams, colleges, fraternities, sororities, churches and community organizations.

During this parade, we also honor our military and our police and fire departments. The parade has evolved over time and now represents the diverse community here in San Diego. It is something that lends itself to how we celebrate him nationally with his own holiday and memorial.

The MLK memorial in Washington, D.C., was established by Alpha Phi Alpha Fraternity Inc. in 1996. Forty years earlier, Dr. King had been honored by Alpha Phi Alpha Fraternity Inc. with the Alpha Award of Honor for Christian leadership in the cause of first-class citizenship for all mankind.

It is in this spirit that we, the local chapter of Alpha Phi Alpha Fraternity Inc., decided to contribute to protecting our community by postponing the parade until 2023. We are thankful for the continued collaboration with San Diego County and look forward to returning in 2023 to the same route celebrating the diversity of our community.

Our fraternity works to provide service to our community through national and local programs. Our national programs are Project Alpha, Go-To-High-School, Go-To-College and A Voteless People Is A Hopeless People. Our local programs, along with the MLK parade, are the San Diego Multicultural Festival and our Holiday Scholarship Ball. The Multicultural Festival celebrates the diversity reflected throughout San Diego. This event is planned for April 24. This past December, we celebrated the Holiday Scholarship Ball, which assists with our scholarship fundraising.

Project Alpha was developed collaboratively with the March of Dimes to educate African Americans on the consequences of teenage pregnancy from the male perspective. This program assists young men in developing an understanding of their role in preventing untimely pregnancies and sexually transmitted infections through responsible attitudes and behaviors.

Our Go-To-High-School, Go-To-College programs focus on the importance of completing secondary and collegiate education as a road to advancement. We believe school completion is the single best predictor of future economic success. Locally, we partner with middle and high schools to mentor African American young men towards this goal. Through our fundraising efforts, we provide scholarships for college-bound students.

Our A Voteless People Is A Hopeless People program focuses on political awareness, empowerment and the importance of voting. Our local chapters program will consist of a no-touch voter registration drive on Saturday from 9 a.m. to 12 p.m. at 312 Euclid Avenue in San Diego. We invite members of the community to come out and safely confirm their voting registration in preparation for the coming elections.As we navigate our new normal, Alpha Phi Alpha Fraternity Inc. will reaffirm our commitment to social justice, community advocacy, economic development and mobility, education and health-care equity.

View original post here:
Opinion: Alpha Phi Alpha develops leaders and promotes brotherhood - The San Diego Union-Tribune

Altos bursts out of stealth with $3B, a dream team C-suite and a wildly ambitious plan to reverse disease – FierceBiotech

Altos Labs just redefined big in biotech. Where to start? The $3 billion in investor support? The C-suite staffed by storied leadersBarron, Bishop, Klausneridentifiable by one name? Or the wildly ambitious plan to reverse disease for patients of any age? Altos is all that and more.

Early details of Altos leaked out last year when MIT Technology Review reported Jeff Bezos had invested to support development of technology that could revitalize entire animal bodies, ultimately prolonging human life. The official reveal fleshes out the vision and grounds the technology in the context of the nearer-term opportunities it presents to improve human health.

It's clear from work by Shinya Yamanaka, and many others since his initial discoveries, that cells have the ability to rejuvenate, resetting their epigenetic clocks and erasing damage from a myriad of stressors. These insights, combined with major advances in a number of transformative technologies, inspired Altos to reimagine medical treatments where reversing disease for patients of any age is possible, Hal Barron, M.D., said in a statement.

Barron is set to take up the CEO post when he leaves GlaxoSmithKline in August, completing a C-suite staffed by some of the biggest names in life sciences. The former Genentech executive will join Rick Klausner, M.D., and Hans Bishop at the top of Altos. Klausner, co-founder of companies including Juno Therapeutics and Grail, is taking up the chief scientific officer post. Bishop, who used to run Juno and Grail, is Altos president. The leadership team is rounded out by Chief Operating Officer Ann Lee-Karlon, Ph.D., formerly of Genentech.

RELATED: Barron quits GSK to take CEO post at $3B biotech startup

The team will use $3 billion in capital committed by investors including Arch Venture Partners to try to turn breakthroughs in our understanding of cellular rejuvenation into transformational medicines. That effort will build on the work of a galaxy of academic scientists Altos has brought under its umbrella.

Aiming to integrate the best features of academia and industry, the startup is setting up Altos Institutes of Science in San Francisco, San Diego and Cambridge, U.K. Juan Carlos Izpisua Belmonte, Ph.D., Wolf Reik, M.D., and Peter Walter, Ph.D., will lead the three institutes, overseeing the work of a current roster of almost 20 principal investigators across the sites. The scientific leadership team also features Thore Graepel, Ph.D., co-inventor of AI breakthrough AlphaGo, and Shinya Yamanaka, M.D., Ph.D., a Nobel laureate who gives Altos ties to Japan.

Klausner, who founded Altos with Bishop, and his colleagues brought the scientists together and created a board of directors that features luminaries such as CRISPR pioneer Jennifer Doudna, Ph.D., and fellow Nobel laureates Frances Arnold, Ph.D., and David Baltimore, Ph.D., to help bring cellular rejuvenation out of academic labs and into clinical development.

Altos seeks to decipher the pathways of cellular rejuvenation programming to create a completely new approach to medicine, one based on the emerging concepts of cellular health, Klausner said. Remarkable work over the last few years beginning to quantify cellular health and the mechanisms behind that, coupled with the ability to effectively and safely reprogram cells and tissues via rejuvenation pathways, opens this new vista into the medicine of the future.

Follow this link:
Altos bursts out of stealth with $3B, a dream team C-suite and a wildly ambitious plan to reverse disease - FierceBiotech

DeepMind’s David Silver on games, beauty, and AI’s potential to avert human-made disasters – Bulletin of the Atomic Scientists

DeepMinds David Silver speaks to the Bulletin of the Atomic Scientists about games, beauty, and AIs potential to avert human-made disasters. Photo provided by David Silver and used with permission.

David Silver thinks games are the key to creativity. After competing in national Scrabble competitions as a kid, he went on to study at Cambridge and co-found a video game company. Later, after earning his PhD in artificial intelligence, he led the DeepMind team that developed AlphaGothe first program to beat a world champion at the ancient Chinese game of go. But he isnt driven by competitiveness.

Thats because for Silver, now a principal research scientist at DeepMind and computer science professor at University College London, games are playgrounds in which to understand how mindshuman and artificiallearn on their own to achieve goals.

Silvers programs use deep neural networksmachine learning algorithms inspired by the brains structure and functionto achieve results that resemble human intuition and creativity. First, he provided the program with information about what humans would do in various positions for it to imitate, a learning style known as supervised learning. Eventually, he let the program learn by playing itself, known as reinforcement learning.

Then, during a pivotal match between AlphaGo and the world champion, he had an epiphany: Perhaps the machine should have no human influence at all. That idea became AlphaGo Zero, the successor to AlphaGo that received zero human knowledge about how to play well. Instead, AlphaGo Zero relies only on the games rules and reinforcement learning. It beat AlphaGo 100 games to zero.

I first met Silver at the Heidelberg Laureate Foruman invitation-only gathering of the most exceptional mathematicians and computer scientists of their generations. In Heidelberg, he was recognized for having received the Association for Computing Machinerys prestigious Prize in Computing for breakthrough advances in computer game-playing.

Few other researchers have generated as much excitement in the AI field as David Silver, Association for Computing Machinery President Cherri M. Pancake said at the time. His insights into deep reinforcement learning are already being applied in areas such as improving the efficiency of the UKs power grid, reducing power consumption at Googles data centers, and planning the trajectories of space probes for the European Space Agency. Silver is also an elected Fellow of the Royal Society and was the first recipient of the Mensa Foundation Prize for the best scientific discovery in the field of artificial intelligence.

Silvers stardom contrasts with his quiet, unassuming nature. In this condensed, edited, from-the-heart interview, I talk with Silver about games, the meaning of creativity, and AIs potential to avert disasters such as climate change, human-made pathogens, mass poverty, and environmental catastrophe.

As a kid, did you play games differently from other kids?

I had some funny moments playing in National School Scrabble competitions. In one event, at the end of the final game, I asked my opponent, Are you sure you want to play that? Why not play this other word which scores more points? He changed his move and won the game and championship, which made me really happy.

More than winning, I am fascinated with what it means to play a game really well.

How did you translate that love of games into a real job?

Later on, I played junior chess, where I met [fellow DeepMind co-founder] Demis Hassabis. At that time, he was the strongest boy chess player of his age in the world. He would turn up in my local town when he needed pocket money, play in these tournaments, win the 50-pound prize money, and then go back home. Later, we got to know each other at Cambridge and together we set up Elixir, our games company. Now were back together at DeepMind.

What did this fascination with games teach you about problem solving?

Humans want to believe that weve got this special capacity called creativity that our algorithms dont or wont have. Its a fallacy.

Weve already seen the beginnings of creativity in our AIs. There was a moment in the second game of the [2016] AlphaGo match [against world champion Lee Sodol] where it played a particular move called move 37. The go community certainly felt that this was creative. It tried something new which didnt come from examples of what would normally be done there.

But is that the same kind of broad creativity that humans can apply to anything, rather than just moves within a game?

The whole process of trial-and-error learning, of trying to figure out for yourself, or asking AI to figure out for itself, how to solve the problem is a process of creativity. You or the AI start off not knowing anything. Then you or it discover one new thing, one creative leap, one new pattern or one new idea that helps in achieving the goal a little bit better than before. And now you have this new way of playing your game, solving your puzzle, or interacting with people. The process is a million mini discoveries, one after the other. It is the essence of creativity.

If our algorithms arent creative, theyll get stuck. They need an ability to try out new ideas for themselvesideas that were not providing. That has to be the direction of future research, to keep pushing on systems that can do that for themselves.

If we can crack [how self-learning systems achieve goals], its more powerful than writing a system that just plays go. Because then well have an ability to learn to solve a problem that can be applied to many situations.

Many thought that computers could only ever play go at the level of human amateurs. Did you ever doubt your ability to make progress?

When I arrived in South Korea [for the 2016 AlphaGo match] and saw row upon row of cameras set up to watch and heard how many people [over 200 million] were watching online, I thought, Hang on, is this really going to work? It was scary. The world champion is unbelievably versatile and creative in his ability to probe the program for weaknesses. He would try everything in an attempt to push the program into weird situations that dont normally occur.

I feel lucky that we stood up to that test. That spectacular and terrifying experience led me to reflect. I stepped back and asked, Can we go back to the basics to understand what it means for a system to truly learn for itself? To find something purer, we threw away the human knowledge that had gone into it and came up with AlphaZero.

Humans have developed well-known strategies for go over millennia. What did you think as AlphaZero quickly discovered, and rejected, these in favor of novel approaches?

We set up board positions where the original version of AlphaGo had made mistakes. We thought if we could find a new version that gets them right, wed make progress. At first, we made massive progress, but then it appeared to stop. We thought it wasnt getting 20 or 30 positions right.

Fan Hui, the professional player [and European champion] we were working with, spent hours studying the moves. Eventually, he said that the professional players were wrong in these positions and AlphaZero was right. It found solutions that made him reassess what was in the category of being a mistake. I realized that we had an ability to overturn what humans thought was standard knowledge.

After go, you moved on to a program that mastered StarCrafta real-time strategy video game. Why the jump to video games?

Go is one narrow domain. Extending from that to the human brains breadth of capabilities requires a huge number of steps. Were trying to add any dimensions of complexity where humans can do things, but our agents cant.

AlphaStar moves toward things which are more naturalistic. Like human vision, the system only gets to look at a certain part of the map. Its not like playing go or chess where you see all of your opponents pieces. You see nearby information and have to scout to acquire information. These aspects bring it closer to what happens in the real world.

Whats the end goal?

I think its AI agents that are as broadly capable as human brains. We dont know how to get there yet but we have a proof of existence in the human brain.

Replicating the human brain? Do you really think thats realistic?

I dont believe in magical, mystical explanations of the brain. At some level, the human brain is an algorithm which takes inputs and produces outputs in a powerful and general way. Were limited by our ability to understand and build AIs, but that understanding is growing fast. Today we have systems that are able to crack narrow domains like go. Weve also got language models which can understand and produce compelling language. Were building things one challenge at a time.

So, you think theres no ceiling to what AI can do?

Were just at the beginning. Imagine if you run evolution for another 4 billion years. Where would we end up? Maybe we would have much more sophisticated intelligences which could do a much better job. I see AI a little bit like that. There is no limit to this process because the world is essentially infinitely complex.

And so, is there a limit? At some point, you hit physical limits, so its not that there are no bounds. Eventually you use up all of the energy in the universe and all of the atoms in the universe in building your computational device. But relative to where we are now, thats essentially limitless intelligence. The spectrum beyond human intelligence is vast, and thats an exciting thought.

Stephen Hawking, who served on the Bulletins Board of Sponsors, worried about unintended consequences of machine intelligence. Do you share his concern?

I worry about the unintended consequences of human intelligence, such as climate change, human-made pathogens, mass poverty, and environmental catastrophe. The quest for AI should result in new technology, greater understanding, and smarter decision making. AI may one day become our greatest tool in averting such disasters. However, we should proceed cautiously and establish clear rules prohibiting unacceptable uses of AI, such as banning the development of autonomous weapons.

Youve had many successes meeting these grand challenges through games, but have there been any disappointments?

Well, supervised learningthis idea that you learn from exampleshas had an enormous mainstream impact. Most of the big applications that come out of Google use supervised learning somewhere in the system. Machine translation systems from English to French, for example, in which you want to know the right translation of a particular sentence, are trained by supervised learning. It is a very well understood problem and weve got clear machinery now that is effective at scaling up.

One of my disappointments at the moment is that we havent yet seen that level of impact with self-learning systems through reinforcement learning. In the future, Id love to see self-learning systems which are interacting with people, in virtual worlds, in ways that are really achieving our goals. For example, a digital assistant thats learning for itself the best way to accomplish your goals. That would be a beautiful accomplishment.

What kinds of goals?

Maybe we dont need to say. Maybe its more like we pat our AI on the back every time it does something we like, and it learns to maximize the number of pats on the back it gets and, in doing so, achieves all kinds of goals for us, enriching our lives and helping us doing things better. But we are far from this.

Do you have a personal goal for your work?

During the AlphaGo match with Lee Sedol, I went outside and found a go player in tears. I thought he was sad about how things were going, but he wasnt. In this domain in which he had invested so much, AlphaGo was playing moves he hadnt realized were possible. Those moves brought him a profound sense of beauty.

Im not enough of a go player to appreciate that at the level he could. However, we should strive to build intelligence where we all get a sense of that.

If you look aroundnot just in the human world but in the animal worldthere are amazing examples of intelligence. Im drawn to say, We built something thats adding to that spectrum of intelligence. We should do this not because of what it does or how it helps us, but because intelligence is a beautiful thing.

See the article here:
DeepMind's David Silver on games, beauty, and AI's potential to avert human-made disasters - Bulletin of the Atomic Scientists

U.S. vs. China Rivalry Boosts Techand Tensions – IEEE Spectrum

In June 2020, OpenAI, an independent artificial-intelligence research lab based in San Francisco, announced GPT-3, the third generation of its massive Generative Pre-trained Transformer language model, which can write everything from computer code to poetry.

A year later, with much less fanfare, Tsinghua Universitys Beijing Academy of Artificial Intelligence released an even larger model, Wu Dao 2.0, with 10 times as many parametersthe neural network values that encode information. While GPT-3 boasts 175 billion parameters, Wu Dao 2.0s creators claim it has a whopping 1.75 trillion. Moreover, the model is capable not only of generating text like GPT-3 does but also images from textual descriptions like OpenAIs 12-billion parameter DALL-E model, and has a similar scaling strategy to Googles 1.6 trillion-parameter Switch Transformer model.

Tang Jie, the Tsinghua University professor leading the Wu Dao project, said in a recent interview that the group built an even bigger, 100 trillion-parameter model in June, though it has not trained it to convergence, the point at which the model stops improving. We just wanted to prove that we have the ability to do that, Tang said.

This isnt simple one-upmanship. On the one hand, its how research progresses. But on the other, it is emblematic of an intensifying competition between the worlds two technology superpowers. Whether the researchers involved like it or not, their governments are eager to adopt each AI advance into their national security infrastructure and military capabilities.

That matters, because dominance in the technology means probable victory in any future war. Even more important, such an advantage likely guarantees the longevity and global influence of the government that wields it. Already, China is exporting its AI-enabled surveillance technologywhich can be used to quash dissentto client states and is espousing an authoritarian model that promises economic prosperity as a counter to democracy, something that the Soviet Union was never able to do.

Ironically, China is a competitor that the United States abetted. Its well known that the U.S. consumer market fed Chinas export engine, itself outfitted with U.S. machines, and led to the fastest-growing economy in the world since the 1980s. Whats less well-known is how a handful of technology companies transferred the know-how and trained the experts now giving the United States a run for its money in AI.

Blame Bill Gates, for one. In 1992, Gates led Microsoft into Chinas fledgling software market. Six years later, he established Microsoft Research Asia, the companys largest basic and applied computer-research institute outside the United States. People from that organization have gone on to found or lead many of Chinas top technology institutions.

China is a competitor that the United States abetted. A handful of U.S. tech companies transferred their know-how and trained some of China's top AI experts.

Ever hear of TikTok? In 2012, Zhang Yiming, a Microsoft Research Asia alum, founded the video-sharing platforms parent company, ByteDance, which today is one of the worlds most successful AI companies. He hired a former head of Microsoft Research Asia, Zhang Hongjiang, to lead ByteDances Technical Strategy Research Center. This Zhang is now head of the Beijing Academy the organization behind Wu Dao 2.0, currently the largest AI system on the planet. That back-and-forth worries U.S. national-security strategists, who plan for a day when researchers and companies are forced to take sides.

Todays competition has roots in an incident on 7 May 1999, when a U.S. B-2 Stealth Bomber dropped bombs on the Chinese embassy in Belgrade, Serbia, killing three people.

That's when the Chinese started saying, We're moving beyond attrition warfare to what they referred to as systems confrontation, the confrontation between their operational system and the American operational system, says Robert O. Work, former U.S. Deputy Secretary of Defense and vice chairman of the recently concluded National Security Commission on Artificial Intelligence. Their theory of victory is what they refer to as system destruction.

The Chinese and the Americans see this much the same way, says Work, calling it a hot competition. If one can blow apart their adversarys battle network, the adversary won't be able to operate and won't be able to achieve their objectives.

System-destruction warfare is part and parcel of what the Peoples Liberation Army thinks of as intelligentized warfare, in which war is waged not only in the traditional physical domains of land, sea, and air but also in outer space, nonphysical cyberspace, and electromagnetic and even psychological domainsall enabled and coordinated with AI.

Work says the first major U.S. AI effort toward intelligentized warfare was to use computer vision to analyze thousands of hours of full-motion video being downloaded from dozens of drones. Today, that effort, dubbed Project Maven, detects, classifies, and tracks objects within video images, and it has been extended to acoustic data and signals intelligence.

The Chinese have kept pace. According to Georgetown Universitys Center for Security and Emerging Technology, China is actively pursuing AI-based target recognition and automatic-weapon-firing research, which could be used in lethal autonomous weapons. Meanwhile, the country may be ahead of the United States in swarm technology, according to Work. Georgetowns CSET reports that China is developing electromagnetic weapon payloads that can be attached to swarms of small unmanned aerial vehicles and flown into enemy airspace to disrupt or block the enemy's command and decision-making.

I worry about their emphasis on swarms of unmanned systems, says Work, adding that the Chinese want to train swarms of a hundred vehicles or more, including underwater systems, to coordinate navigation through complex environments. While we also test swarms, we have yet to demonstrate the ability to employ these types of swarms in a combat scenario.

Chinese firm Baiduwhose comparatively modest Sunnyvale, Calif. office is pictured here in 2018is one of the largest Internet companies in the world. Smith Collection/Gado/Getty Images

This type of research and testing has prompted calls for preemptive bans on lethal autonomous weapons, but neither country is willing to declare an outright prohibition. Barring a prohibition, many people believe that China and the United States, along with other countries, should begin negotiating an arms-control agreement banning the development of systems that could autonomously order a preemptive or retaliatory attack. Such systems might inadvertently lead to flash wars, just as AI-driven autonomous trading has led to flash crashes in the financial markets.

Neither of us wants to get into a war because an autonomous-control system made a mistake and ordered a preemptive strike, Work says, referring to the United States and China.

All of this contributes to a dilemma facing the twin realms of AI research and military modernization. The international research community, collaborative and collegial, prefers to look the other way and insist that it only serves the interest of science. But the governments that fund that research have clear agendas, and military enhancement is undeniably one.

Geoffrey Hinton, regarded as one of the godfathers of deep learning, the kind of AI transforming militaries today, left the United States and moved to Canada largely because he didnt want to depend on funding from the Defense Advanced Research Projects Agency, or DARPA. The agency, the largest funder of AI research in the world, is responsible for the development of emerging technologies for military use.

Hinton instead helped to put deep learning on the map in 2012 with a now-famous neural net called AlexNet when he was at the University of Toronto. But Hinton was also in close contact with the Microsoft Research Lab in Redmond, Wash., before and after his group validated AlexNet, according to one of Hintons associates there, Li Deng, then principal researcher and manager and later chief scientist of AI at Microsoft.

In 2009 and 2010, Hinton and Deng worked together at Microsoft on speech recognition and Deng, then Editor-In-Chief of the IEEE Signal Processing Magazine, was invited in 2011 to lecture at several academic organizations in China where he said he shared the published success of deep learning in speech processing. Deng said he was in close contact with former Microsoft colleagues at Baidu, a Chinese search engine and AI giant, and a company called iFlyTek, a spin off from Dengs undergraduate alma mater.

When Hinton achieved his breakthrough with backpropagation in neural networks in 2012, he sent an email to Deng in Washington, and Deng said he shared it with Microsoft executives, including Qi Lu who led the development of the companys search engine, Bing. Deng said he also sent a note to his friends at iFlyTek, which quickly adopted the strategy and became an AI powerhousefamously demonstrated in 2017 with a convincing video of then-president Donald Trump speaking Chinese.

Qi Lu went on to become COO of Baidu where Deng said another Microsoft alum, Kai Yu, who also knew Hinton well, had already seized on Hintons breakthrough.

Chinas theory of victory is what they refer to as system destruction.

Robert O. Work, former U.S. Deputy Secretary of Defense

Literally within hours of Hintons results, according to Deng, researchers in China were working on repeating his success.

Had they not learned of Hintons work through the research grapevine, they still would have read about it in published papers and heard about it through international conferences. Research today has no borders. It is internationally fungible.

But the United States has since tried to limit this crosspollination, barring Chinese nationals known to have worked for Chinas military or intelligence organizations from working with U.S. research institutions. Yet research continues to flow back and forth between the two countries: Microsoft maintains its research lab in Beijing, and the Chinese Internet and AI giant Baidu has a research lab in Silicon Valley, for example.

Tsinghua Universitys Tang said decoupling the two countries would slow Chinas AI researchnot because it would stop the flow of ideas, but because it would cut China off from the advanced semiconductors needed to train AI models. He said his group is working on chip designs to speed AI training. China, meanwhile, is working to build extreme ultraviolet lithography machines and upgrade its semiconductor foundries to free itself from Western control.

While the U.S. government must negotiate with private sector organizations and researchers to participate in its military modernization, Chinas National Intelligence Law compels its companies and researchers to cooperate when asked.

China began pouring billions of dollars into AI research in 2017, following Google subsidiary DeepMinds success at defeating the world Go champion with its AI model AlphaGo. Among the organizations set up with that funding was Tsinghuas Beijing Academy, where Tang and his team built Wu Dao 2.0.

We hope that we can do science for the world, not just the one country, Tang says. But, he added, we should do something on demand based on the national project research plan.

By most metrics, Wu Dao 2.0 has surpassed OpenAIs GPT-3. Tang says it was trained on 4.9 terabytes of clean data, including Chinese-language text, English-language text, and images. OpenAI has said that GPT-3 was trained on just 570 gigabytes of clean, primarily English-language text.

Tang says his group is now working on video with the goal of generating realistic video from text descriptions. Hopefully, we can make this model do something beyond the Turing test, he says, referring to an assessment of whether a computer can generate text indistinguishable from that created by a human. That's our final goal.

Continue reading here:
U.S. vs. China Rivalry Boosts Techand Tensions - IEEE Spectrum

Monte Carlo Tree Search Tutorial | DeepMind AlphaGo

Introduction

A best of five game series, $1 million dollars in prize money A high stakes shootout. Between 9 and 15 March, 2016, the second-highest ranked Go player, Lee Sidol, took on a computer program named AlphaGo.

AlphaGo emphatically outplayed and outclassed Mr. Sidol and won the series 4-1. Designed by Googles DeepMind, the program has spawned many other developments in AI, including AlphaGo Zero. These breakthroughs are widely considered as stepping stones towards Artificial General Intelligence (AGI).

In this article, I will introduce you to the algorithm at the heart of AlphaGo Monte Carlo Tree Search (MCTS). This algorithm has one main purpose given the state of a game, choose the most promising move.

To give you some context behind AlphaGo, well first briefly look at the history of game playing AI programs. Then, well see the components of AlphaGo, the Game Tree Concept, a few tree search algorithm, and finally dive into how the MCTS algorithm works.

AI is a vast and complex field. But before AI officially became a recognized body of work, early pioneers in computer science wrote game-playing programs to test whether computers could solve human-intelligence level tasks.

To give you a sense of where Game Playing AI started from and its journey till date, I have put together the below key historical developments:

And this is just skimming the surface! There are plenty of other examples where AI programs exceeded expectations. But this should give you a fair idea of where we stand today.

The core parts of the Alpha Go comprise of:

In this blog, we willfocus on the working of Monte Carlo Tree Searchonly. This helps AlphaGo and AlphaGo Zero smartly explore and reach interesting/good states in a finite time period which in turn helps the AI reach human level performance.

Its application extends beyond games. MCTS can theoretically be applied to any domain that can be described in terms of {state,action} pairs and simulation used to forecast outcomes. Dont worry if this sounds too complex right now, well break down all these concepts in this article.

Game Trees are the most well known data structures that can represent a game. This concept is actually pretty straightforward.

Each node of a game tree represents a particular state in a game. On performing a move, one makes a transition from a node to its children. The nomenclature is very similar to decision trees wherein the terminal nodes are called leaf nodes.

For example, in the above tree, each move is equivalent to putting a cross at different positions. This branches into various other states where a zero is put at each position to generate new states. This process goes on until the leaf node is reached where the win-loss result becomes clear.

Our primary objective behind designing these algorithms is to find best the path to follow in order to win the game. In other words, look/search for a way of traversing the tree that finds the best nodes to achieve victory.

The majority of AI problems can be cast as search problems, which can be solved by finding the best plan, path, model or function.

Tree search algorithms can be seen as building a search tree:

The tree branches out because there are typically several different actions that can be taken in a given state. Tree search algorithms differ depending on which branches are explored and in what order.

Lets discuss a few tree search algorithms.

Uninformed Search algorithms, as the name suggests, search a state space without any further information about the goal. These are considered basic computer science algorithms rather than as a part of AI. Two basic algorithms that fall under this type of search are depth first search (DFS) and breadth first search (BFS). You can read more about them in this blog post.

The Best First Search (BFS) method explores a graph by expanding the most promising node chosen according to a specific rule.The defining characteristic of this search is that, unlikeDFSorBFS(which blindly examine/expand a cell without knowing anything about it), BFS uses an evaluation function (sometimes called a heuristic) to determine which node is the most promising, and then examines this node.

For example, A* algorithm keeps a list of open nodes which are next to an explored node. Note that these open nodes have not been explored. For each open node, an estimate of its distance from the goal is made. New nodes are chosen to explore based on the lowest cost basis, where the cost is the distance from the origin node plus the estimate of the distance to the goal.

For single-player games, simple uninformed or informed search algorithms can be used to find a path to the optimal game state. What should we do for two-player adversarialgames where there is another player to account for? The actions of both players depend on each other.

For these games, we rely on adversarial search. This includes the actions of two (or more) adversarial players. The basic adversarial search algorithm is called Minimax.

This algorithm has been used very successfully for playing classic perfect-information two-player board games such as Checkers and Chess. In fact, it was (re)invented specifically for the purpose of building a chess-playing program.

The core loop of the Minimax algorithm alternates between player 1 and player 2, quite like the white and black players in chess. These are called the min player and the max player. All possible moves are explored for each player.

For each resulting state, all possible moves by the other player are also explored. This goes on until all possible move combinations have been tried out to the point where the game ends (with a win, loss or draw). The entire game tree is generated through this process, from the root node down to the leaves:

Each node is explored to find the moves that give us the maximum value or score.

Games like tic-tac-toe, checkers and chess can arguably be solved using the minimax algorithm. However, things can get a little tricky when there are a large number of potential actions to be taken at each state. This is because minimax explores all the nodes available. It can become frighteningly difficult to solve a complex game like Go in a finite amount of time.

Go has a branching factor of approximately 300 i.e. from each state there are around 300 actions possible, whereas chess typically has around 30 actions to choose from. Further, the positional nature of Go, which is all about surrounding the adversary, makes it very hard to correctly estimate the value of a given board state. For more information on rules for Go, please refer this link.

There are several other games with complex rules that minimax is ill-equipped to solve. These include Battleship Poker with imperfect information and non-deterministic games such as Backgammon and Monopoly. Monte Carlo Tree Search, invented in 2007, provides a possible solution.

The basic MCTS algorithm is simple: a search tree is built, node-by-node, according to the outcomes of simulated playouts. The process can be broken down into the following steps:

Before we delve deeper and understand tree traversal and node expansion, lets get familiar with a few terms.

UCB Value

UCB1, or upper confidence bound for a node, is given by the following formula:

where,

What do we mean by a rollout? Until we reach the leaf node, we randomly choose an action at each step and simulate this action to receive an average reward when the game is over.

Flowchart for Monte Carlo Tree Search

Tree Traversal & Node Expansion

You start with S0, which is the initial state. If the current node is not a leaf node, we calculate the values for UCB1 and choose the node that maximises the UCB value. We keep doing this until we reach the leaf node.

Next, we ask how many times this leaf node was sampled. If its never been sampled before, we simply do a rollout (instead of expanding). However, if it has been sampled before, then we add a new node (state) to the tree for each available action (which we are calling expansion here).

Your current node is now this newly created node. We then do a rollout from this step.

Lets do a complete walkthrough of the algorithm to truly ingrain this concept and understand it in a lucid manner.

Iteration 1:

Initial State

Rollout from S1

Post Backpropogation

The way MCTS works is that we run it for a defined number of iterations or until we are out of time. This will tell us what is the best action at each step that one should take to get the maximum return.

Iteration 2:

Backpropogation from S2

Iteration 3:

Iteration 4:

That is the gist of this algorithm. We can perform more iterations as long as required (or is computationally possible). The underlying idea is thatthe estimate of values at each node becomes more accurate as the number of iterations keep increasing.

Deepminds AlphaGo and AlphaGo Zero programs are far more complex with various other facets that are outside the scope of this article. However, the Monte Carlo Tree Search algorithm remains at the heart of it. MCTS plays the primary role in making complex games like Go easier to crack in a finite amount of time. Some open source implementations of MCTS are linked below:

Implementation in Python

Implementation in C++

I expect reinforcement learning to make a lot of headway in 2019. It wont be surprising to see a lot more complex games being cracked by machines soon. This is a great time to learn reinforcement learning!

I would love to hear your thoughts and suggestions regarding this article and this algorithm in the comments section below. Have you used this algorithm before? If not, which game would you want to try it out on?

Related

See the article here:
Monte Carlo Tree Search Tutorial | DeepMind AlphaGo