Archive for the ‘Machine Learning’ Category

U of Texas will stop using controversial algorithm to evaluate Ph.D. applicants – Inside Higher Ed

In 2013, the University of Texas at Austins computer science department began usinga machine-learning system called GRADE to help make decisions about who gets into its Ph.D. program -- and who doesnt. This year, the department abandoned it.

Before the announcement, which the department released in the form of a tweet reply, few had even heard of the program. Now, its critics -- concerned about diversity, equity and fairness in admissions -- say it should never have been used in the first place.

Humans code these systems. Humans are encoding their own biases into these algorithms, said Yasmeen Musthafa, a Ph.D. student in plasma physics at the University of California, Irvine, who rang alarm bells about the system on Twitter. What would UT Austin CS department have looked like without GRADE? Well never know.

GRADE (which stands for GRaduate ADmissions Evaluator) was created by a UT faculty member and UT graduate student in computer science, originally to help the graduate admissions committee in the department save time. GRADE predicts how likely the admissions committee is to approve an applicant and expresses that prediction as a numerical score out of five. The system also explains what factors most impacted its decision.

The UT researchers who made GRADE trained it on a database of past admissions decisions. The system uses patterns from those decisions to calculate its scores for candidates.

For example, letters of recommendation containing the words best, award, research or Ph.D. are predictive of admission -- and can lead to a higher score -- while letters containing the words good, class, programming or technology are predictive of rejection. A higher grade point average means an applicant is more likely to be accepted, as does the name of an elite college or university on the rsum. Within the system, institutions were encoded into the categories elite, good and other, based on a survey of UT computer science faculty.

Every application GRADE scored during the seven years it was in use was still reviewed by at least one human committee member, UT Austin has said, but sometimes only one. Before GRADE, faculty members made multiple review passes over the pool. The system saved the committee time, according to its developers, by allowing faculty to focus on applicants on the cusp of admission or rejection and review applicants in descending order of quality.

For what its worth, GRADE did appear to successfully save the committee time. In the 2012 and 2013 application seasons, developers said in a paper about their work, it reduced the number of full reviews per candidate by 71percent and cut the total time reviewing files by 74percent. (One full review typically takes 10 to 30 minutes.) Between the years 2000 and 2012, applications to the computer science Ph.D. program grew from about 250 to nearly 650, though the number of faculty able to review those applications remained mostly constant. In the years since 2012, the number of applications has reached over 1,200.

The universitys use of the technology escaped attention for a number of years, until this month, when the physics department at the University of Maryland at College Park held a colloquium talk with the two creators of GRADE.

The talk gained attention on Twitter as graduate students accused GRADEs creators of further disadvantaging underrepresented groups in the computer science admissions process.

We put letters of recommendation in to try to lift people up who have maybe not great GPAs. We put a personal statement in the graduate application process to try to give marginalized folks a chance to have their voice heard, said Musthafa, who is also a member of the Physics and Astronomy Anti-Racism Coalition. The worst part about GRADE is that it throws that out completely.

Advocates have long been concerned about the potential for human biases to be baked into or exacerbated by machine-learning algorithms. Algorithms are trained on data. When it comes to people, what those data look like is a result of historical inequity. Preferences for one type of person over another are often the result of conscious or unconscious bias.

That hasnt stopped institutions from using machine-learning systems in hiring, policing and prison sentencing for a number of years now, often to great controversy.

Every process is going to make some mistakes. The question is, where are those mistakes likely to be made and who is likely to suffer as a result of them? said Manish Raghavan, a computer science Ph.D. candidate at Cornell University who has researched and written about bias in algorithms. Likely those from underrepresented groups or people who dont have the resources to be attending elite institutions.

Though many women and people who are Black and Latinx have had successful careers in computer science, those groups are underrepresented in the field at large. In 2017, whites, Asians and nonresident aliens received 84percent of degrees awarded for computer science in the United States.

At UT, nearly 80percent of undergraduates in computer science in 2017 were men.

Raghavan said he was surprised that there appeared to be no effort to audit the impacts of GRADE, such as how scores differ across demographic groups.

GRADEs creators have said that the system is only programmed to replicate what the admissions committee was doing prior to 2013, not to make better decisions than humans could. The system isnt programmed to use race or gender to make its predictions, theyve said. In fact, when given those features as options to help make its predictions, it chooses to give them zero weight. GRADEs creators have said this is evidence that the committees decisions are gender and race neutral.

Detractors have countered this, arguing that race and gender can be encoded into other features of the application that the system uses. Womens colleges and historically Black universities may be undervalued by the algorithm, theyve said. Letters of recommendation are known to reflect gender bias, as recommenders are more likely to describe female students as caring rather than assertive or trailblazing.

In the Maryland talk, faculty raised their own concerns. What a committee is looking for might change each year. Letters of recommendation and personal statements should be thoughtfully considered, not turned into a bag of words, they said.

Im kind of shocked you did this experiment on your students, Steve Rolston, chair of the physics department at Maryland, said during the talk. You seem to have built a model that builds in whatever bias your committee had in 2013 and youve been using it ever since.

In an interview, Rolston said graduate admissions can certainly be a challenge. His department receives over 800 graduate applications per year, which takes a good deal of time for faculty to evaluate. But, he said, his department would never use a tool like this.

If I ask you to do a classifier of images and youre looking for dogs, I can check afterwards that, yes, it did correctly identify dogs, he said. But when Im asking for decisions about people, whether it's graduate admissions, or hiring or prison sentencing, theres no obvious correct answer. You train it, but you dont know what the result is really telling you.

Rolston said having at least one faculty member review each application was not a convincing safeguard.

If I give you a file and say, Well, the algorithm said this person shouldnt be accepted, that will inevitably bias the way you look at it, he said.

UT Austin has said GRADE was used to organize admissions decisions, rather than make them.

"It was never used to make decisions to admit or reject prospective students, asat least one faculty member directly evaluates applicants at each stage of the review process," a spokesperson for the Graduate School said via email.

Despite the criticism around diversity and equity, UT Austin has said GRADE is being phased out because it is too difficult to maintain.

Changes in the data and software environment made the system increasingly difficult to maintain, and its use was discontinued, the spokesperson said via email. The Graduate School works with graduate programs and faculty members across campus to promote holistic application review and reduce bias in admissions decisions.

For Musthafa, the fact that GRADE may be gone for good does not impact the existing inequity in graduate admissions.

The entire system is steeped in racism, sexism and ableism, they said. How many years of POC computer science students got denied [because of this]?

Addressing that inequity -- as well as the competitiveness that led to the creation of GRADE -- may mean expanding committees, paying people for their time and giving Black and Latinx graduate students a voice in those decisions, they said. But automating cannot be part of that decision making.

If we automate this to any extent, its just going to lock people out of academia, Musthafa said. The racism of today is being immortalized in the algorithms of tomorrow.

Continued here:
U of Texas will stop using controversial algorithm to evaluate Ph.D. applicants - Inside Higher Ed

Information gathering: A WebEx talk on machine learning – Santa Fe New Mexican

Were long past the point of questioning whether machines can learn. The question now is how do they learn? Machine learning, a subset of artificial intelligence, is the study of computer algorithms that improve automatically through experience. That means a machine can learn, independent of human programming. Los Alamos National Laboratory staff scientist Nga Thi Thuy Nguyen-Fotiadis is an expert on machine learning, and at 5:30 p.m. on Monday, Dec. 14, she hosts the virtual presentation Deep focus: Techniques for image recognition in machine learning, as part of the Bradbury Science Museums (1350 Central Ave., Los Alamos, 505-667-4444, lanl.gov/museum) Science on Tap lecture series. Nguyen-Fotiadis is a member of LANLs Information Sciences Group, whose Computer, Computational, and Statistical Sciences division studies fields that are central to scientific discovery and innovation. Learn about the differences between LANLs Trinity supercomputer and the human brain, and how algorithms determine recommendations for your nightly viewing pleasure on Netflix and the like. The talk is a free WebEx virtual event. Follow the link from the Bradburys event page at lanl.gov/museum/events/calendar/2020/12 /calendar-sot-nguyen-fotaidis.php to register.

Read this article:
Information gathering: A WebEx talk on machine learning - Santa Fe New Mexican

LeanTaaS Raises $130 Million to Strengthen Its Machine Learning Software Platform to Continue Helping Hospitals Achieve Operational Excellence -…

SANTA CLARA, Calif.--(BUSINESS WIRE)--LeanTaaS, Inc., a Silicon Valley software innovator that increases patient access and transforms operational performance for healthcare providers, announced a $130 million Series D funding round led by Insight Partners with participation from Goldman Sachs. The funds will be used to invest in building out the existing suite of products (iQueue for Operating Rooms, iQueue for Infusion Centers and iQueue for Inpatient Beds) as well as scaling the engineering, product and GoToMarket teams, and expanding the iQueue platform to include new products.

LeanTaaS is uniquely positioned to help hospitals and health systems across the country face the mounting operational and financial pressures exacerbated by the coronavirus. This funding will allow us to continue to grow and expand our impact while helping healthcare organizations deliver better care at a lower cost, said Mohan Giridharadas, founder and CEO of LeanTaaS. Our company momentum over the past several years - including greater than 50% revenue growth in 2020 and negative churn despite a difficult macro environment - reflects the increasing demand for scalable predictive analytics solutions that optimize how health systems increase operational utilization and efficiency. It also highlights how weve been able to develop and maintain deep partnerships with 100+ health systems and 300+ hospitals in order to keep them resilient and agile in the face of uncertain demand and supply conditions.

With this investment, LeanTaaS has raised more than $250 million in aggregate, including more than $150 million from Insight Partners. As part of the transaction, Insight Partners Jeff Horing and Jon Rosenbaum and Goldman Sachs Antoine Munfa will join LeanTaaS Board of Directors.

Healthcare operations in the U.S. are increasingly complex and under immense pressure to innovate; this has only been exacerbated by the prioritization of unique demands from the current pandemic, said Jeff Horing, co-founder and Managing Director at Insight Partners. Even under these unprecedented circumstances, LeanTaaS has demonstrated the effectiveness of its ML-driven platform in optimizing how hospitals and health systems manage expensive, scarce resources like infusion center chairs, operating rooms, and inpatient beds. After leading the companys Series B and C rounds, we have formed a deep partnership with Mohan and team. We look forward to continuing to help LeanTaaS scale its market presence and customer impact.

Although health systems across the country have invested in cutting-edge medical equipment and infrastructure, they cannot maximize the use of such assets and increase operational efficiencies to improve their bottom lines with human based scheduling or unsophisticated tools. LeanTaaS develops specialized software that increases patient access to medical care by optimizing how health systems schedule and allocate the use of expensive, constrained resources. By using LeanTaaS product solutions, healthcare systems can harness the power of sophisticated, AI/ML-driven software to improve operational efficiencies, increase access, and reduce costs.

We continue to be impressed by the LeanTaaS team. As hospitals and health systems begin to look toward a post-COVID-19 world, the agility and resilience LeanTaaS solutions provide will be key to restoring and growing their operations, said Antoine Munfa, Managing Director of Goldman Sachs Growth.

LeanTaaS solutions have now been deployed in more than 300 hospitals across the U.S., including five of the 10 largest health networks and 12 of the top 20 hospitals in the U.S. according to U.S. News & World Report. These hospitals use the iQueue platform to optimize capacity utilization in infusion centers, operating rooms, and inpatient beds. iQueue for Infusion Centers is used by 7,500+ chairs across 300+ infusion centers including 70 percent of the National Comprehensive Cancer Network and more than 50 percent of National Cancer Institute hospitals. iQueue for Operating Rooms is used by more than 1,750 ORs across 34 health systems to perform more surgical cases during business hours, increase competitiveness in the marketplace, and improve the patient experience.

I am excited about LeanTaaS' continued growth and market validation. As healthcare moves into the digital age, iQueue overcomes the inherent deficiencies in capacity planning and optimization found in EHRs. We are very excited to partner with LeanTaaS and implement iQueue for Operating Rooms, said Dr. Rob Ferguson, System Medical Director, Surgical Operations, Intermountain Healthcare.

Concurrent with the funding, LeanTaaS announced that Niloy Sanyal, the former CMO at Omnicell and GE Digital, would be joining as its new Chief Marketing Officer. Also, Sanjeev Agrawal has been designated as LeanTaaS Chief Operating Officer in addition to his current role as the President. "We are excited to welcome Niloy to LeanTaaS. His breadth and depth of experience will help us accelerate our growth as the industry evolves to a more data driven way of making decisions," said Agrawal.

About LeanTaaSLeanTaaS provides software solutions that combine lean principles, predictive analytics, and machine learning to transform hospital and infusion center operations. The companys software is being used by over 100 health systems across the nation which all rely on the iQueue cloud-based solutions to increase patient access, decrease wait times, reduce healthcare delivery costs, and improve revenue. LeanTaaS is based in Santa Clara, California, and Charlotte, North Carolina. For more information about LeanTaaS, please visit https://leantaas.com/, and connect on Twitter, Facebook and LinkedIn.

About Insight PartnersInsight Partners is a leading global venture capital and private equity firm investing in high-growth technology and software ScaleUp companies that are driving transformative change in their industries. Founded in 1995, Insight Partners has invested in more than 400 companies worldwide and has raised through a series of funds more than $30 billion in capital commitments. Insights mission is to find, fund, and work successfully with visionary executives, providing them with practical, hands-on software expertise to foster long-term success. Across its people and its portfolio, Insight encourages a culture around a belief that ScaleUp companies and growth create opportunity for all. For more information on Insight and all its investments, visit insightpartners.com or follow us on Twitter @insightpartners.

About Goldman Sachs GrowthFounded in 1869, The Goldman Sachs Group, Inc. is a leading global investment banking, securities and investment management firm. Goldman Sachs Merchant Banking Division (MBD) is the primary center for the firms long-term principal investing activity. As part of MBD, Goldman Sachs Growth is the dedicated growth equity team within Goldman Sachs, with over 25 years of investing history, over $8 billion of assets under management, and 9 offices globally.

LeanTaaS and iQueue are trademarks of LeanTaaS. All other brand names and product names are trademarks or registered trademarks of their respective companies.

Read more:
LeanTaaS Raises $130 Million to Strengthen Its Machine Learning Software Platform to Continue Helping Hospitals Achieve Operational Excellence -...

Runway Raises $8.5M Series A to Build the Next Generation of Creative Tools – Business Wire

NEW YORK--(BUSINESS WIRE)--Runway, a start-up building the next generation of digital creative tools, announced today an $8.5 million Series A led by Amplify Partners with participation from Lux Capital and Compound Ventures. Using machine learning, the company is pioneering image and video creation techniques for synthetic content manipulation and media editing, allowing users to create and generate content with cutting-edge AI and graphics technology. With an active and growing community, Runway has emerged as a leader in the future of content production.

Deep learning techniques are bringing a new paradigm to content creation with synthetic media and automation, Runway founder Cristobal Valenzuela explained. With Runway, were building the backbone of that creative revolution, allowing creators to do things that were impossible until very recently.

Today, Runways user community includes designers, filmmakers, and other creative professionals at R/GA, New Balance, Google, and IBM. A favorite among educators, Runway has been incorporated into the design curriculum at NYU, RISD, and MIT. So far, Runway users have trained more than 50,000 AI models, uploaded over 24 million files to the platform, and run more than 900,000 models.

By making sophisticated machine learning algorithms unimaginably accessible, Runway challenges a designers own visual and muscle memories, pushing them out of their comfort zones to create unexpected, innovative work, said Onur Yuce Gun, Creative Manager of Computation Design at New Balance.

Most recently, Runway released Green Screen, a web tool that uses machine learning to automate the process of rotoscoping, saving users significant time when they want to remove objects from a background. Professional editors and content creators are using Green Screen to edit content faster and expand the visual effects tools of their projects.

The ways in which we distribute content have changed radically in recent years; however, the tools that creative professionals use to make content have not, said Sarah Catanzaro, Partner at Amplify Partners. For the first time in decades, creatives now have a radically better suite of tools to generate and edit images, video, and other media with AI. Runway is not only automating routine work for creatives, but also enabling new forms of perception and creative expression. We are thrilled to partner with a talented team as they develop the next-gen creative toolkit.

The investment brings Runways total funding since launching to $10.5M and will help Runway to hire within their research and engineering teams as they continue building cutting-edge synthetic media tools while growing their community of creative users.

Valenzuela co-founded Runway with Anastasis Germanidis and Alejandro Matamala, all graduates of NYUs Interactive Telecommunications Program.

About Runway: Runway is building the next generation of creative tools that makes machine learning easy and accessible for all types of creatives. With an active and growing community, Runway is pioneering how content and media are created. With a focus on video automation and synthetic media, Runway reduces the costs of creating visual media across creative industries. To learn more and sign up for a free account visit http://www.runwayml.com.

Read more here:
Runway Raises $8.5M Series A to Build the Next Generation of Creative Tools - Business Wire

What is Machine Learning? | IBM

Machine learning focuses on applications that learn from experience and improve their decision-making or predictive accuracy over time.

Machine learning is a branch of artificial intelligence (AI) focused on building applications that learn from data and improve their accuracy over time without being programmed to do so.

In data science, an algorithm is a sequence of statistical processing steps. In machine learning, algorithms are 'trained' to find patterns and features in massive amounts of data in order to make decisions and predictions based on new data. The better the algorithm, the more accurate the decisions and predictions will become as it processes more data.

Today, examples of machine learning are all around us. Digital assistants search the web and play music in response to our voice commands. Websites recommend products and movies and songs based on what we bought, watched, or listened to before. Robots vacuum our floors while we do . . . something better with our time. Spam detectors stop unwanted emails from reaching our inboxes. Medical image analysis systems help doctors spot tumors they might have missed. And the first self-driving cars are hitting the road.

We can expect more. As big data keeps getting bigger, as computing becomes more powerful and affordable, and as data scientists keep developing more capable algorithms, machine learning will drive greater and greater efficiency in our personal and work lives.

There are four basic steps for building a machine learning application (or model). These are typically performed by data scientists working closely with the business professionals for whom the model is being developed.

Training data is a data set representative of the data the machine learning model will ingest to solve the problem its designed to solve. In some cases, the training data is labeled datatagged to call out features and classifications the model will need to identify. Other data is unlabeled, and the model will need to extract those features and assign classifications on its own.

In either case, the training data needs to be properly preparedrandomized, de-duped, and checked for imbalances or biases that could impact the training. It should also be divided into two subsets: the training subset, which will be used to train the application, and the evaluation subset, used to test and refine it.

Again, an algorithm is a set of statistical processing steps. The type of algorithm depends on the type (labeled or unlabeled) and amount of data in the training data set and on the type of problem to be solved.

Common types of machine learning algorithms for use with labeled data include the following:

Algorithms for use with unlabeled data include the following:

Training the algorithm is an iterative processit involves running variables through the algorithm, comparing the output with the results it should have produced, adjusting weights and biases within the algorithm that might yield a more accurate result, and running the variables again until the algorithm returns the correct result most of the time. The resulting trained, accurate algorithm is the machine learning modelan important distinction to note, because 'algorithm' and 'model' are incorrectly used interchangeably, even by machine learning mavens.

The final step is to use the model with new data and, in the best case, for it to improve in accuracy and effectiveness over time. Where the new data comes from will depend on the problem being solved. For example, a machine learning model designed to identify spam will ingest email messages, whereas a machine learning model that drives a robot vacuum cleaner will ingest data resulting from real-world interaction with moved furniture or new objects in the room.

Machine learningmethods (also called machine learning styles) fall into three primary categories.

Supervised machine learning trains itself on a labeled dataset. That is, the data is labeled with information that the machine learning model is being built to determine and that may even be classified in ways the model is supposed to classify data. For example, a computer vision model designed to identify purebred German Shepherd dogs might be trained on a data set of various labeled dog images.

Supervised machine learning requires less training data than other machine learningmethods and makes training easier because the results of the model can be compared to actual labeled results. But, properly labeled data is expensive to prepare, and there's the danger of overfitting, or creating a model so closely tied and biased to the training data that it doesn't handle variations in new data accurately.

Learn more about supervised learning.

Unsupervised machine learning ingests unlabeled datalots and lots of itand uses algorithms to extract meaningful features needed to label, sort, and classify the data in real-time, without human intervention. Unsupervised learning is less about automating decisions and predictions, and more about identifying patterns and relationships in data that humans would miss. Take spam detection, for examplepeople generate more email than a team of data scientists could ever hope to label or classify in their lifetimes. An unsupervised learning algorithm can analyze huge volumes of emails and uncover the features and patterns that indicate spam (and keep getting better at flagging spam over time).

Learn more about unsupervised learning.

Semi-supervised learning offers a happy medium between supervised and unsupervised learning. During training, it uses a smaller labeled dataset to guide classification and feature extraction from a larger, unlabeled data set. Semi-supervised learning can solve the problem of having not enough labeled data (or not being able to afford to label enough data) to train a supervised learning algorithm.

Reinforcement machine learning is a behavioral machinelearning model that is similar to supervised learning, but the algorithm isnt trained using sample data. This model learns as it goes by using trial and error. A sequence of successful outcomes will be reinforced to develop the best recommendation or policy for a given problem.

The IBM Watson system that won the Jeopardy! challenge in 2011 makes a good example. The system used reinforcement learning to decide whether to attempt an answer (or question, as it were), which square to select on the board, and how much to wagerespecially on daily doubles.

Learn more about reinforcement learning.

Deep learning is a subset of machine learning (all deep learning is machine learning, but not all machine learning is deep learning). Deep learning algorithms define an artificial neural network that is designed to learn the way the human brain learns. Deep learning models require large amounts of data that pass through multiple layers of calculations, applying weights and biases in each successive layer to continually adjust and improve the outcomes.

Deep learning models are typically unsupervised or semi-supervised. Reinforcement learning models can also be deep learning models. Certain types of deep learning modelsincluding convolutional neural networks (CNNs) and recurrent neural networks (RNNs)are driving progress in areas such as computer vision, natural language processing (including speech recognition), and self-driving cars.

See the blog post AI vs. Machine Learning vs. Deep Learning vs. Neural Networks: Whats the Difference? for a closer look at how the different concepts relate.

Learn more about deep learning.

As noted at the outset, machine learning is everywhere. Here are just a few examples of machine learning you might encounter every day:

IBM Watson Machine Learning supports the machine learning lifecycle end to end. It is available in a range of offerings that let you build machine learning models wherever your data lives and deploy them anywhere in your hybrid multicloud environment.

IBM Watson Machine Learning on IBM Cloud Pak for Data helps enterprise data science and AI teams speed AI development and deployment anywhere, on a cloud native data and AI platform. IBM Watson Machine Learning Cloud, a managed service in the IBM Cloud environment, is the fastest way to move models from experimentation on the desktop to deployment for production workloads. For smaller teams looking to scale machine learning deployments, IBM Watson Machine Learning Server offers simple installation on any private or public cloud.

To get started, sign up for an IBMid and create your IBM Cloud account.

Go here to see the original:
What is Machine Learning? | IBM