Archive for the ‘Machine Learning’ Category

Google Will Utilize AI And Machine Learning Algorithms To Fine Tune Gmail Suggestions – Digital Information World

Google's setting up Gmail for an upgrade, as it employs ML models to help users with better search suggestions.

Google has been tampering with, and soon after disposing of, new technology and updates since its inception. We have the Google Graveyard as a testament to such haphazard progress. However, let's be real: the tech giant knows when it's onto a good thing, and Gmail is that good thing. The email-based service is one of Google's trademark platforms and continues to flourish even as emails are considered more and more restricting. No one types out an email to converse when they can use WhatsApp or Instagram. However, rarely do people utilize the same platforms to send out CVs or letters of recommendation: even if they could, it would be considered informal and inappropriate.

That is the sort of environment Google has managed to cultivate with Gmail. It's what Hotmail and AOL chat rooms could never really muster up; Gmail is a brand that resonates with professionalism and the company leans into it as well. This brings us to today's feature of interest, and how Google will utilize it to the email service's full effect.

ML or machine learning algorithms have evolved from being a novel, rarely used concept that evoked "oohs" and "aahs", into a tool that's almost regularly utilized by every social media platform with skin in the game. Honestly, they fit the social media market quite well: ML algorithms rely on using tons of data to generate whatever they're ordered. Automated messages can appear more natural if AI spends its time looking at and attempting to emulate examples of normal speech. With social media platforms (I'm willing to loosen the definition enough to include email platforms) providing a near-endless well of such information to draw from, it's open season for ML programs.

Google intends on utilizing ML models to help AI-based suggestions grow with the writer. To be clearer, Gmail's word suggestions for users will come to reflect what the user is aiming for based on previous interactions. This way, a consumer's suggested vernacular can essentially come forward as a more elegant version of what they'd employ in day-to-day life.

These suggestions will also be heavily utilized in helping users sift through folders and such, attempting to look for prior emails or other similar content. ML models can learn from the keywords that a user utilizes in looking up content in the past, and help narrow down searches in the future.

Google will be rolling out these new updated ML models for user testing across Android audiences.

See more here:
Google Will Utilize AI And Machine Learning Algorithms To Fine Tune Gmail Suggestions - Digital Information World

The imperative need for machine learning in the public sector – VentureBeat

We are excited to bring Transform 2022 back in-person July 19 and virtually July 20 - 28. Join AI and data leaders for insightful talks and exciting networking opportunities. Register today!

The sheer number of backlogs and delays across the public sector are unsettling for an industry designed to serve constituents. Making the news last summer was the four-month wait period to receive passports, up substantially from the pre-pandemic norm of 6-8 weeks turnaround time. Most recently, the Internal Revenue Service (IRS) announced it entered the 2022 tax season with 15 times the usual amount of filing backlogs, alongside its plan for moving forward.

These frequently publicized backlogs dont exist due to a lack of effort. The sector has made strides with technological advancements over the last decade. Yet, legacy technology and outdated processes still plague some of our nations most prominent departments. Todays agencies must adopt digital transformation efforts designed to reduce data backlogs, improve citizen response times and drive better agency outcomes.

By embracing machine learning (ML) solutions and incorporating advancements in natural language processing (NLP), backlogs can be a thing of the past.

Whether tax documents or passport applications, processing items manually takes time and is prone to errors on the sending and receiving sides. For example, a sender may mistakenly check an incorrect box or the receiver may interpret the number 5 as the letter S. This creates unforeseen processing delays or, worse, inaccurate outcomes.

But managing the growing government document and data backlog problem is not as simple and clean-cut as uploading information to processing systems. The sheer number of documents and citizens information entering agencies in varied unstructured data formats and states, often with poor readability, make it nearly impossible to reliably and efficiently extract data for downstream decision-making.

Embracing artificial intelligence (AI) and machine learning in daily government operations, just as other industries have done in recent years, can provide the intelligence, agility and edge needed to streamline processes and enable end-to-end automation of document-centric processes.

Government agencies must understand that real change and lasting success will not come with quick patchworks built upon legacy optical character recognition (OCR) or alternative automation solutions, given the vast amount of inbound data.

Bridging the physical and digital worlds can be attained with intelligent document processing (IDP), which leverages proprietary ML models and human intelligence to classify and convert complex, human-readable document formats. PDFs, images, emails and scanned forms can all be converted into structured, machine-readable information using IDP. It does so with greater accuracy and efficiency than legacy alternatives or manual approaches.

In the case of the IRS, inundated with millions of documents such as 1099 forms and individuals W-2s, sophisticated ML models and IDP can automatically identify the digitized document, extract printed and handwritten text, and structure it into a machine-readable format. This automated approach speeds up processing times, incorporates human support where needed and is highly effective and accurate.

Alongside automation and IDP, introducing ML and NLP technologies can significantly support the sectors quest to improve processes and reduce backlogs. NLP isan area of computer science that processes and understands text and spoken words like humans do, traditionally grounded in computational linguistics, statistics and data science.

The field has experienced significant advancements, like the introduction of complex language models that contain more than 100 billion parameters. These models could power many complex text processing tasks, such as classification, speech recognition and machine translation. These advancements could support even greater data extraction in a world overrun by documents.

Looking ahead, NLP is on course to reach the level of text understanding capability similar to that of a human knowledge worker, thanks to technological advancements driven by deep learning.Similar advancements in deep learning also enable the computer to understand and process other human-readable content such as images.

For the public sector specifically, this could be images included in disability claims or other forms or applications consisting of more than just text. These advancements could also improve downstream stages of public sector processes, such as ML-powered decision-making for agencies determining unemployment assistance, Medicaid insurance and other invaluable government services.

Though weve seen a handful of promising digital transformation improvements, the call for systemic change has yet to be fully answered.

Ensuring agencies go beyond patching and investing in various legacy systems is needed to move forward today. Patchwork and investments in outdated processes fail to support new use cases, are fragile to change and cannot handle unexpected surges in volume. Instead, introducing a flexible solution that can take the most complex, difficult-to-read documents from input to outcome should be a no-brainer.

Why? Citizens deserve more out of the agencies who serve them.

CF Su is VP of machine learning at Hyperscience.

Welcome to the VentureBeat community!

DataDecisionMakers is where experts, including the technical people doing data work, can share data-related insights and innovation.

If you want to read about cutting-edge ideas and up-to-date information, best practices, and the future of data and data tech, join us at DataDecisionMakers.

You might even considercontributing an articleof your own!

Read More From DataDecisionMakers

Read more here:
The imperative need for machine learning in the public sector - VentureBeat

A technique to improve both fairness and accuracy in artificial intelligence – MIT News

For workers who use machine-learning models to help them make decisions, knowing when to trust a models predictions is not always an easy task, especially since these models are often so complex that their inner workings remain a mystery.

Users sometimes employ a technique, known as selective regression, in which the model estimates its confidence level for each prediction and will reject predictions when its confidence is too low. Then a human can examine those cases, gather additional information, and make a decision about each one manually.

But while selective regression has been shown to improve the overall performance of a model, researchers at MIT and the MIT-IBM Watson AI Lab have discovered that the technique can have the opposite effect for underrepresented groups of people in a dataset. As the models confidence increases with selective regression, its chance of making the right prediction also increases, but this does not always happen for all subgroups.

For instance, a model suggesting loan approvals might make fewer errors on average, but it may actually make more wrong predictions for Black or female applicants. One reason this can occur is due to the fact that the models confidence measure is trained using overrepresented groups and may not be accurate for these underrepresented groups.

Once they had identified this problem, the MIT researchers developed two algorithms that can remedy the issue. Using real-world datasets, they show that the algorithms reduce performance disparities that had affected marginalized subgroups.

Ultimately, this is about being more intelligent about which samples you hand off to a human to deal with. Rather than just minimizing some broad error rate for the model, we want to make sure the error rate across groups is taken into account in a smart way, says senior MIT author Greg Wornell, the Sumitomo Professor in Engineering in the Department of Electrical Engineering and Computer Science (EECS) who leads the Signals, Information, and Algorithms Laboratory in the Research Laboratory of Electronics (RLE) and is a member of the MIT-IBM Watson AI Lab.

Joining Wornell on the paper are co-lead authors Abhin Shah, an EECS graduate student, and Yuheng Bu, a postdoc in RLE; as well as Joshua Ka-Wing Lee SM 17, ScD 21 and Subhro Das, Rameswar Panda, and Prasanna Sattigeri, research staff members at the MIT-IBM Watson AI Lab. The paper will be presented this month at the International Conference on Machine Learning.

To predict or not to predict

Regression is a technique that estimates the relationship between a dependent variable and independent variables. In machine learning, regression analysis is commonly used for prediction tasks, such as predicting the price of a home given its features (number of bedrooms, square footage, etc.) With selective regression, the machine-learning model can make one of two choices for each input it can make a prediction or abstain from a prediction if it doesnt have enough confidence in its decision.

When the model abstains, it reduces the fraction of samples it is making predictions on, which is known as coverage. By only making predictions on inputs that it is highly confident about, the overall performance of the model should improve. But this can also amplify biases that exist in a dataset, which occur when the model does not have sufficient data from certain subgroups. This can lead to errors or bad predictions for underrepresented individuals.

The MIT researchers aimed to ensure that, as the overall error rate for the model improves with selective regression, the performance for every subgroup also improves. They call this monotonic selective risk.

It was challenging to come up with the right notion of fairness for this particular problem. But by enforcing this criteria, monotonic selective risk, we can make sure the model performance is actually getting better across all subgroups when you reduce the coverage, says Shah.

Focus on fairness

The team developed two neural network algorithms that impose this fairness criteria to solve the problem.

One algorithm guarantees that the features the model uses to make predictions contain all information about the sensitive attributes in the dataset, such as race and sex, that is relevant to the target variable of interest. Sensitive attributes are features that may not be used for decisions, often due to laws or organizational policies. The second algorithm employs a calibration technique to ensure the model makes the same prediction for an input, regardless of whether any sensitive attributes are added to that input.

The researchers tested these algorithms by applying them to real-world datasets that could be used in high-stakes decision making. One, an insurance dataset, is used to predict total annual medical expenses charged to patients using demographic statistics; another, a crime dataset, is used to predict the number of violent crimes in communities using socioeconomic information. Both datasets contain sensitive attributes for individuals.

When they implemented their algorithms on top of a standard machine-learning method for selective regression, they were able to reduce disparities by achieving lower error rates for the minority subgroups in each dataset. Moreover, this was accomplished without significantly impacting the overall error rate.

We see that if we dont impose certain constraints, in cases where the model is really confident, it could actually be making more errors, which could be very costly in some applications, like health care. So if we reverse the trend and make it more intuitive, we will catch a lot of these errors. A major goal of this work is to avoid errors going silently undetected, Sattigeri says.

The researchers plan to apply their solutions to other applications, such as predicting house prices, student GPA, or loan interest rate, to see if the algorithms need to be calibrated for those tasks, says Shah. They also want to explore techniques that use less sensitive information during the model training process to avoid privacy issues.

And they hope to improve the confidence estimates in selective regression to prevent situations where the models confidence is low, but its prediction is correct. This could reduce the workload on humans and further streamline the decision-making process, Sattigeri says.

This research was funded, in part, by the MIT-IBM Watson AI Lab and its member companies Boston Scientific, Samsung, and Wells Fargo, and by the National Science Foundation.

Originally posted here:
A technique to improve both fairness and accuracy in artificial intelligence - MIT News

Trainual Leverages AI and Machine Learning to Give SMBs a Faster Way to Onboard and Train – EnterpriseTalk

Trainual, the leading training management system for small businesses and growing teams, today released an AI-powered documentation engine for outlining roles and responsibilities. The Suggested Roles and Suggested Responsibilities features allow users of its platform to leverage the learnings of thousands of growing organizations around the world by recommending roles by company type, along with the responsibilities associated with those roles. Trainual accomplishes this with proprietary data that connects which types of trainings have been assigned to comparable job titles from similar businesses in every industry.

Small businesses create 1.5 million jobs annually inthe United States, accounting for 64% of annual averages (source). With Suggested Roles and Responsibilities, small business owners and leaders have tools to quickly identify the duties for new roles within their organization, and map training materials to them.

Also Read: Three Approaches for Leveraging Remote Teams in IT Sector

Every small business is unique. As they grow, so does their employee count and the mix of different roles they have within their companies. And along with each role comes a new set of responsibilities that can take lots of time to think up and document, saidChris Ronzio, CEO and Founder of Trainual. We decided to make that process easier. Using artificial intelligence (AI) and machine learning, Trainual is providing small business owners and managers the tools to easily keep their roles up-to-date and the people that hold them, trained in record time.

The process is simple. When a company goes to add a new role, theyll automatically see a list of roles (AKA job titles) that similar businesses have added to their companies. After accepting a suggested role in the Trainual app, theyll see a list of suggested responsibilities, curated utilizing AI and Trainuals own machine learning engine. Owners, managers, and employees can then easily add context to all of the responsibilities for every role in the business by documenting or assigning existing content thats most relevant for onboarding and ongoing training.

Check Out The NewEnterprisetalk Podcast.For more such updates follow us on Google NewsEnterprisetalk News.

Visit link:
Trainual Leverages AI and Machine Learning to Give SMBs a Faster Way to Onboard and Train - EnterpriseTalk

Researchers use AI to predict crime, biased policing in cities – Los Angeles Times

For once, algorithms that predict crime might be used to uncover bias in policing, instead of reinforcing it.

A group of social and data scientists developed a machine learning tool it hoped would better predict crime. The scientists say they succeeded, but their work also revealed inferior police protection in poorer neighborhoods in eight major U.S. cities, including Los Angeles.

Instead of justifying more aggressive policing in those areas, however, the hope is the technology will lead to changes in policy that result in more equitable, need-based resource allocation, including sending officials other than law enforcement to certain kinds of calls, according to a report published Thursday in the journal Nature Human Behavior.

The tool, developed by a team led by University of Chicago professor Ishanu Chattopadhyay, forecasts crime by spotting patterns amid vast amounts of public data on property crimes and crimes of violence, learning from the data as it goes.

Chattopadhyay and his colleagues said they wanted to ensure the system not be abused.

Rather than simply increasing the power of states by predicting the when and where of anticipated crime, our tools allow us to audit them for enforcement biases, and garner deep insight into the nature of the (intertwined) processes through which policing and crime co-evolve in urban spaces, their report said.

For decades, law enforcement agencies across the country have used digital technology for surveillance and predicting on the belief it would make policing more efficient and effective. But in practice, civil liberties advocates and others have argued that such policies are informed by biased data that contribute to increased patrols in Black and Latino neighborhoods or false accusations against people of color.

Chattopadhyay said previous efforts at crime prediction didnt always account for systemic biases in law enforcement and were often based on flawed assumptions about crime and its causes. Such algorithms gave undue weight to variables such as the presence of graffiti, he said. They focused on specific hot spots, while failing to take into account the complex social systems of cities or the effects of police enforcement on crime, he said. The predictions sometimes led to police flooding certain neighborhoods with extra patrols.

His teams efforts have yielded promising results in some places. The tool predicted future crimes as much as one week in advance with roughly 90% accuracy, according to the report.

Running a separate model led to an equally important discovery, Chattopadhyay said. By comparing arrest data across neighborhoods of different socioeconomic levels, the researchers found that crime in wealthier parts of town led to more arrests in those areas, at the same time as arrests in disadvantaged neighborhoods declined.

But, the opposite was not true. Crime in poor neighborhoods didnt always lead to more arrests suggesting biases in enforcement, the researchers concluded. The model is based on several years of data from Chicago, but researchers found similar results in seven other larger cities: Los Angeles; Atlanta; Austin, Texas; Detroit; Philadelphia; Portland, Ore.; and San Francisco.

The danger with any kind of artificial intelligence used by law enforcement, the researchers said, lies in misinterpreting the results and creating a harmful feedback of sending more police to areas that might already feel over-policed but under-protected.

To avoid such pitfalls, the researchers decided to make their algorithm available for public audit so anyone can check to see whether its being used appropriately, Chattopadhyay said.

Often, the systems deployed are not very transparent, and so theres this fear that theres bias built in and theres a real kind of risk because the algorithms themselves or the machines might not be biased, but the input may be, Chattopadhyay said in a phone interview.

The model his team developed can be used to monitor police performance. You can turn it around and audit biases, he said, and audit whether policies are fair as well.

Most machine learning models in use by law enforcement today are built on proprietary systems that make it difficult for the public to know how they work or how accurate they are, said Sean Young, executive director of the University of California Institute for Prediction Technology.

Given some of the criticism around the technology, some data scientists have become more mindful of potential bias.

This is one of a number of growing research papers or models thats now trying to find some of that nuance and better understand the complexity of crime prediction and try to make it both more accurate but also address the controversy, Young, a professor of emergency medicine and informatics at UC Irvine, said of the just-published report.

Predictive policing can also be more effective, he said, if its used to work with community members to solve problems.

Despite the studys promising findings, its likely to raise some eyebrows in Los Angeles, where police critics and privacy advocates have long railed against the use of predictive algorithms.

In 2020, the Los Angeles Police Department stopped using a predictive-policing program called Pred-Pol that critics argued led to heavier policing in minority neighborhoods.

At the time, Police Chief Michel Moore insisted he ended the program because of budgetary problems brought on by the COVID-19 pandemic. He had previously said he disagreed with the view that Pred-Pol unfairly targeted Latino and Black neighborhoods. Later, Santa Cruz became the first city in the country to ban predictive policing outright.

Chattopadhyay said he sees how machine learning evokes Minority Report, a novel set in a dystopian future in which people are hauled away by police for crimes they have yet to commit.

But the effect of the technology is only beginning to be felt, he said.

Theres no way of putting the cat back into the bag, he said.

Read the rest here:
Researchers use AI to predict crime, biased policing in cities - Los Angeles Times