DISCOVERY Presented at ICT Proposer´s Day

ICT Proposers’ Day 2017 took place on 9 and 10 November in Budapest, Hungary. It focused on the 2018 Calls for Proposals of the Horizon 2020 Work Programme in the field of Information & Communication Technologies and Future and Emerging Technologies (FET).DISCOVERY project was presented at Idealist2018 Workshop and participated in international cooperation session for EU and North America cooperation in NGI…

Market, Partnership and Investment Opportunities in the USA Await You!

NearUS, the H2020 initiative aiming at building up ENRICH – European Network of Research and Innovation Centres and Hubs – in the USA has opened four fantastic calls targeted to #researchers, #entrepreneurs and #companies 1) NearUS Match & Pitch @ CES, Las Vegas, 7-11 January 2018 Deadline for applications: 12 November 2017 Target audience: Early stage entrepreneurs and…

NSF investments aim to address growing cybersecurity challenge

NSF announces $74.5 million in awards through its Secure and Trustworthy Cyberspace (SaTC) program October 31, 2017 Today, the National Science Foundation (NSF) announced $74.5 million in funding for foundational research and education that aims to address the growing cybersecurity challenge. This investment, through the NSF Secure and Trustworthy Cyberspace (SaTC) program, is critical to…

Congrats to our Canadian Partner: AI’s problem with disability and diversity

Machine learning is a technology we’ve seen move from the lab to something most of us interact with, directly or indirectly, on a daily basis. It’s a powerful approach to artificial intelligence.

Traditionally, computing relied on humans writing rules that a computer would then have to follow. In machine learning, the computer looks at data and creates patterns to help it understand the world. It predicts how things will act in the future, and then make decisions based on those predictions.

It’s this data that fuels machine learning. But in its raw form, the big data sets used to train AI are bound to contain information that isn’t relevant or that is confusing. So it needs to be refined. The data is “cleaned” by removing outliers or data that disrupts a greater pattern.

Jutta Treviranus is the director of the Inclusive Design Research Centreand a professor at OCAD University. Jutta has been working towards more inclusive and accessible technology for much of her life. Her focus is now on improving artificial intelligence systems so they can better serve everyone, including people with disabilities, and organizes hackathons towards this goal.

“One of the things we want to do when we create data sets for machine learning engines, is we want to increase the signal and decrease the noise,” Jutta says. “Meaning that we want to emphasize the dominant patterns, or the patterns that are going to be used to allow the machines to make useful decisions.

“But in the process of doing that we clean out what we call the outliers, or the edge data. Because that makes it more difficult to draw conclusion for the machines and therefore they’re not as quick to make a useful decision.”

The problem is that outlier data isn’t just irrelevant. It can represent the experience of real world people. And while machine learning allows computers to make their own decisions, those decisions are essentially coloured by the data we provide them with.  “Our research is not very supportive to diversity,” Jutta says, “and we have basically transferred that to machine learning.”

“The silver lining to this is that it’s made it possible to talk about things like stereotype and prejudice and bias and blind spots with hard scientists. Previously what is called soft science was seen as very fuzzy and indeterminate and didn’t have sufficient hard edges. But because the machines make it manifest it becomes a topic that is respected and people will pay attention to and engage in.”

One of the things holding back improving machine learning systems for people with disabilities and other outliers is the use of proprietary software.

“We need open systems,” Jutta says. “We need transparent systems in order to serve individuals or to create systems that work for people with disabilities or that work for diversity. If a system is a black box, if it is a closed system,  then there’s no way to query it, there’s no way to challenge who you’ve excluded.  And that is a really dangerous thing that happens.”

“Now that we have networked systems we can design our systems for diversity. We can personalize them, customize them, make them flexible, so that they address our personal needs. And if there are gaps or things that aren’t there for particular requirements someone migh