Private AI Research Institute | Avast

Private AI Research Institute | Avast
Emma McGowan, 18 January 2021

The Institute fosters the design and development of techniques for privacy-focused machine learning



Artificial intelligence (AI) was once something you only heard about in science fiction — but not anymore. These days, AI is used for everything from computers playing chess to self-driving cars to robots you can actually interact with. But the development of AI has been largely decentralized and siloed, creating a dual sided problem: On one side, researchers don’t have access to the data they need. On the other, expanding access to data creates more possibilities for privacy breaches.
The Private AI Collaborative Research Institute, originally established by Intel’s University Research & Collaboration Office (URC), aims to change that by bringing the private sector together with academics to create a collaborative environment in which researchers from a wide range of backgrounds can advance and develop privacy-focused technologies for decentralized AI.
“Machine learning has the potential to profoundly alter data analysis,” Professor Nicolas Papernot, of University of Toronto, tells Avast. “Machine learning requires diverse and large labeled datasets to perform well and make predictions that are useful. Yet, machine learning algorithms are known to leak private data they analyze in these datasets. I believe the center for Private AI will be instrumental to the design and development of techniques for privacy-preserving machine learning that enable institutions to analyze data in a more respectful manner.”
On the private side, Intel invited Avast and Borsetta to join the Institute. On the public side, eight universities with nine research projects were chosen to kick off the first year and will ..

Support the originator by clicking the read the rest link below.