Facial recognition technology, as we know, hopefully is the low-hanging fruit of dangerous AI. It seems like harm is getting out of context. I will call it dangerous because that's what it is. Yet, we need to have these levels of imagination of banning certain technologies, and facial recognition technologies should be banned.
The public sector can make that choice because it is responsible to the public in the end. The private sector, as it stands, is responsible to the shareholder and to the business model of making more money. This is how capitalism works. This is what we're seeing.
That's not the job of the government. Again, when I say that AIDA should be out and reflected upon as public and private, that is exactly what I'm thinking about. I'm thinking about facial recognition technology used by law enforcement, national security, in IRCC and in immigration. Now it can be used maybe in Service Canada, or maybe in the CRA the way the IRS wanted to use facial recognition for doing taxes. Again, these technologies aren't domain-bound. Just like Palantir went from the military to health, FRT, facial recognition technology, works the same way. The public sector needs to be involved and to be publicly accountable to its people.
I really am coming back to Bianca's points about democracy. Participation is messy, but we need to participate in a way that there is dissent, discussion, non-compliance across the board and consensus, because it is important to make sure that these technologies will no longer be used because they are too dangerous. We saw what happened with Clearview AI. That is a privacy case, but it is also a mass surveillance case, besides the obvious, which are the dangers and harms it has done to so many marginalized groups.