One thing I would say is, as somebody was talking about, this isn't new. In an election campaign, somebody can, the night before an election, send leaflets to a whole constituency with a false rumour about a candidate. This issue that we would legislate around content is just not possible, because a lot of this stuff is the grey, murky, misleading space.
I do think there's something specific around content that makes the election system bumble. For example, we were monitoring the election in Brazil two weeks ago. On election day, there was a great deal of rumour circulating around the fact that the machines weren't working and that you could stay at home to vote via SMS. I think if we're talking about content, that's the kind of space where there is room to say, if the harm is specifically around the election, then there is something that could be done around that.
I think we need more transparency around behaviours, not content. The platforms are moving in this direction, but they need more pressure to be placed on them in terms of what is a behaviour that we can see that we would have a problem with and we would all agree about, whether it's automation, whether the IP address is external to the Canadian border or people using fake accounts.
I think behaviour is something that is worth looking at, but the content part of this is something that is much more challenging. We need more pressure on the platforms to be more transparent about those behaviours, because we don't know what decisions they're making. It's completely opaque at the moment.