Thank you, Mr. Chair.
This time I will focus more on YouTube, a very popular and influential platform, just like Google, of which it is a part. I was talking about it earlier. YouTube sometimes directs users to extreme, unreliable content that reports or occasionally praises conspiracy theories. YouTube makes this content look like real information.
I was wondering if you had any details about the algorithm used for users, once they are on a web page displaying, say, political content, since that is the subject of our discussion today. The algorithm will give them suggestions for other videos on the right of the page they are viewing, or under the video if they are using a mobile phone. What algorithm is used and what is the degree of transparency of this algorithm that suggests content to users when they are on a particular web page?
What mechanism is there to ensure that this content does not praise conspiracy theories or give fake news, unreliable information or, perhaps, unbalanced information, in other words information that may just promote an idea or vision, a political party?
What degree of transparency and what mechanism have you put in place to ensure that the content that is suggested to users is quality content, that it is balanced in terms of public policy, political parties and political ideas as well?