Code of Practice on Disinformation. A Comparative Analysis: Methodological Limitations

There has been much ado in the press over the last month or so concerning the Twitter platform (Twitter X) and its alleged inadequacy in tackling the phenomena of disinformation/misinformation. As discussed in my previous post Potency of the European Union's Code of Practice on Disinformation the European Commission’s (EC) has put into place the Digital Service Act (DSA) which came into effect on August 25, 2023.

On September 26, 2023, Lisa O’Carroll wrote an article published in The Guardian with the following headline: EU warns Elon Musk after Twitter found to have highest rate of disinformation. Ms. O’Carroll reported:

“The EU has issued a warning to Elon Musk to comply with sweeping new laws on fake news and Russian propaganda, after X – formerly known as Twitter – was found to have the highest ratio of disinformation posts of all large social media platforms.”

The laws Ms. O’Carroll referred to are those enshrined in the DSA. Ms. O’Carroll made reference to a report that underpins the EU’s warning issued to Mr. Musk. This September 2023 Semi-Annual Report was compiled by an organisation called TrustLab. The Title of the report is: Code of Practice on Disinformation hereafter referred to as the Code.

Ms. O’Carroll related that the TrustLab report goes about “laying bare for the first time the scale of fake news on social media across the EU”.

Chris Vallance wrote for the BBC on September 26, 2023 reported, an article entitled Disinformation mostactive on X, formerly known as Twitter, EU says. Concerning TrustLab’s report Mr. Vallance wrote:

“It examined over 6,000 unique social media posts across Facebook, Instagram, LinkedIn, TikTok, X, and YouTube.”

Mr. Vallance quotes the EU’s Values and Transparency Commissioner, Vera Jourova’s warning:

"My message for [X] is: you have to comply with the hard law. We'll be watching what you're doing,"

In the light of these remarks, it is worthwhile having a closer look at the details of the TrustLab Report, something, by the way Ms. O’Carroll and Mr. Vallance and the mainstream media in general, in reporting have not tended to offer, in their commentaries much elaboration on. This lack of expansion can lead to a misleading impression that glosses over the challenges involved in identifying disinformation/misinformation in practice.

Neither Ms. O’Carroll and Mr. Vallance clearly report what the authors of the TrustLab report wrote about the report’s methodology and the caveat’s these authors have regarding their conclusions. This is why it is worth looking at what the TrustLab in their report say about their methodology.

The key points made in the opening remarks of the TrustLab Report’s Executive Summary relate to the “ambiguous and fast-changing” nature of “online disinformation” making the measurement of this phenomenon “challenging”.  The authors of the report also mention that their Report forms the “the first empirical application of the Code”. The purpose of this investigation was to “evaluate the prevalence and sources of disinformation across six major social media platforms” operating in Poland, Slovakia and Spain. The platforms covered by the study being Facebook, Instagram, LinkedIn, X (formally Twitter) and YouTube.

The authors, in the Executive Summary outline the “key metrics” examined by the Report. These are:

“…. discoverability, relative post engagement, absolute post engagement, and properties about disinformation actors, including ratio of disinformation actors, their account activities, and engagement with other users.”

In the concluding remarks of their Executive Summary the TrustLab authors write that the “study establishes an initial benchmark for the implementation of the Code” that opens up the opportunity “for further discussion and advances in the measurement of disinformation.” In saying this the authors acknowledge that the measurement of disinformation is challenging “and accordingly our metrics and methodology can, and should be, improved.”

The TrustLab authors in one of the concluding sections of their report entitled Limitations and Implications layout cautionary remarks concerning the methodology and data collection approach.

In terms of the data collection underpinning the report the TrustLab authors admit that when interpreting the data any conclusion drawn comes with “…. the caveat that budget and time constraints have led to small sample sizes and imprecise estimates for some metrics.”

Another limitation highlighted, by the TrustLab authors relates to the “…. lack of direct access to platform data in the current study [which] limits its capacity to measure broader dimensions of online disinformation.”

A further limitation that the TrustLab authors mention is that “…. the manual labelling of mis/disinformation content and disinformation actors may be subject to human error.”  In saying this they acknowledge that even with sufficient “pre-training” and “ongoing feedback” in the data collection process “… a small number of labelling errors are possible.”  They hope to further reduce such errors in future projects.

Another limiting factor, according to the TrustLab authors is the fact that “… platform heterogeneity can still present challenges.”  They elaborate on this point as follows:

“Not all types of engagement are accessible across all platforms, and even the same type of engagement may carry different meanings and implications depending on the specific content being engaged with and the functionality and established norms of the platform.”

As well as these limitations the TrustLab authors point to how their methodology can be improved through more access to platform data coupled with more time and budgetary resources. They point out that more resources would enable them to augment their methodology with a “feed-based approach”, a tool that was not used in their study due to budgetary constraints; given the resources such an approach could be included in future studies.

Finally, the TrustLab authors describe a limitation of their study revolving around trend analysis:

“The current study provided three biweekly measurements of disinformation, which does not inform meaningful time trends.”

They point out that for future analyses “Longitudinal analysis that tracks the same metrics over a long period of time can yield more robust platform measures and better capture time trends.”

The TrustLab authors to their credit disclose that:

“… ‘foreign interference in the information space’ is outside the scope of the current study. Better access to platform data and alignment on operational definitions can enable future measurements to address this limitation.”

The TrustLab authors admit that the science behind their study is still in its infancy requiring further discussion and refinement. So, we can see that the caution with which the TrustLab authors approach their work is not reflected in the mass media’s commentary, including Ms. O’Carroll’s in the Guardian and Mr. Vallance’s in the BBC which give their readers the impression, without mentioning any of the caveats that the science behind the measurement of online disinformation is based on well-established practices, rather than an evolving methodology and data collection processes.

So given the fact that the TrustLab report methodology and practice is still in its early stages misleading virtue signalling, sensationalism and finger pointing of online platforms such as Twitter X by news media needs to be avoided lest we lose sight of the science and allow politics to drive the public’s understanding of what constitutes disinformation/misinformation and what does not.

The Fundamental point, one that needs to be addressed first and before any discussion of methodology or data collection approaches that is not really discussed, by the press in any great depth is what is it that constitutes misinformation or disinformation? The assumption, in the mass media seems to be that these concepts are well defined and beyond doubt. As mentioned above this assumption is not shared by the TrustLab authors who see that our understanding of these terms is still evolving. What these terms mean is still open to debate.  The TrustLab authors accept the EU’s definitions, under the Code of Practice. They write that the basis of their definitions is:

“…. based on an amalgamation of peer-reviewed studies that TrustLab considers to be broadly aligned with industry standards that were then adopted for TrustLab’s policies. These definitions are aligned with the European Union’s 2022 Code of Practice on Disinformation.”

They go out of their way to stress “…. that consistency of methodology and metrics will further improve as the discussions among stakeholders about detailed definitions of misinformation, disinformation, engagement, and other factors continue.” 

One cannot stress enough how important it is that the terms: misinformation, disinformation and Mis/Disinformation be defined precisely because there is much at stake. We need to ensure that in a quest to rid ourselves of misinformation and disinformation we don’t impact on the free flow of ideas. The fear of being fined would likely force online platforms such as Twitter X and the others to erroneously, on the basis of imprecise definitions and political pressure censor content.  The risk in this situation is that the overzealous curtailing of news and opinions on the internet could very well impede the flow of the truth.

The Australian Human Rights Commission in a submission about the Australian Federal Government’s proposed Communications Legislation Amendment (Combating Misinformation and Disinformation) Bill has concerns regarding this draft bill, which, I contend can be extended to all attempts to define misinformation expressed in the following warning:

“Drawing a clear line between truth and falsehood is not always simple, and there may be legitimate differences in opinion as to how content should be characterised. The broad definitions used here risk enabling unpopular or controversial opinions or beliefs to be subjectively labelled as misinformation or disinformation, and censored as a result.”

 

Comments

Popular posts from this blog

Reflections on Bluntness and "Push Back' in International Discourse

A Discourse on Laurel and Hardy Statecraft