This report analyzes the Misinformation, Disinformation, and Malinformation (MDM) space, also referred to as Information Disorder, within the Middle East and North Africa (MENA) region while considering the rise of generative AI. The report attempts to answer key research questions such as understanding the scale of MDM in MENA, identifying who is behind them, and identifying emerging patterns and trends from multifaceted perspectives. Additionally, the report highlighted the role of AI and what key players in MENA we should keep an eye on, for those countering the information disorder.
We applied a combination of quantitative and qualitative research methods on AFH’s fact-checked posts database as well as our series of publications called Eye on Twitter. The combined analysis covered campaigns and posts published between 2021 and the end of Q3 in 2023. Our fact-checked posts database and publications cover stories that involve multiple platforms and actors. Our research and findings focus on Egypt mainly, and to a lesser extent, Lebanon, Iraq, Jordan, and Yemen. Additionally, we conducted a literature survey focusing on the practical side of MDM using some of the well-documented incidents and campaigns across MENA that spanned the period from 2011 until 2023.
The research report explores the scale of MDM in MENA, how it is evolving over time, and whether it is becoming a full-fledged, mature industry and sense the maturity level of the ecosystem. It is also seeking to understand the key actors behind MDM campaigns in MENA and whether new types of actors are getting more involved in driving the information disorder.
It analyzes the key drivers behind MDM as well as the main tactics applied in MENA. The report discusses some tactics in detail, some newly emerging tactics and compiles others into the appendix. The study also focuses on MENA’s fact checkers and captures emerging patterns in their operations, driven by fact checkers and by the surrounding ecosystem.
The study lists the key themes/topics targeted by MDM campaigns in MENA based on AFH data. It also analyzes the key themes that drive online MDM in MENA towards physical harm.
Finally, the study surveys available literature and identifies trends on the role of AI in media and journalism in MENA, and how MDM is going to be affected by the rise of GenAI. A list of benevolent players in Media and AI who can be part of mitigating the MDM risks in MENA was created.
This report draws from pioneering research titled “Meeting the Challenges of Information Disorder in the Global South,” [1] led by Professor Herman Wasserman, which maps the actors currently leading the counter-information disorder, and their approaches, tools, and methods. Furthermore, it studies the research landscape and identifies key areas for further research. Our report focuses on those creating the information disorder in MENA, their tactics, and how they exploit some of the challenges identified by Wasserman’s research in the counter-information disorder ecosystem.
The following are selected key findings highlighted by this research study:
Key findings on the scale of MDM in MENA:
Key findings on MDM actors in MENA:
Key findings on MDM tactics in MENA:
To MENA Fact Checkers
To MENA Governments
Arabi Facts Hub is a nonprofit organization dedicated to researching information disorder in Arabic content on the Internet and providing innovative solutions to detect and identify it. Information disorder can be described as publishing false information without the intent of harming, like misinformation, or with the intent of harming like disinformation and malinformation (MDM). This study refers to information disorder and MDM interchangeably.
In our first research report, we studied the MDM space covering the Middle East and North Africa (MENA) region. This was done while considering the rise of generative AI. We focused on the offensive side of the ecosystem by trying to answer the following questions:
Then, we moved to the defensive side of the ecosystem and highlighted the emerging patterns among MENA actors and who are the key players in Media & AI in MENA that we should keep an eye on, for those countering the information disorder.
We relied mainly on AFH’s fact-checked posts database as well as our series of publications called Eye on Twitter to conduct this research. The combined analysis covered campaigns and posts published between 2021 and the end of Q3 of 2023, where platforms and involved actors were included in the study. Our research and findings focused on Egypt mainly, and to a lesser extent, Lebanon, Iraq, Jordan, and Yemen. Furthermore, we reviewed well-documented incidents and campaigns across MENA that spanned the period from 2011 until 2023.
As part of this research, we identified areas that needed further exploration and gaps in the Arabic MDM data landscape that should be bridged. AFH aims to play an active role in the MDM research, policy, and data spaces in MENA through publishing more in-depth and data-informed reports.
There is a general agreement that the MDM ecosystem in MENA is alarmingly massive, targeting wide audiences who have relatively high connectivity, device availability, and platform usage trends in that region [3]. Yet, it is very difficult to find an independent research or media report that quantifies the scale of what has likely morphed into a full-fledged MDM industry, with supply and demand dynamics. Probably the obvious reason behind that is the restrictions on freedom of press and expression in MENA, which are needed to uncover one the main actors behind MDM in MENA: Arab regimes[4] Another reason is the fact that MDM and Influence Operations (IO) are secretive and deceptive[5] . There are many IO happening, not just in MENA but worldwide, and people behind it conceal their identities. As a result, we lack sufficient information to easily identify the IO and who orchestrates them. But we can somewhat infer the scale by looking into the known incidents of IO that targeted a specific country in MENA or had a regional scope.
Since Twitter, now X, is one of the main mediums that is targeted by disinformation campaigns in MENA, Twitter Transparency and Twitter Moderation Research Consortium (TMRC) published a public archive of state-backed information operations in October 2018[6] . Soon enough, in September 2019, Twitter announced two MENA-related takedowns. The first one took down a network of 267 accounts based in the United Arab Emirates and Egypt, which were involved in influencing operations targeting Qatar and Iran. This incident highlighted the role of private companies in such operations, as the private company behind these accounts was identified to be an Abu Dhabi and Cairo-based company called DotDev[7] . Simultaneously, Twitter suspended six accounts posing as independent media outlets, yet promoting narratives favorable to the Saudi government. The second takedown involved the removal of 4,258 accounts operating exclusively from the UAE and tweeted mostly about Qatar and Yemen[8] .
Two months following that, in December 2019, Twitter announced the largest state-backed disinformation network it ever detected, which originated from Saudi Arabia. The core network included 5,929 accounts, on which TMRC published comprehensive data. However, these accounts were the tip of the iceberg as they controlled an extended network of 88,000 accounts, all of which were suspended amid the announcement[9] . Twitter was able to trace back all these accounts to a social media and marketing company based in Saudi Arabia called Smaat. As part of this operation[10] , two Saudi nationals linked to the matter, Ali Alzabarah and Ahmed Almutairi/Aljibreen, were charged and listed on the FBI's Counterintelligence Wanted list[11] [12] . This behavior of groups of pages and people working together to mislead others about who they are, their perception, and what they are doing is called Coordinated Inauthentic Behavior (CIB)[13] .
Coordinated disinformation campaigns are not new in MENA, but the 2019 takedowns were the largest, indicating how mature and vicious this industry evolved with time. Probably the earliest documented campaign was initiated from Iran in 2010, targeting Saudi Arabia beside other countries. The operation included 783 Facebook pages, accounts, and groups[14] . Moreover, 2015 witnessed the inception of disinformation campaigns led by two companies: New Waves in Egypt and Newave in the United Arab Emirates[15] . The Disinformation Visualizer project published by the Atlantic Council's DFRLab and Google's Jigsaw showcases the identified coordinated disinformation campaigns globally between 2010 and 2019[16] .
Reviewing the identified CIB MDMs beginning in 2020, it can be noticed that the MDM scale in MENA expanded in terms of regions, amount of published content, and campaigns launched by foreign governments. The Stanford Internet Observatory (SIO) studied many of Twitter's seized accounts managed from Saudi, UAE, and Egypt during that year and concluded that the content was regional and mainly criticized the Syrian government and Iranian influence in Iraq, and promoted Haftar in Libya[17] .
Shortly after that, SIO and Graphika analyzed 99 Pages, 6 Groups, 181 profiles, and 22 Instagram accounts identified by Facebook/Meta to be inauthentic and linked to the Russian Internet Research Agency (IRA). The analysis concluded that they targeted Libya, Syria, and Sudan[18] . Further analysis by SIO showed hundreds of accounts spreading hundreds of thousands of tweets, originating in Iran, and others linked to Russian networks, targeting Arabic-speaking audiences with coordinated disinformation. Many of these accounts were sock puppets and fake personas whom SIO believes they were affiliated with the Internet Research Agency (IRA) and Russian government-linked actors[19] .
Beside the scalability increase in regional targeting, the country-specific targeting among MENA rivalries continued. Some of these campaigns even had "defining characteristics" like the network of 33 fake accounts claimed to be dissident members of the Qatari royal family but turned out to be sock puppet accounts linked to the government of Saudi Arabia. These accounts "repeatedly changed their screen names and deleted earlier tweets" as highlighted by SIO[20] . Some of these campaigns capitalized on nonpolitical topics to push political propaganda, like the network of 9,000 pro–United Arab Emirates network that used Coronavirus hashtags to criticize Turkey’s military intervention in Libya and praise the government of UAE[21] .
Domestic coordinated disinformation campaigns were on the rise, too. A notable takedown was announced by Facebook on August 6, 2020, where Yemen-Based 28 Pages, 15 Groups, 69 Facebook accounts, and ten Instagram accounts were suspended for impersonating government agencies to push anti-Houthi content[22] .
Researchers from Stanford University curated an important dataset documenting all known Facebook and Twitter takedowns centered on Influence Operations (IO) in the MENA region, especially between 2020 and 2021. The dataset contains 46 information operations, originating from ten MENA countries. Iran was the highest with 20 of the 46 takedowns. Egypt was second with 10 takedowns, then the UAE with 6, and Saudi Arabia was fourth with 5. The researchers found that among the 46 takedowns, 24% were linked to a government and 26% were attributed to a marketing, PR, or IT firm[23] . This last finding raises an important question: does the absence of specific countries like Qatar in most of the published disinformation analysis and takedowns imply that they refrain from directly engaging or investing in counter-disinformation campaigns? Or maybe private companies are filling that vacuum by providing propaganda-as-a-service for them and acting as fronts? There is no conclusive answer, but some of AFH's investigations in the next section of this report may provide some leads.
Moving to 2023, and despite the measures taken by major social media platforms to fight MDM, the CIB campaigns continue to flourish and their tactics are advancing. DFRLab published two analyses highlighting some of these tactics. The first analysis was in Sudan which involved the use of nine hundred revived dormant accounts; accounts that became active after years of inactivity, possibly hijacked. These accounts amplified content shared by Sudan’s Rapid Support Force (RSF) and Mohamed Hamdan Dagalo, known as Hemedti[24] . The second analysis highlighted how hundreds of thousands of accounts pushed hashtags promoting Egypt’s support for human rights. Essentially, these accounts tried to hijack the narrative and widely acknowledged perception of the poor human rights conditions in Egypt[25] . This activity included running slandering campaigns targeting activists like Gamal Eid[26] .
AFH Research Findings on the Scale of MDM in MENA
Our Open-Source Intelligence (OSINT) team at AFH, in collaboration with Daraj Media, published original analyses covering a wide range of MDM campaigns in MENA, which identified the increasing scale of campaigns and evolving tactics during 2023. These campaigns were selected by the AFH editorial team based on topic significance to regional and global events, the campaigns reach and trend levels, breadth and depth of used platforms and tactics. Some campaigns may appear very local, however, this should be considered with a level of abstraction, as we believe these campaigns’ information disorder dynamics are similar to other campaigns not within MENA countries only, but across the globe.
Jordan
We started with Jordan, where we investigated the role of an online campaign, which included many fake accounts, and how it interfered with the passing of a cybercrime bill in the parliament. Pro-Jordan monarchy tweets in mass numbers flooded twitter as private pictures of the crown prince and his new bride on the beach were leaked. These online campaigns seemed to have set the tone for the discussion around the controversial anti-cybercrime law which was passed in August 2023. These online campaigns seemed to be working in coordination, and 24,000 tweets came from accounts that did not disclose its geographic location. The campaigns operated in 5 main hashtags, involving 10,000 accounts, producing 34,000 tweets, viewed 24 million times, and had a reach of 32.5 million. Retweets were above the average retweet number, which was at 64.5%. This is a sign of coordinated and inauthentic engagement. Engagement with the hashtags from Egypt (465 tweets) contributed to the hashtag, resulting in hashtags showing up on regional trends[27] .
Our analysis showed that this campaign is an example of a coordinated and multifaceted campaign, with the plausible involvement of state and nonstate actors. Moreover, looking into the interaction volume and the number of accounts involved highlights the increase in scale of MDM campaigns over the years in MENA, as such a volume is relatively huge in a country like Jordan.
Lebanon
Moving to Lebanon, AFH OSINT team investigated anti-LGBTQ activity on social media in Lebanon and how it expanded to many Arab countries. The story started after Hassan Nasrallah’s, the Secretary-General of Hezbollah, remarks on ‘standing up to the LGBTQ movement.’ A chain reaction occurred from allies in the region (militias in Iraq and Yemen) amplifying the message. This investigative report looked into organized online campaigns from June 2023 (pride month) to August 2023. Search on words affiliated with LGBTQ terminology soared by 74%. During that same period, there were 248 thousand posts on the topic, with an average of 3.5 thousand tweets daily. These campaigns generated almost a million engagements, on average 14,000 engagements per day.
243,000 out of the 248,000 were on twitter, coming from 34 countries. These tweets were viewed 133 million times on twitter. In June of the same year, the Islamist Sadrist movement in Iraq launched a similar anti-LGBTQ campaign, generating 108,000 tweets, with a reach of 2 million. The report uncovered that 74% of the hashtag activity were retweeted, and only 15.9% of the tweets were original. Nasrallah’s remarks led to a ban in Iraq on the use of the terms (gender) and (homosexuality) in traditional and new media, and instead the use of more inflammatory terms.
AFH analysis highlights the scale that such campaigns can reach especially when ignited by top-down disinformation, i.e., coming from reputable and influential figures in the society. What makes this disinformation campaign unique is its ability to mobilize masses in the real world and incite them to inflict physical harm to marginalized groups.
Yemen
In Yemen, 4 hashtags surfaced after Nasrallah’s remarks. In total, 22,000 tweets were generated with a reach of 4 million. 72.8% of the engagement on the hashtags were inauthentic, and the majority of the tweets came from Yemen, then Saudi Arabia, Iraq, Lebanon, USA, Kuwait, Bahrain, Egypt, China and Syria[28] .
AFH analysis of this campaign highlighted how disinformation can travel physical borders and reach all the way from Lebanon to Yemen. Later, this report will discuss the drivers behind the MDM diffusion and its scalable spread, but in a few words, this campaign highlights how globalization and wide access to technology among MENA youth are possible factors behind the wide participation in these hashtags and the MDM entailed in them. Probably the way this campaign was framed and the fear mongering has resembled an existential threat to a large audience of conservatives, whether socially or religiously. This all played a role in causing the regional spread of this campaign.
Egypt
Moving to Egypt, we investigated the online campaigns that supported a man accused of murdering a woman on camera (Nayyira Ashraf). These campaigns spread misinformation and gender-based violence. After the defendant was convicted of murder, 9 Facebook groups (closed and open groups) were created with a total of 86,000 members. One of the largest groups (with 18,000 members) posted 3,500 posts within a month of its launch. These groups focused on supporting the defendant and claimed to “uncover” the lies against him, influence public opinion with misinformation, and spread conspiracy theories. A hashtag in support of the defendant was generated on twitter, with 41,000 interactions (between tweets, retweets and likes). The investigation names the key groups and accounts involved in this coordinated behavior[29] .
Staying in Egypt, we investigated the Egyptian accounts tweeting in support of a pro-Qatar hashtag. A hashtag surfaced in October 2022 after the German Foreign Minister criticized Qatar for poor human rights conditions in preparation for FIFA world cup 2022. The hashtag grossed 15,000 tweets, 39,900 retweets, and 16,000 likes. Total accounts that participated in the hashtag were 2,039 accounts from countries such as Qatar, Egypt, Jordan, Saudi Arabia, Yemen, Oman and Kuwait.
Some of the most active accounts on the hashtag had affiliations with Muslim Brotherhood and anti-regime views. Other accounts are known for questionable promotion tactics by usually advertising for commercial purposes or being hired to raise hashtags on the trend list. These accounts contributed to the hashtag by 300-600 tweets per account[30] .
AFH Quantitative Findings on the Scale of MDM in Egypt
AFH is building the largest database of Arabic fact-checked posts covering the MENA region. The primary goal is to fill a crucial research gap in terms of having an accessible and updated database of high-quality fact-checked posts in Arabic. The value proposition of this database is:
Built by Fact Checkers
Unlike many published datasets, AFH does not scrape data and train people to further annotate it. Our database is populated and curated by fact checkers from various organizations who add domain-expertise and richer insights.
Serves Multiple Stakeholders
Currently published data mainly serves researchers who wish to build machine learning models. Ours target additional audiences like media and policy researchers.
Dialects and Multimedia
The database we will be building covers a wide array of data types and formats as well as dialects. Published datasets generally tend to cover modern standard Arabic content per se, and textual content. We will go beyond that.
Continuously Updated
Most published datasets are not maintained as they were used for the research that they were involved in. Our database will be continuously updated to always keep up with changes in the MDM ecosystem.
AFH is still in the process of populating this database, but for the purpose of this report, we decided to rely on the data we collected from our partner: Matsada2sh, a fact-checking organization that focuses on debunking wide-spread claims related to Egypt, mainly. We made the decision of merely analyzing Matsada2sh data because we obtained all their data with no gaps for the time period the report aims to cover, through an API integration with Matsada2sh site. We understand that this might make our findings skewed towards Egypt and from the point of view of Matsada2sh, but we tried our best to crosscheck the emerging trends and findings we detected with other MENA countries to reach better conclusions. Also, this report is the first publication in a series of reports that will continue to analyze more data with higher coverage and depth. Furthermore, Egypt has a unique and mature information ecosystem that witnessed various dynamics of information disorder, making it exemplary of what is actually (and could) happen in other MENA countries. Please note that the dataset used for this study can be shared with researchers upon request.
We analyzed 1,642 fact-checked posts that span 2021, 2022, and until the end of the third quarter (Q3) of 2023. The number of posts is shown below in Table 1. By the end of Q3 for the years 2021 until 2023, the year 2021 had 465 posts, the year 2022 had 427 posts, and the year 2023 had 485. This shows that although the data is until Q3 for 2023, the number of fact-checked posts in this year exceeded the previous years at the same quarter, which can indicate that the scale of MDM is on the rise. Note that Matsada2sh maintained a consistent fact-checking capacity across these years, hence, the increase is not attributed to any major increase in Matsada2sh operations.
YEAR |
Count of posts |
2021 |
586 |
2022 |
571 |
2023 |
485 |
Grand Total |
1,642 |
These posts are related to different topics as shown in Figure 1. The change in topics distribution over the years highlights major shifts in targeted topics, which can be attributed to many factors like the nature of events that took place in these periods, or the type of MDM tactics being used which could be more effective in certain topics. For example, the figure shows how in 2023 (Q1 to Q3) the topic “Economy” almost doubled in size compared to previous years. This is also true to a lesser extent for “Science & Tech”.
Figure 1: Count of Matsada2sh posts by topic
To understand the distribution over time, with more granularity as well as better comparability, Figure 2 shows the posts over time covering Q1, Q2, and Q3 only for the three years. The figure does not show any specific seasonality affecting the distribution, but it is noticeable that the year 2023 witnessed certain days where the number of posts increased substantially compared with the same time in previous years. One of the future research reports that we will work on is to overlay the actual events that took place in Egypt over this distribution to identify the correlation between the events and fact-checked posts. Moreover, to identify if the velocity of publishing MDM claims is getting lower over the years, which could be an indicator of how, unfortunately, efficient the MDM campaigns are becoming.
Figure 2: Count of posts covering Q1, Q2, and Q3 only for the three years
The brief analysis of Matsada2sh data definitely indicates an increase in volume and variety of MDM spreading in Egypt. Upon a closer look into the content itself of Matsada2sh posts, we found that traditional narratives under certain topics, for example how the previous presidential period of Mohamed Morsi is demonized, or how the LGBTQ community is targeted, still exist at a slightly increasing pace. However, we found a number of evolving and conflicting narratives under the same topic, like economic and monetary policies. What is unique about these narratives is they are loaded logical fallacies that likely aim to game the public opinion and to contrast a narrative against the other but not with the intention to raise the awareness, but to deceive and polarize the public.
Matsada2sh data showed a surge in fact-checked posts count including fake AI-generated images amid the global hype of ChatGPT, which started to pick up in April/May 2023 as shown in Figure 3. This was very clear, it was a trend that went viral globally and in the MENA. However, Matsada2sh data further shows that this trend of using fake AI-generated content was not sustainable and faded away, which suggests that the MDM ecosystem’s actors in MENA did not yet integrate the Generative AI in their MDM campaigns to a scalable level. This report will later discuss the potential challenges of generative AI on MDM in greater detail.
Figure 3: Google Trend analysis of interest over time in “ChatGPT” term[31]
We found that sports-related fact-checked posts represent the second highest share of posts in identified topics. Our analysis of Matsada2sh’s posts content led us to conclude that sport-related posts are the main vehicle used to gain higher reach and, consequently, financial gains. Rogue page admins and account owners abuse the popularity of sports in Egypt, especially football, to spread MDM and create trends. In certain instances, sport-related posts were used to push the political agenda of the government, like targeting specific football hashtags to publish MDM content aimed at rival countries.
As we moved towards 2023, we noticed a general increase in environment-related topics as part of the fact-checked posts in Matsada2sh. The main driver behind this increase is domestic as well as foreign due to the likely use of MDM campaigns by the Egyptian and Ethiopian governments targeting political and economic issues caused by the Grand Ethiopian Renaissance Dam. The data shows that Egyptian audiences were targeted by these local and foreign MDM campaigns, where competing narratives were pushed. In other cases and at a much lower scale, specific environment-related MDM posts seemed to be pushed by organized groups of climate change deniers and flat earthers. Such content generally echoed global conspiracy theories and recycled viral content from somewhere around the world, with no noticeable innovation or localization.
The numbers reported in this section show a very grim situation. The scale of MDM is on the rise, getting more complex and effective. The role of generative AI, which will be discussed later in this report, is also rubbing salt into the wound. Is it a lost battle? How can someone beat advanced organizations like GlavNIVTs, Russia’s main Scientific Research Computing Center, which dedicates resources to crafting disinformation networks like Fabrika, with its operators claiming that only 1 percent of its fake accounts are detected by social media platforms?[32]
Illustrating the scale of MDM in MENA does not promote a cynical view, and by no means implies the lack of agency of authentic users. There are reports that showed how grassroots narratives prevailed over government-backed ones[33]. Moreover, the recent war on Gaza shows how collective efforts can be self-organized on Twitter, Facebook, TikTok, and others to counter massive disinformation well-funded MDM campaigns. The MDM applied to the recent war on Gaza can inspire a study on its own as a new era has emerged in terms of players, channels, tactics, scale, tools, and so on. This report was being finalized when the war started, so Gaza-related MDM is not covered extensively.
There are various actors who are involved in MDM campaigns in MENA. Some complex campaigns can involve multiple actors where prior coordination exists. In other cases, the dynamics of information diffusion on social media can attract additional, and sometimes unintended, actors to get involved. The nature of actors behind MDM in MENA is attributed to their access to power and resources, not only in the digital space, but the physical space too. Hence, these actors can be state and non-state actors as follows:
This report will cover selected actor types who tend to have the major share in backing MDM campaigns in MENA.
Domestic and foreign governments and governmental organizations:
Facebook/Meta quarterly publishes reports on Influence Operations (IO) and Adversarial Threat reports to track campaigns with Coordinated Inauthentic Behavior (CIB) including their countries of origin. These reports generally confirm that IO are correlated with geopolitical events, and their scale and modus operandi is aligned with very powerful actors who most likely are affiliated to states. To put this in perspective, the threat report on the state of (IO) between 2017-2020 shows MENA as a very active place in terms of number of CIB campaigns originating from this area [35] as shown in the figure below. The CIB activity of these campaigns is sharply increasing towards being covert and advanced with time, as opposed to being overt, as reported in Facebook/Meta’s 2022 Q4 report when analyzing the Russian-origin IO campaigns related to Russia’s war in Ukraine[35].
Figure 4: Global CIB disruptions as reported by Facebook/Meta between 2017-2020
One of the early state-backed CIB IO that targeted MENA countries like Egypt, Yemen, and Sudan was discovered by Reuters in 2018 to have originated from Iran. The Iran-linked actors ran multiple campaigns through a network of 70 websites publishing content to 15 countries in Asia and Africa. A year later, the Citizen Lab uncovered an advanced IO operation led by Iran[36] targeting Saudi Arabia, USA, and Israel and was called “Endless Mayfly.”[37] The unique aspect of this IO is not just the use of a network of inauthentic websites and online personas, but the fact that all published content was systematically deleted after a period of time. Such an ephemeral approach allowed the content to last enough to serve the Iranian interests, but then disappear to cover any tracks. The network included 135 inauthentic articles, 72 domains, 11 personas, one fake organization, and a pro-Iran republishing network. Another Iran-linked network targeting MENA, beside other areas, was dismantled by the US Justice Department in 2020. The network included 92 web domains used in campaigns run by IRGC to spread disinformation while posing as independent media outlets[38].
Russia is an example of a foreign state actor that targeted MENA audiences with disinformation IO for a surprisingly long time. Dr. Nadia Oweidat, assistant professor at Kansas State University, explained how her research on disinformation in the Arab World led her to find evidence that Russia deployed systematic and consistent campaigns on digital mediums since 2010[39]. The evidence includes Russian bots or intelligence officers masquerading as “authentic Egyptian voices that are anti-colonial, anti-American, anti-Western.”
Within MENA itself, there are many documented IO of CIB originating from Arab countries targeting their people or other rivaling countries, following political, economic, cultural, or even sport-related conflicts. A key example comes from Saudi Arabia where an open source investigation by Bellingcat in 2019 identified a network of bots and tactics, including hacking, that were used to manipulate the public opinion of Saudi citizens and neighboring countries. The investigation led to a specific man called Saud Al-Qahtani who then had an official and powerful position in the Saudi government[40]. Such findings are likely the fruits of an institutionalized effort made by Saudi Arabia to build a disinformation apparatus, as uncovered by The New York Times in 2018. The disinformation apparatus recruits hundreds of individuals for an equivalent of $3,000 monthly salary to run the “troll farm” that targets dissidents[41].
Another CIB IO was attributed to Saudi Arabia state media in which hashtags discussing COVID-19 pandemic were used, or abused, to spread disinformation targeting the governments in Qatar, Iran, and Turkey[42]. Some went as far as blaming these governments for actively spreading the virus. Marc Owen Jones’s important book titled “Digital Authoritarianism in the Middle East” provides detailed accounts on many disinformation campaigns in the Arab world, especially those originating from Gulf countries[43].
Surprisingly, some MENA governments who actively practice disinformation and stand behind MDM operations were actually trained and equipped by “democratic” countries. For instance, it was reported that European Union Agency for Law Enforcement (CEPOL) provided extensive training in 2019 to Moroccan police forces in online disinformation tactics, which included creating “sock puppet” accounts, purchasing SIM cards for fake accounts, and using software to run multiple fake identities simultaneously[44]. Similar training was provided by CEPOL in April 2019 to the Algeria’s police force “National Gendarmerie” to apply digital surveillance and disinformation tactics like the using software to track the location of devices, use of sock puppets for open-source research, purchase different sim cards for different accounts, use picture editing tools, and to post frequently and outside of work hours[45]. Similarly, it was reported in 2023 that the forensic technology to counter cybercrimes that was donated by United Nations Office on Drugs and Crime (UNODC) to the Tunisian government was actually used to monitor online activities of political activists and arrest them on charges of conspiring against the state[46].
It took the state actors in MENA including domestic, foreign governments, and governmental organizations a few years to realize the power of social media and how technology penetrated the lives of Arab-speaking people in major MENA countries. When this was understood, the old tactic of divide and conquer quickly manifested in the form of state backed disinformation campaigns on Twitter and Facebook. Among the key goals of such campaigns was polarizing public opinion and manipulating hashtags that critique governments as a way to diffuse the public tension from accumulating to physical uprising[47]. MENA governments doubled down their disinformation strategy on social media and we witnessed how some organic hashtags and community-driven campaigns were hijacked and poisoned by state-controlled bots and troll armies. We even witnessed disinformation campaigns that were completely manufactured and amplified by troll and bots, creating what Marc Owen Jones calls a “pseudo–civil society.” [48]
Since their inception, social media campaigns have offered the MENA governments a cheaper medium to launch their campaigns; a way to automate campaigns with a huge volume of content to repeatedly expose the people to their narrative; a medium more trusted than traditional state-run media outlets; and an efficient space to reach multiple segments of audiences and measure the reach and effectiveness with data-driven insights.[49]
While domestic and foreign governments targeting MENA with disinformation campaigns generally use similar tactics, they differ in their strategies. Clint Watts, a senior fellow at the Foreign Policy Research Institute, describes Iranian disinformation campaigns to be strategic in design but “very hasty in execution”, unlike the more sophisticated campaigns backed by Russia that are carefully executed in alignment with their strategy[50]. Another difference in the Iranian and Russian campaigns is that the former tries to promote its own narrative and politics, whereas the latter tends to inflict polarization and deception in the countries it targets[51].
A key effort in understanding the scale of global disinformation and the countries it originates from was done by the Australian Strategic Policy Institute (ASPI). The institute found that most of the state-linked disinformation on Twitter can be attributed to Russia, Iran, Saudi Arabia, China and Venezuela. In October 2019, Saudi Arabia based actors published 2.3 million tweets, a number that surpassed even China in that year by 20 folds[52]. The institute published insights on Russia-backed disinformation campaigns including those focused on Syria to discredit the White Helmets volunteer group. ASPI maintains worth-checking datasets and interactive dashboards on global disinformation[53].
That being said, MENA governments do not only act on the offense side of information disorder, but on the defense side, too. Knowing that MDM is a double-edged sword, many Arab countries have officially established dedicated authorities to counter information disorder with supportive legislation and laws. This side of the ecosystem was extensively studied in a pioneering research titled “Meeting the Challenges of Information Disorder in the Global South,”[54] which was led by Professor Herman Wasserman with contribution of four media organizations in the Global South: Research ICT Africa (RIA) covering Africa region, InternetLab covering Latin America, LIRNEasia covering Asia, and Arab Reporters for Investigative Journalism (ARIJ) covering the Middle East and North Africa. The research study surveyed and interviewed many actors in MENA countering the information disorder. But it shed light on how sensitive this topic is when it reported that a number of government agencies in Saudi Arabia and Egypt refused to cooperate or grant the research team interviews. Other organizations and individuals in Bahrain, Jordan, Egypt, and Tunisia refrained from completing the questionnaires. One key finding the study reported is that in many Arab countries where freedom of expression tends to be suppressed, the legislation criminalizing the publication of misleading information “has been used to suppress freedom of expression.”
Platforms:
The conundrum that social media platforms face as profit-seeking companies, while simultaneously accountable for maintaining global values like freedom of speech, can be very challenging. They could be successful when deploying features that promote such values, yet they could fail and get sued for their complicity with governments. An example is the allegation made against Twitter by lawyers representing Abdulrahman al-Sadhan who was one of thousands of Saudis whose confidential information was obtained by Saudi agents working at Twitter in 2014 and 2015.[55]
From a MENA perspective, it is controversial whether to consider major social media platforms to be part of the disinformation challenge, but the presence of their regional offices in countries like the UAE raises eyebrows. Afef Abrougui and Mohamad Najem, CEO of SMEX, a non-profit that advocates for and advances human rights in digital spaces across MENA, argue that technology for GCC governments is used as a “lever of power to control dissent and populations,” so they question the presence of big technology companies in GCC and highlight the intertwined relationship between politics and economy that could hinder digital rights in the Arab region[56]. This challenge is further exacerbated by the lack of dynamic regulations that protect the users, including data protection laws, privacy acts, cybercrime laws, and freedom of information.
On a deeper level, social media platforms constantly change their policies and algorithms, sometimes driven by financial gains, and at other times just following the views of their owners who do not necessarily put into consideration the whole situation. A major, and recent, example is X’s new owner, formerly Twitter, who was at the verge of potentially ending online anonymous activism and put the lives of Arab dissidents in danger when he suggested the “authenticating all humans” feature[57], which entails requiring anonymous users to disclose their true identities. Moreover, a DFRLab assessment found that around March 29, 2023, Twitter changed how its algorithms treat Russian, Chinese, and Iranian state media outlets[58]. NPR further confirmed such a finding and indicated that Twitter decided to stop filtering government accounts in these countries[59], and Reuters reported that the “government-funded” label was removed too[60]. These changes led the European Commission to conclude that Twitter under its new ownership heavily contributed in allowing Russian disinformation campaigns about Ukraine to achieve higher reach before the war began[61]. Moreover, lack of transparency of the new policy of allowing suspended accounts on Twitter to appeal suspension and achieve reinstatement, could open the door for malicious users to abuse it[62].
The changes made by the social media platforms do not generally consider the complicated context in various regions including MENA. This is especially apparent in how content related to uprisings and protests in MENA is usually disfavored by the algorithms[63], and other content is highly censored due to political reasons, like the pro-Palestine voices amid the current Israel’s war on Gaza[64]. On that note, the platforms’ political bias could overshadow their existing policies. This manifests in platforms' approach to mitigate compliance risk with powerful governments by submitting to their requests to remove opposing content. A staggering 94% of the flagged content by the Israeli state prosecutor’s office has been removed by Meta and TikTok according to the Israeli agency[65]. It is a legitimate concern to question how major platforms are siding with the aggressor in this war and whether they are cooperative in removing the disinformation content produced by that side or willful blindness is in play. Major platforms should be more transparent about their content moderation policies, as observing their decisions could make them look biased, or even incompetent as connotated by the experience reported by Marc Owen Jones when he informed Twitter about 1,800 accounts involved in spreading disinformation and hate speech in Bahraini hashtags, and how the accounts were only suspended due to their spam-like behavior[66].
The Public (Opposition and Pro-Regime):
The scale of disinformation practiced by state-actors in MENA led some opposition groups, whether political, economic, environmental, or social groups to adopt disinformation tactics to fight back. Although such tactics are similar, this by no means equates the scale of operations between the state and non-state actors, as the former has access to more resources, and tends to be centralized, whereas the latter is more decentralized.
One of the key opposition groups that orchestrated disinformation campaigns in MENA and was thoroughly analyzed is the Muslim Brotherhood (MB). Stanford Internet Observatory, which analyzed many state-backed disinformation campaigns in MENA, closely studied a MB-Linked network that was taken down by Facebook in November 2020[67]. The analysis covered 25 pages, 31 profiles, and 2 Instagram accounts which originated in Egypt, Turkey, and Morocco and targeted Egyptian audiences directly and other MENA countries indirectly. The network was cross-platform with access to YouTube and Telegram channels that have a large followers base. They published disinformation content criticizing Saudi Arabia, Egypt, and UAE while praising Turkey and Qatar.
According to another report, the disinformation campaigns led by MB-linked networks can still target MENA audiences but for an overseas cause. The India-based “The Disinfo Lab” uncovered online disinformation campaigns launched in September 2020, which called for boycotting India for its role in Kashmir conflict, originated from troll factories linked to MB groups in Turkey and Pakistan. The report mapped various NGOs, companies, and alleged fronts that help MB in their media disinformation campaigns[68]. Note that we were not able to verify the credibility of this report, nor the independence of “The Disinfo Lab” from the Indian state, so we clearly ask readers to view this report and the reported networks distrustfully.
The general public in MENA is more decentralized than opposition groups but can form spontaneous and short-term networks that can actively contribute to the disinformation discourse. As any audience, the public can be polarized where some segments become pro-government, while others are against. For more information on the disinformation tactics that can be practiced by the public, researchers from Harvard University and University of Toronto published a study on the three I’s: inauthenticity (use of automation and fabrication tools), inequality (reliance on the impact of influencers), and insecurity (use of intimidation and repression) explaining the key tactics and how they differ from a centralized, i.e. state, entity[69].
Private Companies:
Private companies play a vital role in advancing the disinformation campaigns in MENA. Propaganda-as-a-service is not new, but with the rise of social media, many players entered into this space including traditional PR firms, based in MENA and around the world. The previous section of this report highlighted the scale of MDM operations in MENA, which gives an idea on the size of this industry and spendings. Probably one of the many dark sides of this industry is the lack of accountability of owners and workers who operate such companies. This report will later discuss how online MDM can cause physical harm, but stories like Cambridge Analytica on how online manipulation transcended into undermining democracy[70], and although the firm was dissolved, its executive moved freely to another firm. A detailed investigation report by Wendy Siegelman uncovered that “Project Associates” contracted with “SCL Social Limited” (in September 2017), whose director was Alexander Nix, former CEO of Cambridge Analytica, for $330,000 to launch campaigns on Facebook against Qatar[71]. This lack of accountability and simply playing by the free market rules cannot stop the proliferation of the disinformation economy.
There are two main scenarios on how the private disinformation industry operates in MENA. The first scenario is when a predator country contracts with a foreign company to target Arab-speaking audiences with disinformation. The predator country could be foreign. This is the story of the UK-registered media firm called “Yala News”, which was exposed by the BBC's Disinformation Team to be spreading and mirroring Russian state disinformation targeting MENA audiences[72]. Another variation is when the predator country is within MENA. An example is the team of Israeli contractors who developed and used a bot-management solution called Advanced Impact Media Solutions, or Aims, and offered it through an Israeli company called Demoman International. The software controls thousands of fake profiles on Twitter, LinkedIn, Facebook, Telegram, and other mediums. The Guardian uncovered the operations of this team, and interviewed Tal Hanan, a 50-year-old former Israeli special forces operative, who claims to have manipulated more than 30 elections around the world using hacking and automated disinformation on social media. The Guardian and its reporting partners tracked Aims-linked bot activity across the internet and found that it was behind fake social media campaigns including the United Arab Emirates[73].
The other scenario is when a predator country contracts with a domestic company to target Arab-speaking audiences with disinformation. In 2019, Facebook disabled a network of accounts linked to Egypt and UAE-based firms. The companies were “New Waves” in Egypt and “Newave” in the United Arab Emirates (UAE), which targeted MENA audiences across Egypt, the UAE, and neighboring countries with influence operation campaigns[74]. Other examples of such companies are Elite Media Group in Saudi Arabia, whose activities were detailed by Marc Owen Jone[75]s, and the Union Group in Egypt that has indirect control over disinformation campaigns in Egypt[76].
AFH Research Findings on the Actors of MDM in MENA
Our analyses of Jordan, Lebanon, Yemen, and Egypt by the AFH’s OSINT team has helped us uncover many interesting facts and trends regarding the possible actors behind the MDM campaigns in the MENA.
In Jordan, misinformation campaigns showed how the lack of reasonable regulations on MDM that considers the end-users’ point of view can be ineffective in meeting its very objectives. It can also lead to accusing regulators for politicizing the regulations to limit freedom of expression. Furthermore, the lack of regulatory vision and transparency immediately lost a key stakeholder in the ecosystem: the general public. This dynamic has most likely fueled the incident we studied and caused a reactive action from influential actors upon the leak of private pictures of the crown prince and his new bride, hence leading to orchestrated MDM campaigns. In addition, this might have sent the wrong message that a regulation was truly applied when powerful people in society were affected. It was promoted and lobbied for by the state media pro-government influencers. It is believed that the unwise management of this situation backfired and caused more users to search for the leaked images, which is a classical story in MENA countries: forbidden fruit is the sweetest.
The analysis further showed that some influencers played a misleading role in these campaigns, which reinforced the top-down MDM. Such influencers did not call for certain community values that could help fight disinformation and promote regulations that are in tandem with freedom of expression. Instead, they shared arguments implicitly favoring the elite groups in power, mixed their discourse with conservative tribal customs, calling for going back to society’s roots, so they became, intentionally or unintentionally, part of advancing the disinformation campaigns around the cybercrime law in Jordan. This is “astroturfing” tactic at best and will be further discussed later in this report.
Moving to Lebanon, our analysis on the actors behind the MDM campaign that targeted the LGBTQ community shows that it started with influential figures in Lebanon who have a conservative anti-liberal mindset across different religions, but quickly spread beyond borders, reaching countries like Yemen, Egypt, and the Gulf. Polarized masses with ideological worldviews who are being exposed to identity politics discussions in the West were ready for mobilization in this campaign. This quickly led the MDM campaign in Lebanon targeting LGBTQ community to become a regional trend with involvement of more actors. In the near future, we will develop a research report trying to answer important questions in these campaigns like: what are the MDM campaign dynamics that cause these hashtags to go viral? What are the social, political, economic, and cultural drivers affecting that? Do these hashtags grow organically then they are hijacked or orchestrated by organized groups? Or the opposite? Is it a hybrid of both? Did some influential state and non-state actors contribute to these campaigns in a self-managed or disorganized way, yet it converged to organized coordination overtime? Are there organized groups across MENA that join forces on specific topics?
The story analysis further showed how nuanced and different the discourses are, even those belonging to the same MDM campaign, which corresponds with the unique local context of each country or society. Furthermore, these nuances magnify the different social structures, governance models, and rule of law across MENA countries. Some MDM campaigns liberally called for violent physical actions while some stayed with bullying and harassment on the digital spaces. Also, we found that conservative actors, mainly influencers and political/religious figures, participate in campaigns on these topics to advance existing discourse but in a modernized way that speaks to young generations. It is like piggybacking active discussions to spread certain ideas. This further involves recycling extremists’ “Western” rhetoric and repackaging it for Arabic audiences to give it some legitimacy – that even the West, the propagator of these ideas is suffering. This is usually a slippery slope for spreading MDM, intentionally or unintentionally, because it is often mixed with emotionally charged arguments where logical fallacies are commonly used. In certain topics, as we saw in the LGBTQ MDM campaign in Lebanon, demonizing marginal groups was part of it.
These research findings led us to ask questions that we will focus on in our upcoming studies. For example, what can MDM campaigns tell us about the real biases and imbalances that exist in our physical and digital ecosystems? How does violence cascade in MENA societies and cause cycles of abuse to trickle down? These are very serious questions in MENA countries because the lack of sufficient state action (rule of law and regulations) and lack of evolving societal protocols/practices does not only prevent MDM campaigns from targeting marginalized groups, but it extends to punishing any sympathizers with them. It helps manifest everything wrong in society.
When we analyzed the story that our OSINT team uncovered on Egypt, it was clear that actors in some countries have higher maturity in terms of organizing MDM campaigns than others. It is not related to having access to more technical resources per se, but it includes deeply understanding the behavior of the targeted audiences. This plays a role in making specific MDM campaigns more pervasive than other campaigns, especially within specific topics like gender-based violence. The story of Nayera Ashraf in Egypt showed how certain narratives prevail and how the public focus can be shifted from blaming the perpetrator into blaming the victim. It was noticeable seeing a unique set of actors who claimed to be relatives of the victim and then in a very deceptive and subtle way, tried to exonerate the perpetrator. This is a double-fold challenge for truth-seekers and fact-checkers, as debunking needs to consider the identity of the actor as well as the narrative it publishes. What is more challenging, and is one of the trends that we identified in the MENA’s information disorder ecosystem, is the wide-ranging use of hard to fact-check content like live streams, especially those coming from someone claiming to be close to the victim, or just asking intentionally intrusive questions to distract the audience.
In Egypt, like in Gulf countries, it was very clear how many pages/accounts systematically participated in trending topics to gain more reach, visibility, and eventually financial gains. Such participation did not necessarily include publishing or amplifying MDM content, but the modus operandi of these pages, accounts, and the private-sector companies behind them highlight how they strive to maximize their reach and gains. This begs a crucial question about the business models that drive MDM campaigns into these vicious cycles to gain more reach and financial proceeds. One of the future research objectives of AFH is to study the media business models in MENA that drive MDM campaigns. One starting point is to extend the fascinating “actor mapping” work done by Global Media Registry and its partners in what they called: “Media Ownership Monitor.”[77] which includes MENA countries like Egypt, Lebanon, Morocco, and Tunisia. One of the key findings reported by this project is that the media in MENA is generally owned by small elites who are close to governing regimes. Some unique characteristics distinguish these countries, like how the media monopoly in Lebanon is driven by specific families directly linked to political parties and religious chiefs, and how specific private-sector companies in Egypt own major media outlets and act as merely fronts to the security apparatus or the Egyptian army.
One recurring dynamic that we discovered in Egypt, again like in Gulf countries, was the weaponization of MDM campaigns by political actors to score points against regional rivalries. What we saw is very similar to the examples mentioned in the literature survey of this section, where both regimes and their opposition are involved in initiating MDM campaigns, although at different scales. The involvement strategy of specific MENA countries in the MDM campaigns is very overt, where they have invested in state resources and private sector companies to conduct the campaigns, whereas other countries are likely using covert strategies to push MDM campaigns through informal media arms, opposition (of the targeted country) living in exile, and public relation firms.
The MDM ecosystem in MENA can be very hostile, leaving physical implications on societies. It is very unfortunate to witness how various actors recklessly launch MDM campaigns during political conflicts that drag the public into severe battles. But when governments find a compromise, people are just asked to celebrate and rarely does someone care about the damage and rift caused as a consequence. This is probably telling on how the public is looked at and considered in this information disorder ecosystem. When some watchdog groups try to hold these governments accountable, they are either intimated, arrested, simply ignored, or in some MENA countries, governments would use the law to avoid giving information or be transparent.
Another unfortunate reality is to see some opposition groups acquiring similar traits of their oppressors and gradually becoming a replica of what they originally opposed. Finally, this unregulated landscape opens the doors wide open to entrepreneurs who wish to utilize the attention economy for their gains and interests, as well as media platform owners who change their policies with little regard to how such changes could enable more MDM campaigns.
AFH Quantitative Findings on the Actors of MDM in Egypt
We continued our analysis of Matsada2sh fact-checked posts, spanning 2021, 2022, until the end of the third quarter (Q3) of 2023, but we introduced a new variable called “Plausible/Possible Actor.” This variable aims to identify the probable beneficiary behind launching each of the 1,642 identified MDM campaigns and verified claims by Matsada2sh. As part of their verification process, Matsada2sh focuses on the content itself, whether textual, image, or video, but they also identify, when possible, who took the main role in initiating the claim. This helped us construct this variable by associating the claim initiator to domestic governments, the opposition, organized groups, foreign governments, pages seeking popularity and reach, or simply unidentified. Furthermore, we created a sample of these claims and corroborated them with publicly available sources that fact-checked them and directly linked them to actual actors.
Please note that this variable identifies probable associations and does not provide a decisive adjudication. Hence, we used this constructed variable as an indicator, and the association identification methodology will be enhanced in future research. As a result, and given the nature of the data, and despite our best efforts to cross-check the information, we cannot guarantee the accuracy of all details. We ask anyone interested in this topic to contact us with revision suggestions.
Our analysis shows that the state of Egypt, denoted as “Domestic Gov” in Figure 5 below, is possibly the main actor behind the majority of MDM campaigns in Egypt, especially under the topics of politics and economy. We noticed a 39% increase in the volume of these campaigns until the end of Q3 of 2023, compared to previous years. This number is most likely increasing as we reach the end of Q4. These campaigns are usually amplified by influencers and mouthpieces. This provides a possible indication that state-backed MDM campaigns are on the rise in Egypt.
Figure 5: Matsada2sh fact-checked posts linked to possible actors over the years 2021, 2022, 2023 (Q3)
In addition, we noticed that besides focusing on politics and economy, the government is using societal issues as well as regional conflicts to advance its agenda and score political points against its rivals.
Foreign countries' involvement in MDM campaigns targeting Egypt seem to have reduced in 2023 compared to previous years. This is probably due to the decrease in discussions around the Grand Ethiopian Renaissance Dam (GERD) as well as some key political rapprochement and reconciliations that took place in MENA in 2023.
Pages and accounts that seek to maximize reach and financial gains mainly used sport-related topics amplify or publish MDM content. Our analysis showed that they are the second largest actor in Egypt. However, their involvement seemed to have declined in 2023 compared to previous years, this is possibly due to restrictions enforced by platform owners on sponsored content and tactics that game or manipulate or trick the content ranking algorithms.
On the sport-related MDM campaign, we noticed government’s involvement in specific campaigns that aimed to conduct “sportswashing”, which is using sports to improve or enhance reputations damaged by previous wrongdoing[78]. This is becoming a trend in other MENA countries, like in Gulf countries, but on a global scale[79].
Posts associated with the opposition maintained its volume over the years, but this could slightly increase in Q4 of 2023 especially during the presidential elections in Egypt, which is taking place in December 2023. Most of the opposition’s MDM campaigns that Matsada2sh identified were initiated by opposition TV anchors or YouTube influencers living in exile.
The last actor type we analyzed was the organized groups, which are more active in technology and science topics, where their content mainly focuses on conspiracy theories. Another organized groups that seem to be rising in Egypt and getting more involved in MDM campaigns are the Kemetism[80] and Afrocentrism[81] supporters.
One of the future research topics that AFH will cover is to conduct a comparative analysis between MENA countries on topic-specific MDM campaigns. For example, while analyzing Matsada2sh data, we did not notice in Egypt any pattern of environment-related disinformation campaigns backed by private companies. This is unlike the situation in Iraq where private companies actively get involved in such campaigns to hide their wrongdoings on the environment.
To conclude our analysis, we tried to categorize the drivers behind government-backed MDM campaigns in Egypt and noticed that they usually fall within the following:
On the 13th of August 2023, a private plane landed in Lusaka, the capital of Zambia's international airport boarding from Cairo, the capital of Egypt. What seemed to be a regular business trip unfolded to a rather extremely bizarre scandal involving $5m in cash, fake gold, guns, ammunition, and corrupt individuals linked to the Egyptian authorities[82]. AFH's partner, Matsda2sh[83], played a courageous role in investigating the story and uncovering major details, highlighting the wrongdoings conducted by influential elites in the government. When the government realized its initial response of bluntly denying all the evidence published by Matsda2sh, it mobilized its digital arsenal and implemented various misinformation tactics to absorb the heat of public opinion. This included publishing half-truths, discrediting Matsda2sh, mass reporting its Facebook page, publishing counter hashtags, and demonizing it and its funders. A few days later, harassment and threats transcended the digital space and manifested in arresting Karim Asaad, a member of Matsda2sh editorial team, which caused a global uproar that eventually led to releasing him and admitting certain details by the government regarding the private plane scandal.
This recent story highlights some of the key drivers behind the use of MDM, like advancing financial interests, political gains, and cultural hegemony, as well as perpetuating dictatorship and the status quo of lack of transparency and accountability. As highlighted in previous sections of this study, the MDM actors are of different nature, state and non-state actors, but the state actors remain the most resourceful and the most vicious in restricting the freedom of speech, conducting mass surveillance, and manipulating the public opinion. This is not limited to countries known for restricting press and publications, as measured by Reporters Without Borders’ Press Freedom Index[84] and The Economist Democracy Index[85] (note how MENA countries are ranked among the lowest in 2023). This extends to democratic countries where political polarization and competing for influence led some state actors to be involved in MDM activities. One of the consequences of such a challenge in Europe led the EU to create a European Democracy Action Plan in 2020 that called for protecting the integrity of elections and democratic process from disinformation applied in online campaigning[86].
The World Economic Forum declared in its 2023 Global Risk Report [87]that MDM is a short- and long-term global risk, categorized under "societal" risk. Furthermore, the report considered the MDM risk to be directly interconnected with "erosion of social cohesion" and "cyber insecurity"[88] risks as well as a major driver in the "interstate conflict" risk[89]. Unfortunately, the political and economic gains facilitated by MDM is hindering serious efforts required by resourceful actors, like states, and was clearly highlighted in WEF's 2023 report when it considered the perceptions around preparedness and governance to counter this risk by local governments, businesses, and other stakeholders to range between "highly ineffective" and "ineffective."[90]
Drivers of MDM, like the political interests, evolve as the geopolitics landscape changes. This leads to continuous evolvement of the MDM tactics as well as an increase in the operating model's maturity of the state actors. One manifestation of such a maturity dynamic can be seen in establishing dedicated entities mandated to handle such operations. Reporters Without Borders (also known as RSF) tracked such dedicated actors and published in 2020 a list of what it called: press freedom’s digital predators[91]. Digital predators are companies and government agencies that leverage technology to harass journalists; apply mass censorship; run disinformation campaigns; and spy on journalists and activists to jeopardize people's ability to access authentic news and information.
Some of the MENA digital predators according to RSF's 2020 list are:
Geopolitical disputes and competition drive foreign state actors to weaponize MDM and use it against their rivals. Disinformation crafted by foreign state-actors usually seeks to undermine trust in public institutions[92] and to influence decision-makers in the target state to follow a predetermined set of actions[93]. At the public's psychological level, MDM applied by foreign state actors aims to create a cognitive overload that leads to poor judgements and decision-making as well as exert a continuous influence on people’s reasoning even if it has been corrected.[94]
The prevalence of the MDM drivers is creeping into emerging domains, in what can be called the MDM arms race. One of the most recent domains that was quickly used in MDM is generative AI, specifically the large language models, which enables the AI to eloquently summarize, generate, and understand text sentiment. A very detailed and visionary research of how influence operations can be supported by generative language models to achieve higher automation, scalability, and personalization was published in January 2023 by researchers in Georgetown University, Stanford Internet Observatory, OpenAI and others[95]. The research identifies a wide range of actions to mitigate the potential risks of generative AI in influence operations and calls for the collective action of many actors including governments, AI developers, technology platforms, and more.
Advancing financial interests, political gains, and cultural hegemony, as well as perpetuating dictatorship, the status quo, and other MDM drivers are intertwined in the very fabric of dominating socio-political and economic power structures in MENA and various parts of the world. This makes it very challenging to mitigate the ensuing risks and consequences by one actor. It calls for a collective action that includes local and global governments, businesses, civil societies, and citizens.
The range of MDM and Influence Operations (IO) tactics applied around the world is constantly evolving. Some of them are covert while the others are overt. With advancements in technologies, the tactics are becoming more sophisticated, like the use of deep fakes; more nested, like orchestrating a multi-level attack; and alarmingly, more accessible to small players. It is not only governments or entities with possession of large resources that can carry out an advanced attack, but small groups can execute a harmful campaign using the right tools and approach.
One of the latest take downs of a sophisticated MDM campaign was announced by Meta on 29th of August 2023, in which China's “Spamouflage” operation was dismantled[96]. Thousands of Twitter, Facebook, and Instagram accounts spammed users and hashtags with politically-opinionated content including videos and pictures as a camouflage. Such massive spam behavior is not new to MENA. For example, in October 2020, the Stanford Internet Observatory uncovered various Twitter accounts that used to automatically tweet verses of the Quran prior to 2019 and applied follow-back rings for years to build an audience base. But at the beginning of that year, they deleted their tweets and started to change their usernames to names of known Qatari dissidents and engaged in a political spam behavior.[97]
China’s Spamouflage operation involved setting up off-platform websites, which receives the traffic from the spam on social media[98]. This resembles Russia's Doppelganger campaign against European countries[99]. However, China’s campaign covered the Middle East where various websites claiming to be independent posted about biased politics.
The technique of using cross-platform publishing that targets MENA users and setting up fake websites is a recurring theme. In some cases, it could include having fake journalists who are either non-existent (use AI-generated photos) or turn out to be fraudsters, like the story of Liliane Khalil[100]. Another variant is silently hacking a legitimate website and modifying its content, or using what is called “typosquatting”[101] to take advantage of common mistakes like setting up a fake website at aljazeerah.net, where the real site is aljazeera.net. Another advanced variant is what Marc Owen Jones calls “hijacking the narrative” where domains and websites are created with the names of rivals, like the site hizbullahnewsagency.com was created by rival countries in the Gulf[102]. Traditional opposition actors in MENA adopt similar tactics, too. In November 2020, Stanford Internet Observatory uncovered a Muslim Brotherhood-linked information operation and a Facebook network originating from Iran and Afghanistan, which used a complex cross-platform operation with fake pages and accounts to advance its reach. [103]
A very common tactic is to direct troll armies at opposition figures, and the MENA region has witnessed many cases similar to this. A study by Brookings Institute explained the rationale behind this indicating that “such constant harassment makes online life miserable for these individuals, wearing them down through constant abuse. It also deters ordinary people from following them, as they seek to avoid the cesspool of recrimination and abuse associated with the critics’ accounts.”[104] Another approach is to make Twitter bots mass follow an influential journalist, like the case of the Yemeni political and information analyst Hisham al-Omeisy, with the goal of undermining the individual’s credibility, the account engagement rate, and eventually leads to account suspension due to the suspicion that the followers were bought.[105]
Hacking is an enabler to MDM/IO in MENA, like hijacking Twitter accounts to spread MDM like in Sudan[106], or directly hacking the phone/laptops of dissidents to potentially hijack their accounts as reported many times by Citizen Lab[107], or even planting an insider in Twitter like the story of former employee Ali Alzabarah[108], who spied on the accounts of Saudi dissidents and others and is now wanted by the FBI. [109]
Hacking can be at the cognitive level, too. Meta’s threat report on the state of influence operations between 2017-2020, described “perception hacking” as the threat actors’ endeavors to “capitalize on the public’s fear of IO to create the false perception of widespread manipulation of electoral systems, even if there is no evidence.” [110]
Some tactics deployed by perpetrators in MENA become popular for a while then fade away, but some may come back. In the Gulf, the sockpuppet accounts was a popular MDM tactic, where accounts that used to belong to someone else are deceptively acquired to act as a mouthpiece for a specific state-backed agenda. In some cases, this tactic was accompanied by another one that involves mass reporting (or coordinated abusive reporting) to silence accounts that disclosed these sockpuppet accounts. In February 2023, Facebook/Meta highlighted in its “Quarterly Adversarial Threat Report” the impact that mass reporting can inflict.[111]
By observing MENA’s MDM/IO ecosystem for years and using preliminary research findings through the fact-checked database consolidated by AFH, some frequent tactics in the MDM machine in MENA were identified. Eight tactics will be highlighted below, and more can be found in Appendix A. Note that these tactics are not comprehensive.
Coordinated Inauthentic Behavior (CIB): This behavior is widely applied in MENA where a wide combination of fake, duplicate, stolen, and real accounts are in coordination to achieve certain objectives. The accounts’ behavior is inauthentic when compared with normal users, hence their operation is considered inauthentic. In the Arab world, some campaigns on Twitter are initiated by troll armies formed on ideological or political basis who use private Telegram groups (like some Muqtada al-Sadr followers as uncovered by a published AFH investigation)[112], Facebook groups, or obscure forums (similar to 4Chan) for coordination. Whereas other campaigns are initiated by a single Twitter user publishing certain keywords to trigger armies of bots and fake accounts to start the campaign.
Shooting the messenger: Some campaigns in the Arab world that gained huge tractions focused on the pattern of shooting the messenger. They focused on the harassment, intimidation, and doxing of whistleblowers, activists, or journalists exposing certain wrongdoings of state-actors.
Amplifying bots and Hashtag Poisoning: Twitter is infested with bots that amplify campaigns discourse with retweets, likes, and promoting certain hashtags, and they became the cornerstone of any misleading campaign globally. However, in the Arab world, there are certain nuances making such bots more difficult to detect, as the misinformation actors are becoming more experienced in crafting such bots, let alone the state-level funding usually allocated to achieve this.
Sock puppet accounts: This is one of the emerging patterns in the Arab world in which certain accounts, which look genuine, are rumored to have been controlled by other figures. Analyzing the discourse used by such accounts, even performing linguistic forensics, can uncover certain patterns to detect such accounts.
Catfishing and fake persona accounts: Some campaigns are triggered by fake identities who are suspiciously given some legitimacy and endorsement by other users to target certain victims. For example, a suspicious account claiming to be the relative of a victim who was killed in an honor killing incident, to spread a victim-blaming narrative. Another example, in light of Elon Musk’s new policies in Twitter, is allowing anyone to buy a verified account badge.
Shill or Stooge accounts: These are credible accounts that exonerate or publicly support other accounts spreading misinformation without declaring that they have close affiliation or relationship. This is an emerging pattern in the Arab world and in most cases involves social media influencers endorsing other state-backed accounts.
Astroturfing: Defined in Wikipedia as “the practice of masking the sponsors of a message or organization (e.g., political, advertising, religious or public relations) to make it appear as though it originates from and is supported by grassroots participants.[113]” This pattern is emerging in the Arab world yet very difficult to detect as it requires elaborate analysis for a period of time.
Crafted discussion threads: Some misinformation campaigns use well-crafted discussion threads that highlight opposing arguments but push for a desired conclusion or narrative. Logical fallacies are usually used like the straw man fallacy[114], Storm of tweets and made-up discussion by troll/bot farms to draw out real conversations by real users. Sometimes the sequence of a trend and its initiation, rather than account creation date, can be useful in detecting this tactic.[115]
To view more MDM/IO tactics that are in use in MENA and around the world, refer to Appendix A which was compiled by the authors of this report. We find some of these tactics very interesting like ephemeral disinformation[116], in which all the published content is systematically deleted after the campaign to evade detection. Another tactic is challenging and interrogating the platforms’ algorithms to manipulate how they work. We believe that the sophistication level of these tactics will keep on the rise, and will be a “fusion” of tactics, where campaigns apply a mixture of tactics supported by high levels of automation and orchestration.
On the emerging MDM tactics that the MENA region witnessed recently, we noticed two tactics that took place during the current Israeli aggression against Gaza in October 2023 (note that this report was being finalized when the war started, so Gaza-related MDM is not covered extensively). The first tactic is disinformation through mistranslation. This was committed by a prestigious media outlet, the BBC, and was uncovered by Respond Crisis Translation. On November 25, a released Palestinian prisoner from Israeli prison appeared in a video posted by BBC News, and she was describing horrendous abuses she faced. The released prisoner said in Arabic: “Israel held us in the cold without electricity, sprayed us with pepper spray and left us to die." However, according to Respond Crisis Translation, the BBC translated this to “No one helped us. Only Hamas cared. ... We love them very much.".[117]= The BBC corrected the translation later, but when it did, the video had already been watched over 12 million times.
The second tactic is an innovation that can be potentially abused for MDM. As a response to Facebook/Meta’s systemic censorship of pro-Palestine content on Instagram and Facebook[118], some activists invented a way to evade censorship, in which the real news content was mixed deliberately with sarcastic or unrelated visual content as shown in Figure 6. This interesting approach to evade censorship begs the question on possibilities of misusing it to intentionally publish disinformation content.
Figure 6: The image caption reads “Here is a recipe to prepare coffee with creme, ingredients are within the image. Share it with your friends if you liked it”, however the content in the image is about Israeli security forces’ aggressive activities in Al-Aqsa Mosque and the occupied East Jerusalem’s Sheikh Jarrah neighborhood.
To conclude with this section, we further analyzed the Matsada2sh data that we have access to through our database to look for interesting tactics used in Egypt. We found that many of the aforementioned tactics in this section have been used over the years. But what stood out was the heavy use of the “deceiving with statistics” tactic. It is not uncommon to find MDM campaigns in Egypt initiated by a government officer using statistics to back his/her claim to sound credible. However, many of these claims are not true, and the irony is that fact-checkers like Matsada2sh usually debunk such claims using the government data itself, especially datasets published by the Central Agency for Public Mobilization and Statistics[119]. This tactic is sometimes blended with other tactics like publishing and amplifying the claims through verified accounts, creating the impression that the claim is credible and exploiting the existing perception that verified accounts were actually assessed by Twitter for authenticity and notability[120]. However, this changed in X and the verification badge can be acquired through premium subscriptions by virtually anyone[121]. Another notable tactic that goes along with “lying with statistics” is to initiate the misleading content by citing a foreign media outlet, research center, author, a personal friend living in the West, or a source who cannot be disclosed (i.e. source protection), without naming any of them. This passive framing of the information source aims to legitimize the information itself and makes it harder to verify especially when blended with half-truths.
As part of actuating its vision, AFH constantly observes the fact checking ecosystem in MENA and gets involved in discussions with its partners and key institutions to detect emerging patterns. AFH recently conducted training sessions after assessing the needs of its partners through a registration form[123], and we noticed that mature fact checking organizations have made progress in expanding the use of fact checking and Open Source Intelligence Tools (OSINT) as part of their day-to-day investigations to cover domains like geospatial analysis using satellite imagery. Also, various fact checking organizations indicated that they are looking to use paid tools to analyze larger content (bypass the API limits) or leverage advanced features in the tools to identify deeper insights during the investigation. The list below shows some of the tools being used by some of our partners who responded to our survey, highlighting the range of domains and tools used by MENA fact checkers:
Throughout some discussions with our partners, we noticed that leading fact checking organizations in MENA are steadily moving towards institutionalization and increasing the maturity of their operating model. This includes identifying strategic initiatives and building internal capabilities in fields like data analytics, data management, and AI. A very positive progress on this track was announced by the Arab Reporters for Investigative Journalism (ARIJ) that launched its AI Strategy for Small & Medium size Newsrooms[124] in July 2023. This is exemplary work in the fact checking ecosystem in MENA, and each organization, including AFH, should celebrate this step and consider designing its own AI strategy.
One of the key emerging patterns that is affecting the MENA fact checking ecosystem as a whole is driven by the platform providers. The desire of social media platforms like Facebook and Twitter/X to maximize their revenues affected the tools used in standard fact checking investigation. Two notable examples are the limits enforced by Twitter on its free API[125], which removed many free tools, and the expected closure of CrowdTangle, the social media monitoring tool by Meta[126]. Furthermore, some shortsighted changes that can be enforced by platform owners without considering the situation that fact checkers in regions like MENA face can jeopardize their lives. An example is the “Authenticate All Humans'' proposed by Elon Musk shortly after acquiring Twitter[127], which could potentially harm journalists and activists in MENA. This is different from opening regional offices in MENA countries that have a history of targeting journalists and fact checkers. It is changing the whole platform to disfavor these groups. The proposed change was not implemented in Twitter/X, but it is an eye opener on how such decisions can deeply change the fact checking landscape and whether we are ready for it.
The ecosystem is constantly evolving and patterns are emerging. This includes the people, processes, technologies, and governance of the fact checking ecosystem. In addition, it includes the dynamics and relationships among the actors operating within it. The levels of complexities added to the ecosystem will create more challenges, yet open the door for more opportunities.
AFH studied the recurring themes which MDM campaigns in the Arab-speaking world usually target, using its partners’ data and the research it conducted. AFH identified ten key themes as shown in the following list. These themes overlap with globally observed MDM themes, however, the depth and breadth within each theme varies depending on regional context of MENA:
It is widely perceived that exposure to MDM causes psychological harm. But the level of harm varies depending on the type of MDM. For instance, satire/parody content aiming to fool users is less harmful compared to completely fabricated content targeting certain individuals or groups.[130]
It is imperative to develop an understanding that MDM operations are not confined in the digital world – they can also lead to severe physical harm. One of the early accounts in MENA was the story of Hamza Kashgari in 2012, who was targeted by disinformation campaigns that reacted to a tweet he published and considered It an act of blasphemy. Despite the clarification that Hamza provided to explain his tweet, a certain narrative framed him and led to his arrest after receiving mass threats from angry mobs[131]. Covid-19 can also be considered as a global case in which MDM played a role in causing physical harm to vulnerable people and communities. Similarly, "A Genocide Incited on Facebook" was the title of an article published by the New York Times[132] explaining how Myanmar's military used disinformation campaigns on Facebook to justify targeting the country’s mostly Muslim Rohingya minority group. Back to MENA, and specifically to Iraq, the years 2022 and 2023 witnessed a number of cases highlighting the dangers faced by independent journalists and environmentalists who are exposing the wrongdoings of governments, private companies, and special interest groups against the environment. In February 2023, Human Rights Watch stated that environmental activists in Iraq are facing threats and detention by government officials and armed groups[133]. This statement was announced amid the release of the prominent Iraqi environmentalist Jassim al-Asadi, following his two-week abduction by an unidentified armed group. A few months later, Ali Hussein Julood died of cancer due to exposure to pollution caused by British Petroleum BP’s oil fields in Iraq[134]. The BBC covered the story prior to Ali’s death in a documentary in 2022[135], where many Iraqi activists and journalists who participated remained anonymous fearing for their lives. Reading into the details of these stories, we can conclude that the common factors across them is the use of climate and environment-related disinformation campaigns on social media and traditional media outlets to discredit the journalists and activists. This online MDM was a proxy to justify potential physical harm and to silence the public opinion's denouncement for such targeting.
Physical harm can be on a wider scale sometimes. Societies in the Arab world can turn against each other on certain occasions, whether being provoked by social, political, or other types of incidents. Sectarian-based hashtags, like Sunni vs. Shia, or ethnicity-based ones are full of misleading claims and fake content to widen the rift between each sect. Such disinformation campaigns can be organic in some cases, and driven by others, like governments, in other cases. Sometimes a peaceful event like a football match can turn into a very obscene situation like the match between Algeria and Egypt in 2009, which witnessed large amounts of fake content spread across traditional media outlets and social media sphere[136]. Another example is Jnoud El-Rab, a far-right Christian group in Lebanon, which attacked an LGBTَ-friendly bar in Beirut in August 2023. The violent attack was preceded by various inciting speeches by political figures as well as disinformation campaigns.[137]
Social media platforms are aware of the physical harm that can be facilitated by their technologies, so automated and manual content moderations are in place. However, major platforms are not coping well with languages other than English like Arabic and Persian. Journalist Farnaz Fassihi, who works at the New York Times tweeted on how sexual harassment and death threats are not censored in Twitter due to the limited moderation coverage for Persian language[138]. Negar Mortazavi, a Washington DC based journalist, emphasized that harmful content moderation should support Arabic and Persian languages in a better way and must be context aware[139]. Such distinction between the content and the context is essential because various activists in the Middle East reported that their content is sometimes censored with a very broad justification that does not factor in the events' context in the Middle East.
Social media platform companies are revenue-seeking vehicles and the manual harmful content moderation adds financial burden on their budgets. The lack of incentive to absorb this burden and lack of proper governmental regulations to enforce moderation practices against harmful content complicate the matter. However, there must be a continuous discussion on mitigating such evolving risks by concerned stakeholders including the public.
Throughout this report, the role of technology and AI has been highlighted, like in MDM automation. But the rise of generative AI techniques like Generative Adversarial Networks (GANs)[140], which offered advanced text to image generation and manipulation beside other use cases, has equipped propagandists and digital predators with a new set of tools. For example, the use of fake images and fabricated personas was essential in creating a coordinated network of fake journalists who published articles in websites like Arab Eye and Persia Now, and eventually found a way to get their stories published in known right-wing media outlets, as uncovered by the DailyBeast[141]. A further advanced version of this tactic is known as “narrative laundering” or “information laundering,” used by Russia’s “Inside Syria Media Center” (ISMC) news website, in which the GRU (The Main Directorate of the General Staff of the Armed Forces of the Russian Federation) relied on fake journalist personas and media fronts to produce content that was republished as pundits’ quotes within ideologically aligned websites.[142]
Another major improvement that took the world by storm was ChatGPT and its foundation of Large Language Models (LLMs). It quickly crept into journalism, and the challenge of generating a made up, yet seemingly authentic article has quickly emerged, as reported by The Guardian[143]. The Guardian team explained how they are reacting to this new technology and how they formed a team of journalists and engineers to learn and assess the potential and limits of ChatGPT and LLMs. ChatGPT and new AI-driven technologies are posing further threats to press freedom as much as they bring new opportunities to it. Various stories covered how some famous AI models had internalized gender and racial biases as part of their training process and generated hateful content as a result.
A reference was made in a previous section of this report to a very detailed and visionary research published in January 2023, which studied the emerging threats and mitigations of using generative LLMs in Influence Operations (IO)[144]. The authors indicated that automated text generation, summarization, rewriting, and extraction offered by generative AI would lead to immediate increase in the number of actors entering this space, as the cost of producing content will drastically drop, the scalability of operations will expand, and the accessibility to automated IO operations increases (think of IO/propaganda-as-a-service in the MDM space). Instead of using the infamous copypasta technique[145], auto-generated content in the MDM space can be efficiently personalized and produced in real time. The misleading content will be more authentic, convincing, and harder to detect. The research further explains how propagandists can build adaptable generative models that exploit explicitly deceitful content using a process known as fine-tuning[146]. Although a number of powerful pre-trained generative language models are publicly available, most of the recent models are private, and some of them can be potentially leaked to the public by insiders: for example, the recent story of Meta’s LLaMa model being leaked on Github and Torrent sites[147]. Propagandists do not require huge computational power to train and build their models. Prompt Engineering and fine-tuning for available or leaked models can adjust the behavior of these powerful models to generate misleading content with affordable computational powers. Furthermore, these models can be fine-tuned and optimized for languages and dialects.
The aforementioned research suggested lists of mitigations across different stages of the generative model creation: model design and construction, model access, content dissemination, and belief formation and called governments, AI developers, technology platforms, and researchers to implement these mitigation measures[148]. Reflecting on how these advancements can hinder press freedom in the Arab world, we should call for mitigation methods covering the end-to-end process of generative language model-enabled disinformation campaigns. The feasibility of each mitigation method should be assessed from the perspective of accountable decision makers: for example, some mitigation would be for AI model providers to impose strict usage restrictions, whereas journalists to use fact-checking specific methodology, while governments impose access controls on AI hardware, and simultaneously, the social media platform providers coordinate to identify AI generated content.
The generative AI wave is evolving and growing, and a wise approach by fact checkers and journalists is to embrace it and devise ways to integrate it in the daily operations of media outlets. A notable effort to collect AI-driven tools that can be used by media outlets was done by Partnership on AI nonprofit in their compiled database of AI tools for local newsrooms[149]. This rapid advancements in generative language models would most likely proliferate digital repression including: disinformation campaigns, digital misogyny, harassment campaigns and defamation against journalists as well as undermine digital civic spaces and marginalized groups not only by domestic/foreign states but by actors of various types. This decentralizes powerful technologies and puts them at the disposal of everyone. While this may implicate an abundant future of opportunities, it is crucial to mitigate the risks of abusing these technologies given the preexisting polarization and rifts in our societies. There is no silver bullet to mitigate the risks of generative AI in MDM. Instead, orchestrated mitigation actions by governments, journalists, AI researchers, civil society, companies, organizations, and citizens must be designed.
As part of understanding the trends in MDM and AI in MENA, the authors of this report mapped key players at the intersection of AI and media in the MENA region who can help counter information disorder. The selection criteria for generating this landscape is simple: to include any organization or company that uses AI in media-centric use cases and supports the Arabic language. Needless to say, this landscape is by no means comprehensive, but it represents a good starting point for those interested to understand this ecosystem in MENA.
This mapping leveraged a previous project done by K4All organization (Knowledge for All)[150], a project funded partially by IDRC[151], which mapped talents, players, knowledge and co-creation hot spots in AI across four global regions, including MENA. The mapping shared in this report is more current and focuses on active AI and media research in MENA. Universities that merely offer academic degrees on AI or companies that do not offer use cases related to media were excluded. The image below presents the ecosystem landscape. Refer to Appendix B for the full list.
Figure 7: Key players at the intersection of AI and media in the MENA region
In addition to the landscape, there is a remarkable effort done by the Digital Innovations for Peace (DIP) initiative where more than 150 open-source solutions and resources on information disorder and Media and Information Literacy (MIL) were collected in a directory[152]. This directory is constantly being updated (currently 167 projects) and it lists key projects in MENA and the globe that combat MDM and hate speech.
Furthermore, there are very useful directories of Arab-speaking individuals working in AI research around the world, like Egyptians in AI Research[153] and Moroccans in AI Research.[154]
The US Cybersecurity and Infrastructure Security Agency (CISA) published a useful guide on key disinformation tactics, with examples and mitigation actions.[156]
# |
Name |
Type |
Country |
Category |
Details |
Link |
1 |
AILLA The AI in Language Learning & Assessment Lab |
Research Group - NFP |
Global |
Publications |
Develops language learning and assessment applications in fields like text readability, AI writing assistant, Multimodal language learning... |
|
2 |
arbml |
Research Group - NFP |
Global |
Tools|Datasets |
Community of researchers working on Arabic NLP research and development |
|
3 |
Research Group - NFP |
Saudi Arabia |
Knowledge Sharing|Raising Awareness |
Offering courses and articles to teach AI in Arabic |
||
4 |
Arabic Speech |
Research Group - NFP |
Global |
Speech Technologies |
A community that runs for the benefit of Arabic Speech Science and Speech Technologies |
|
5 |
Iwan |
Research Group - NFP |
Saudi Arabia |
Publications |
Research group in King Saud University, focusing on NLP research |
|
6 |
Mada |
Nonprofit - Private Partnerships |
Qatar |
Raising Awareness |
Promoting digital inclusion and building a technology-based community that meets the needs of persons with disabilities (PWDs) in Qatar |
|
7 |
KACST |
Governmental Organization |
Saudi Arabia |
Publications|Tools|Datasets |
Publishing papers, datasets, and tools for Arabic |
|
8 |
bigIR |
Governmental Organization |
Qatar |
Publications|Tools|Datasets |
Publishing papers, datasets, and tools for Arabic |
|
9 |
QCRI |
Governmental Organization |
Qatar |
Publications|Tools|Datasets |
Nonprofit multidisciplinary computing research institute founded by the Qatar Foundation, but later became part of KBHU |
|
10 |
CAMeL Lab |
Research Group - NFP |
UAE |
Publications|Tools|Datasets |
The Computational Approaches to Modeling Language (CAMeL) Lab is a research lab |
|
11 |
Nmthgiat |
Research Group - NFP |
Saudi Arabia |
Knowledge Sharing|Raising Awareness |
Promote AI and NLP technologies in the Arab world |
|
12 |
CAMEL-AI |
Research Group - NFP |
Global |
Publications|Tools|Datasets |
Open-source community dedicated to the study of autonomous and communicative agents |
|
13 |
The AI Initiative (AII) |
Governmental Organization |
Saudi Arabia |
Publications|Tools|Datasets |
At KAUST focuses on research excellence, training the future generation of experts, and translation/innovation |
|
14 |
Asas Lab |
Research Group - NFP |
Saudi Arabia |
Publications|Tools|Datasets |
Building a safe and beneficial AI for the MENA region is our mission. |
|
15 |
Dalil Mena |
Nonprofit - Private Partnerships |
Lebanon and Jordan |
Tools |
Streamlines the fact-checking process, using efficient, collaborative and AI-powered workflows. Arij, AFCN, and Siren Analytics |
|
16 |
RDI |
Private Institute |
Egypt |
Tools |
Various Arabic NLP products useful for journalists like Kateb.ai |
|
17 |
Farasa |
Governmental Organization |
Qatar |
Tools |
Farasa is the state-of-the-art full-stack package to deal with Arabic Language Processing |
|
18 |
Taha Zerrouki Projects |
Research Group - NFP |
Algeria |
Tools |
Various tools like Mishkal and Qutrub |
|
19 |
WebRadar |
Private Institute |
Tunis |
Tools |
Online market and media research |
|
20 |
Widebot |
Private Institute |
Egypt |
Tools |
AI-driven chatbots |
|
21 |
Agolo |
Private Institute |
USA |
Tools |
Enterprise-Grade Entity Intelligence (not sure how much they support Arabic?) |
|
22 |
Lisan AI |
Private Institute |
UAE |
Tools |
Writing assistance for Arabic |
|
23 |
Trafalgar AI |
Private Institute |
UK |
Tools |
Developed some critical technology around hate speech, stereotype management and digital slander attacks |
|
24 |
Uktob |
Private Institute |
Saudi Arabia |
Tools |
Generative AI content |
|
25 |
Fenek AI |
Private Institute |
Qatar |
Tools |
Automatic Transcription Platform for Dialectal Languages |
|
26 |
Trint |
Private Institute |
UK |
Tools |
Content creation with AI transcription and live collaboration tools |
|
27 |
Qalam |
Private Institute |
Saudi Arabia |
Tools |
Qalam is a spin-off company from Mawdoo3.com Group and generative AI |
|
28 |
aiXplain |
Private Institute |
USA |
Tools |
aiXplain helps you create and maintain AI systems easily. You can design your own AI pipeline |