I consider preventing indexation of all of MediaWiki 1.33.0's Special pages.
In Hebrew, מיוחד means "Special":
Disallow: /מיוחד:* Disallow: /index.php?title=מיוחד:* Disallow: /index.php%3Ftitle%3D%D7%9E%D7%99%D7%95%D7%97%D7%93:*
Doing so is good in general because many of these pages aren't useful to the avarage surfer (rather to staff only) but some are important to both a regular users and crawlers.
A problem of not indexing "RecentChanges" and "Categories" is plausible though, as both these Special pages serve as "small, dynamic pseudo-sitemaps" that give access to virtually all webpages in the site.
Would you remove
Disallow of MediaWiki special pages from robots.txt altogether?
Would you keep it with a good list exceptions just for "RecentChanges" and "Categories"?
Would you take a totally different approach?