I don,t think anyone's tested it as such, but it's worth noting that no one has reported any ingestion problems. That being said it's likely that the majority of the content from the biggest source that might be affected (I.e. Wikipedia, Fandom and other large wikis) was compiled separately rather than crawled as such.
Is this prompted by the investigation from i think last fall (i don't remember who did it) that suggested that part of the HCU disaster was caused by such ToCs triggering duplicate content penalties?