Across
- 3. The moderation of AI-generated content through outsourced human classification.
- 4. Carbon-based energy sources sustaining most contemporary computing infrastructures.
- 8. Specialised computational capacity, chiefly GPUs, required for machine learning.
- 13. Extracted minerals and metals essential to digital hardware production.
- 14. A vital cooling resource increasingly strained by data-centre expansion.
- 15. The shift of AI from analysing communication to generating cultural expression.
- 16. Media and policy narratives that shape how AI is feared, trusted, or regulated.
- 18. Excess thermal output from servers requiring intensive cooling systems.
- 19. Human micro-labour used to refine, score, and personalise AI model outputs.
Down
- 1. Discarded digital hardware accumulating as toxic and poorly regulated e-waste.
- 2. Global extractive networks supplying resources for chips, batteries, and servers.
- 4. The long-term geological trace left by digital infrastructure and electronic waste.
- 5. Large-scale collections of digital material extracted to train machine-learning models.
- 6. Planetary-scale data-centre infrastructures enabling large AI models to function.
- 7. Statistical systems that synthesise new content from learned patterns in data.
- 9. Critical cartography that exposes hidden power relations behind AI technologies.
- 10. The exponential growth of data, computation, and capital in AI development.
- 11. AI systems that reproduce cultural forms through probabilistic mimicry.
- 12. Electricity consumption required to power and cool AI data centres.
- 15. Venture-backed companies commercialising generative models and coordinating global infrastructures.
- 17. Market mechanisms that compensate emissions without reducing industrial growth.
