Across
- 2. Unfair prejudice in data that leads to discriminatory outcomes
- 4. The EU AI Act classifies systems based on this (Unacceptable, High, etc.)
- 5. The ability to provide a human-understandable reason for an AI's output
- 8. Legal responsibility for damages caused by a product or system
- 11. The set of rules or instructions followed by a computer to solve a problem
- 13. The obligation to explain, justify, and take responsibility for AI decisions
- 14. The moral principles that govern how AI should be designed and deployed
Down
- 1. The structural concept of human oversight: In, On, or Out
- 2. A system where the internal logic is hidden or too complex to interpret
- 3. The landmark European regulation that categorizes AI by risk levels
- 6. A formal review to check if an AI system complies with safety standards
- 7. The principle of being open about how an AI system is developed and used
- 9. A subset of AI based on neural networks that is often difficult to explain
- 10. The capacity of an AI to act or make decisions without human intervention
- 12. The fuel of AI; if this is biased, the output will be biased too
