Sort by
Filtered results
- 4 results found
Sort by
Module B focuses on the risks AI poses for social fairness and trust: how the use of AI-based tools can generate inequality or dishonesty, particularly when human productions differ in nature (e.g. creative vs.
This document examines how AI-driven content curation and recommendation systems affect the quality of public deliberation.
Module C of the Toolkit has two primary objectives: First, to understand AI and big data within the context of a long history of interactions between technological affordances and cultural norms, values, and practices. This recognises that knowledge technologies—such as written language, the printing press, television, radio, etc.—have shaped culture and knowledge production. The relationship between technology and culture is fundamentally mutual and reciprocal. Second, building upon the first objective, Module C focuses on the particular definition of AI and big data as advanced knowledge technologies (AKTs). We analyse the past in this module to better understand the present and—potentially—to anticipate what may lie ahead.
Since our liberal democracies generally employ forms of representativeness to their institutions, the impact of AI on free and fair elections is also one of the key ways in which technology affects our polities.