Human sciences

Manipulation, mistrust and adoption: paradoxical responses to AI in companies

• A study conducted by Natalia Vuori (Aalto University, Finland) has identified four trust configurations (full trust, full distrust, uncomfortable trust, and blind trust) that determine how employees interact with AI.
• Some employees deliberately manipulate AI tools, compromising the accuracy of their results. Deterioration in the reliability of these tools may then create a ‘vicious cycle’ within companies.
• Managers are well-advised to adopt approaches to AI that reflect levels of trust among employees while reassuring them about the use of data and emphasising concrete benefits.
Read the article

“I lost track of time”: how we get caught up in digital applications?

Read the article

Artificial intelligence: how psychology can contribute to AGI

Read the article

A User-Centred Approach to AI Explainability

Read the article

Data and AI Ethics Council, guarantor of responsible AI at Orange

Read the article

Factiverse: reliable AI fact-checking in more than 100 languages

Read the article
A man is crouched on bare ground, holding an object in the air with one hand and a pencil in the other. Next to him, an open laptop suggests he is focused on his outdoor research work.

Geology, geoarchaeology, forensic science: AI reveals history in grains of sand

Read the article