PhilSci Archive

Deep convolutional neural networks are not mechanistic explanations of object recognition

Grujicic, Bojana (2023) Deep convolutional neural networks are not mechanistic explanations of object recognition. [Preprint]

[img]
Preview
Text
DNNs not mechanistic.pdf

Download (4MB) | Preview

Abstract

Given the extent of using deep convolutional neural networks to model the mechanism of object recognition, it becomes important to analyse the evidence of their similarity and the explanatory potential of these models. I focus on one frequent method of their comparison – representational similarity analysis, and I argue, first, that it underdetermines these models as how-actually mechanistic explanations. This happens because different similarity measures in this framework pick out different mechanisms across DCNNs and the brain in order to correspond them, and there is no arbitration between them in terms of relevance for object recognition. Second, the reason similarity measures are underdetermining to a large degree stems from the highly idealised nature of these models, which undermines their status as how-possibly mechanistic explanatory models of object recognition as well. Thus, building models with more theoretical consideration and choosing relevant similarity measures may bring us closer to the goal of reaching a mechanistic explanation.


Export/Citation: EndNote | BibTeX | Dublin Core | ASCII/Text Citation (Chicago) | HTML Citation | OpenURL
Social Networking:
Share |

Item Type: Preprint
Creators:
CreatorsEmailORCID
Grujicic, Bojanabojana.grujicic@maxplanckschools.de0000-0003-2551-5070
Keywords: deep neural networks, explanation, mechanisms, representation, object recognition, similarity measures
Subjects: General Issues > Explanation
General Issues > Models and Idealization
Specific Sciences > Neuroscience
Specific Sciences > Cognitive Science > Perception
General Issues > Technology
Depositing User: Ms Bojana Grujicic
Date Deposited: 06 Oct 2023 03:13
Last Modified: 06 Oct 2023 03:13
Item ID: 22629
Subjects: General Issues > Explanation
General Issues > Models and Idealization
Specific Sciences > Neuroscience
Specific Sciences > Cognitive Science > Perception
General Issues > Technology
Date: 2023
URI: https://philsci-archive.pitt.edu/id/eprint/22629

Monthly Views for the past 3 years

Monthly Downloads for the past 3 years

Plum Analytics

Actions (login required)

View Item View Item