Jump to content

Talk:Convolutional neural network

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Aiden304 (talk | contribs) at 21:15, 26 March 2025 (Update Linguistics in the Digital Age assignment details). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
Did You Know
A fact from this article appeared on Wikipedia's Main Page in the "Did you know?" column on December 9, 2013.
The text of the entry was: Did you know ... that convolutional neural networks have achieved performance double that of humans on some image recognition problems?

Inaccurate information about Convolutional layers

Convolutional layers do not do convolutions. They do what is called "Cross correlation" in DSP, which is different than the statistics definition of cross correlation. https://en.wikipedia.org/wiki/Cross-correlation

This article says multiple times that the convolution operation is being done, and it links to the convolution article https://en.wikipedia.org/wiki/Convolution

This is misleading because it does not do this operation linked in the article. It does the operation linked in the cross correlation articles. -AS

Inacurate information: Convolutional models are not regularized versions of fully connected neural networks

In the second paragraph of the introduction, it is mentioned that "CNNs are regularized versions of multilayer perceptions." I think the idea is inaccurate. The entire paragraph describe convolutional models as regularized versions of fully connected models, and I don't think that is a good description. I think the idea of inductive bias would be better then that of regularization to explain convolutions.

I would also suggest merging the section "Definition" into the introduction. The definition section is only two sentences and it feels it would be better placed at the introduction.

Empirical and explicit regularization?

The section Regularization methods has two different subsections: Empirical and Explicit. What do we mean by empirical? And what do we mean by explicit? —Kri (talk) 12:43, 20 November 2023 (UTC)[reply]

Introduction

"only 25 neurons are required to process 5x5-sized tiles". Shouldn't that be "weights" and not "neurons"? Earlier it said "10,000 weights would be required for processing an image sized 100 × 100 pixels". Ulatekh (talk) 15:53, 19 March 2024 (UTC)[reply]

Absolutely, you're right. I was going to ask the same question. 25 weights for each neuron in the second layer from each neuron in the input layer, and all these 25 weights don't vary as the filter is slid across the input. Do you want to make the correction or should I, since the original editor is not responding? Iuvalclejan (talk) 22:47, 25 January 2025 (UTC)[reply]
I made the change. 2600:6C5D:577F:F44E:B9B2:E830:3647:8315 (talk) 14:20, 27 January 2025 (UTC)[reply]

Big picture

Why are convolutional NNs (or networks with several Convolutional layers as opposed to none) more useful especially for images, than networks with only fully connected layers? You mention something about translational equivariance in artificial NNs and in the visual cortex in brains, but this is a property of the neural network, not of its inputs. It's a way to reduce the number of weights per layer, but why isn't it universally useful (for all inputs and all output tasks), and why is it better for images than other ways of reducing the number of weights per layer? Iuvalclejan (talk) 23:50, 25 January 2025 (UTC)[reply]

Wiki Education assignment: Linguistics in the Digital Age

This article is currently the subject of a Wiki Education Foundation-supported course assignment, between 15 January 2025 and 7 May 2025. Further details are available on the course page. Student editor(s): AshlaMaOmao (article contribs).

— Assignment last updated by Aiden304 (talk) 21:15, 26 March 2025 (UTC)[reply]