Discussion about this post

User's avatar
Scott Burson's avatar

"This is why adding more parameters to a Neural Network adds to the cost exponentially."

Nonsense. Training cost is linear -- not even quadratic -- in the parameter count.

I know: nontechnical people throw around the term "exponentially" as if it only meant "a lot". But it has a very specific technical meaning, and we computer scientists should use it only when that meaning is applicable.

In fact, parameter counts themselves have been growing exponentially. Training these large networks would not be possible if the cost were even quadratic, never mind exponential, in the parameter count.

Expand full comment
3 more comments...

No posts