• GreatDong3000@lemm.ee
      link
      fedilink
      arrow-up
      0
      arrow-down
      1
      ·
      1 month ago

      Man the models can’t store verbatim its training data, the amount of data is turned into a model that is hundreds or thousands of times smaller than the original source data. If it was capable of simply recovering everything that it was trained on this would be some magical compression algorithm and that by itself would be extremely impressive.

        • GreatDong3000@lemm.ee
          link
          fedilink
          arrow-up
          0
          arrow-down
          1
          ·
          edit-2
          1 month ago

          Oh ok, you want to claim this is compressing the entirety of the internet in a model that isn’t even 1 terabyte of data and be unimpressed that is something.

          But it isn’t compression. It is a mathematical fact that neural networks are universal function approximators, this is undisputed, and analytic functions are continuous so to be an analytical function approximator it must be able to fill in the gaps between discrete data points by itself, which necessarily means spiting out data outside of the input distribution, data it has not seen.