• @[email protected]
    link
    fedilink
    English
    982 days ago

    We don’t think in “bits” at all because our brain functions nothing like a computer. This entire premise is stupid.

    • @[email protected]
      link
      fedilink
      English
      33
      edit-2
      1 day ago

      Bit in this context refers to the [Shannon](https://en.wikipedia.org/wiki/Shannon_(unit\)) from information theory. 1 bit of information (that is, 1 shannon) is the amount of information you receive from observing an event with a 50% chance of occurring. 10 bits would be equivalent to the amount of information learned from observing an event with about a 0.1% chance of occurring. So 10 bits in this context is actually not that small of a number.

      • @[email protected]
        link
        fedilink
        English
        31 day ago

        10 shannons, that is, 10 bits, each with 50% probability would be equivalent to the amount of information gained from observing an event with 1/1024 chance of occurring, not 1/10. Thats because this unit gets combined multiplicatively. The wikipedia article mentions that if there are 8 possible events with equal probability, the information content would be 3 shannons.

      • @[email protected]
        link
        fedilink
        English
        15
        edit-2
        1 day ago

        The paper gives specific numbers for specific contexts, too. It’s a helpful illustration for these concepts:

        A 3x3 Rubik’s cube has 2^65 possible permutations, so the configuration of a Rubik’s cube is about 65 bits of information. The world record for blind solving, where the solver examines the cube, puts on a blindfold, and solves it blindfolded, had someone examining the cube for 5.5 seconds, so the 65 bits were acquired at a rate of 11.8 bits/s.

        Another memory contest has people memorizing strings of binary digits for 5 minutes and trying to recall them. The world record is 1467 digits, exactly 1467 bits, and dividing by 5 minutes or 300 seconds, for a rate of 4.9 bits/s.

        The paper doesn’t talk about how the human brain is more optimized for some tasks over others, and I definitely believe that the human brain’s capacity for visual processing, probably assisted through the preprocessing that happens subconsciously, or the direct perception of visual information, is much more efficient and capable than plain memorization. So I’m still skeptical of the blanket 10-bit rate for all types of thinking, but I can see how they got the number.

        • @[email protected]
          link
          fedilink
          English
          71 day ago

          It’s a fair metric IMO.

          We typically judge super computers in FLOPS, floating-point-operations/sec.

          We don’t take into account any of the compute power required to keep it powered, keep it cool, operate peripherals, etc., even if that is happening in the background. Heck, FLOPs doesn’t even really measure memory, storage, power, number of cores, clock speed, architecture, or any other useful attributes of a computer.

          This is just one metric.

    • @[email protected]
      link
      fedilink
      English
      172 days ago

      I also don’t have 10 fingers. That doesn’t make any sense - my hands are not numbers!

      Ooooor “bits” has a meaning beyond what you assume, but it’s probably just science that’s stupid.

        • @[email protected]
          link
          fedilink
          English
          18
          edit-2
          2 days ago

          You say “we don’t think in bits because our brains function nothing like computers”, but bits aren’t strictly related to computers. Bits are about information. And since our brains are machines that process information, bits are also applicable to those processes.

          To show this, I chose an analogy. We say that people have 10 fingers, yet our hands have nothing to do with numbers. That’s because the concept of “10” is applicable both to math and topics that math can describe, just like “bits” are applicable both to information theory and topics that information theory can describe.

          For the record: I didn’t downvote you, it was a fair question to ask.

          I also thought about a better analogy - imagine someone tells you they measured the temperature of a distant star, and you say “that’s stupid, you can’t get a thermometer to a star and read the measurement, you’d die”, just because you don’t know how one could measure it.

            • @[email protected]
              link
              fedilink
              English
              9
              edit-2
              2 days ago

              Maybe try looking into the topic instead of confidently repeating your wrong assertions? You’re literally pulling a “my hand is not a number!” right now.

              Just because you have a limited understanding of a unit, doesn’t mean that unit is only applicable to what you know. Literally the star example I brought up.

                • @[email protected]
                  link
                  fedilink
                  English
                  2
                  edit-2
                  1 day ago

                  Ah, so you just choose to ignore information you don’t already know? What a rational thing to do. You’re not anti-intellectual at all.

                  Or are you seriously trying to gaslight everyone into believing Shannon entropy doesn’t exist?

    • @[email protected]
      link
      fedilink
      English
      42 days ago

      Also supposing it did, I’m quite sure that everyone’s brain would function at different rates. And how do you even measure those people that don’t have an internal monologue? Seems like there is a lot missing here.

      • @[email protected]
        link
        fedilink
        English
        31 day ago

        It’s an average. The difference between two humans will be much less than the difference between humans and machines.