Three possibilities come to mind:

Is there an evolutionary purpose?

Does it arise as a consequence of our mental activities, a sort of side effect of our thinking?

Is it given a priori (something we have to think in order to think at all)?

EDIT: Thanks for all the responses! Just one thing I saw come up a few times I’d like to address: a lot of people are asking ‘Why assume this?’ The answer is: it’s purely rhetorical! That said, I’m happy with a well thought-out ‘I dispute the premiss’ answer.

  • @[email protected]
    link
    fedilink
    2
    edit-2
    5 months ago

    I get where you’re going with that analogy. It’s a bit awkward, just because, as you did, you have to stipulate shelter as opposed to the sheltered area, but with that stipulation it does work, and quite well really.

    And as analogies should be, it’s intriguing.

    But…

    My first reaction is that it’s sort of similar to the “consciousness is an illusion” concept in that it appears to just move the problem back one step rather than solve it.

    It seems to me that what you’re describing is the “space” (or maybe "framework would be better) in which consciousness takes place, but not consciousness itself.

    The problem then (as is the problem with the consciousness is an illusion idea) is that that space/framework/whatever is only of note if a consciousness is introduced.

    At the risk of bringing in too many metaphors, it’s akin to the “tree falling in a forest” thought experiment. The tree falling in the forest certainly generates disturbances in the air that, were there ears to hear them, would register as sound. But without ears to hear them, they’re just disturbances in the air. Similarly, it seems to me that the “shelter” that’s apparently intrinsic to the brain is only rightly considered “shelter” if there’s a consciousness to experience it. Without a consciousness to experience it, it’s just a space/framework/whatever.

    Anyway, do you believe there is any ingredient to consciousness other than the physically of the brain?

    I believe that consciousness in and of itself is obviously that.

    I probably should’ve clarified - when I say “consciouness,” I’m referring to the state/process that’s at least one step removed from immediate perception.

    I see a round red thing and recognize it to be food. That’s just perception.

    I also recognize it to be the thing called an “apple” (in English - other languages have other words). I know that they grow on trees and come in many varieties, and I remember the tree in the side yard of the house I grew up in and how the apples were small and yellow and very good, but I had to generally get a ladder to get any apples, since the deer ate the ones close to the ground (and the ones on the ground, which at least meant I didn’t have to worry about cleaning them up), oh yeah and mom had a recipe for raw apple cake and it was delicious, but she bought the apples for that because the ones from the tree were too firm and tangy to bake with… and so on.

    That’s the part that, to me, corresponds with the “shelter” in your analogy.

    But that’s still not consciousness.

    Consciousness is the apparently entirely non-physical “audience” to all of that - the “I” that’s aware of the process as it’s happening.

    For example, it’s not the part that recognizes an apple, or the part that categorizes it as food, or even the part that remembers the apple tree and the cake and feels nostalgia - it’s the part that’s one step removed from all of that - the internal “audience” (of one) that observes that “I” am experiencing all of that.

    And it seems to me that your view accounts for all of those subsidiary things, but doesn’t account for the “audience” - consciousness. Consciousness is distinct from, and at least one step removed from, all of those things.

    And finally (though this has already gone on quite long) -

    I don’t believe that consciouness is a manifestation of some “spark” or “soul” or anything else external. I think it’s really a relatively mundane function of the brain that we simply haven’t come to understand yet (and for as long as “science” remains blinkered by reductive physicalism, likely won’t be able to come to understand). The key, and the thing (to go all the way back) that ties it in with free will, is that I believe that (as I mentioned before) the communication between brain and consciousness is bidirectional - that there’s some mechanism by which conscious thought alone can at least affect if not wholly direct the path along which neurons fire, and likely not only pioneer new paths, but in some way “flag” them, such that the new path is (nominally) properly fitted into the whole.

    And again - thanks. This is some of the most rewarding mental exercise I’ve had in a long time.

    • @[email protected]
      link
      fedilink
      15 months ago

      Haha “Don’t think about it too hard,” I suggested! But there’s some real value in what you’ve said: I do find your idea of an “audience” very helpful and will have to cogitate on it.

      The first thing it makes me think of is a story about Doctor Who that I once made up while trying to sleep: are you familiar with Doctor Who? It’s a sci-fi show like a home-made British Star Trek with more tinfoil and more time travel.

      In my story, if you’ll indulge me, one of the characters briefly enters a super-perceptual state where she sees some kind of invisible entity steal the consciousness out of her friend’s head, squirts someone else’s consciousness in there, and then makes off with the one originally belonging to her friend.

      It’s like the audience from your analogy walked out of one brain and became the audience for another brain. Her friend wakes up, and doesn’t seem to be any different: the brain his conscientiousness inhabits still has all his memories and nobody else’s, so why would she expect him to be any different? But the “audience” observing their friendship hitherto has been spirited away to who-knows-where and this grieves her.

      The middle of the story consists of her trying to track down the entity, and also her trying, with only ambiguous success, to determine any difference in his behaviours or thoughts. She becomes increasingly desperate as she imagines the invisible entity getting further and further away.

      In the end, she tracks down a way to re-enter the state in which she perceived the entity originally. This time, she is able to look over a large group of people, and the punch line, of course, is that there are countless such entities swapping our “audiences” from head to head constantly. How many times in the last minute has she herself had her conscientiousness swapped for somebody else’s?

      It makes me wonder if the ownership between brain and audience is a thing. Is that connection a necessary thing that is missing from our materialistic definition of a consciousness? Maybe that’s the nagging thing that you call a communication channel.

      I’ve got another thought experiment to share which challenges this idea, but I’ve given you far too much to wade through already so I’ll save it.

      • @[email protected]
        link
        fedilink
        15 months ago

        That’s a fascinating concept.

        And yes - though a yank, I know Doctor Who. ;)

        (And this is the point at which I accidentally tapped “Reply” last time through, which is why there’s a deleted post before this one)

        Anyway…

        My first reaction was that it didn’t make sense that a consciousness could find itself attached to (hosted by?) a different mind and just blithely continue on.

        But the more I think about it, the more I think that’s at least reasonable, and possibly even likely.

        A consciousness might be comparable to a highly sophisticated and self-aware frontend. Any range of data or software can be stored and run through it, and when new data or even a new piece of software is introduced, the frontend/consciousness can and will (if it’s working correctly) integrate it with the system, and it can review the data and software it’s overseeing and find flaws and (unless the ego subsystem intervenes) amend or replace it, and so on.

        And viewed that way, and taking into account the likely mechanics of the whole thing, it really is possible and arguably even likely that it would be essentially content-neutral. It would make sense that while the experience of “I the audience” is itself a distinct thing, the specific details - the beliefs and values and memories and such that make it up - are just data pulled from memory, and it could just as easily pull any other data from any other memory (if it had access to it).

        Fascinating…