Well actually with theoretical physicists because they particularly deal with statistical methods rather than tangibles, are _especially_ drawn to information theory.
Consider where the shift happened from “energy is neither created nor destroyed” and became “information cannot be destroyed”? Theoretical physics.
Theoretical physics is precisely the source of the notion.
“Information (data) is the base unit. There can be nothing smaller than the base unit. Therefore, everything is made of the base unit.”
If you were to do something like, say, try to get “meta” with the information source and start talking about semantics and meaning, social constructs of knowledge and such, the ultimate _human_ basis for all of these systems, complete with all of the flaws that goes with that – well, there wouldn’t be any room for it in a system with information-as-base-unit.
That’s the issue with it ultimately.
That doesn’t invalidate its usefulness on a pragmatic level, but it does show where its limits are.
Oh you’re talking quantum information theory. I think in qubit computer logic, they’re primarily matrix transforms, which definitely gives them more flexibility than the standard binary.
I’d like to see work done on qutrits, but as we’ve barely bothered with ternary logic, I don’t see much benefit to jumping to qutrits ’til we’ve mastered ternary.
In a way, qubits are easier to work with than ternary logic because the probabilities allows some assertions of “trueness and falseness” and maintaining the ability to negotiate relative truth values that are interdependent quantities in an entangled state.
AH ok, the “This is the size of your brain” stuff. How many states are possible. Ok. How big is the checkerboard for us to move our pieces on.
I mean, that’s certainly useful stuff but it’s like a gigantic piece of paper. If I know how big my paper is and how wide my pencil tip is, it still won’t let me draw anything on it and really, pragmatically speaking, we can do more with less. Limits are useful but like an unplugged computer, doesn’t do much.
About two years ago, I did a big search for ternary chips in use. There’s some project, some practical things – and I remember a soviet computer that was ternary – probably that same one you’re talking about… but the closest things I’ve found to ternary logic chips are some analog computing tech that managed to survive (in chip form) because their functionality couldn’t be reproduced as well (or at all) in digital equivalents. But generally, there’s not much out there. I have some work on it saved somewhere… I really expected to see more.
Oh LISP machines … man they had such potential. I wanted to learn LISP too. The professor suggested I go through the proper steps (this is 1990) – start with Pascal and work up but I wanted to jump right into AI.
I took a course in Pascal but it was like, “eh, this is like Basic” and was kinda a waste of time, although I did write a few things in Turbo Pascal in the early 90’s and distributed on BBS’, so not a total loss.
The base pairings are interesting in the rule-sets they create. it’s hard _not to_ compare our gene sequences with some kind of computer code because the similarities are so profound.
My fascination with ternary logic though, is that it _can’t_ be mapped to binary. Aristotle’s Excluded Middle has to be included tongue emoticon Considering how influential his black + white thinking has been to, I dunno, ALL of Western Civilization, it’s no surprise that we have trouble thinking in such terms. The very development of our languages themselves have it built into it, the source of dichotomies, paradoxes and the like.
I’m not saying it’s a bad thing: So much has been accomplished with ONLY THIS NEVER THAT NOTHING INBETWEEN and variations of it but I know when we start being able to think in terms of variations of “I don’t knows” that aren’t necessarily tied into probabilities (which I see as kinda primitive: “On a scale of 0-100, how close is it to X?” – probability is linear at its root.
We can think in terms of “two at once” – the either/or… but once we can three distinctive somethings, things get wonky for us because we’ve never been trained in navigating relativities, uncertainties, unbalanced triads and such.
I found something somewhere about the logic used in the prefrontal cortex being ternary in nature.. which makes sense – and they’ve mapped the circuits (so to speak). I should find it. It was pretty fascinating to see them under magnification. and being traced out.
True, although then wouldn’t it require another level of processing for the “I don’t knows”, like a database or something?
Ah, see there’s the issue I have with the mapping of 3 — 2: The levels of complexity grows beyond the pure logic itself.
Other layers of abstraction are required.
To me, this points to a fundamental flaw in the usage of binary vs ternary: You can’t just feed ternary into binary and have a logic-machine (as it were) process it all to come up with definitive results. You need to add additional algorithms, user input, database abstractions, semantics, language: in short, the “Uncertain bit” seems to be everything that logic is not.
I suppose I seem to asking the impossible but I don’t think it is:
“SOLVE FOR X=”I DON’T KNOW”