This is topic Trek computer system incontinuity.. in forum Starships & Technology at Flare Sci-Fi Forums.


To visit this topic, use this URL:
https://flare.solareclipse.net/ultimatebb.php/topic/6/1404.html

Posted by The Red Admiral (Member # 602) on :
 
A minor point, but it's something I picked up recently that bothered me. In 'Distant Origin' the Voth scientists whilst roaming under the cover of a kind of phasing-cloak on the ship inspect a computer interface terminal, and one states "simple binary system", or words to that effect. Yet in 'Future's End', when Janeway uploads Starling's database to Voyager, Kim says on the other end of the comlink that it would take a few moments to reconfigure to the old binary system (in order to interpret the antiquated data format).

So what's going on here, does 24th century computing utilize binary, or not?! For some reason I hadn't noticed this problem before, even though I've viewed these and all other Trek episodes on multiple occasions (although I haven't seen any Enterprise episodes yet - and for spoiler reason will not participate in any threads until I have).

I had originally thought over the years that Daystrom's initial Duotronic computer system used some ultra hi-tech system which replaced binary. Even today we have the theoretical 'quantum computing' which could possibly one day supersede the traditional binary method of plain 1's and 0's.

So, are there any theories....? Or does anyone know of any further references (one way or the other) to binary in Star Trek...?
 


Posted by Sol System (Member # 30) on :
 
These references always bother me. You can't "improve" upon binary code. It isn't the sort of thing you can add features to. It is useful because it is simple. On. Off. The only thing simpler would be a unary code, which I suspect would prove less than useful.
 
Posted by The Red Admiral (Member # 602) on :
 
You can improve on binary. I don't yet know all the ins and outs of 'quantum computing', but this amazingly advanced and complex system, currently in research phases is an alternative to the simple binary system. But the technology isn't quite there yet. You need to be well versed in quantum physics to grasp how the concept works.

A brief description of quantum computing:
http://www.qubit.org/intros/comp/comp.html
 


Posted by TSN (Member # 31) on :
 
Well, I can think of one improvement offhand. Think about this... In binary, to store, say, the number 42, you need to at least store "101010". That's six units of storage ("bits", in this case). In decimal, you only need two, the '4' and the '2'. The reason we use binary is because is works on switches. Each bit can be turned on or off, and the computer can interpret these as '1' and '0'. To store things in decimal, every "bit" could be any digit from 0-9. Simple "on" and "off" states won't cover that. You'd need a way for a single unit of memory to have ten different states.

If we can do that, we can store things in decimal, and it will be improved over binary. Or we can have sixteen states and use hexadecimal. And so on.
 


Posted by The Red Admiral (Member # 602) on :
 
Precisely, that's how quantum computing works, using various states of quanta to simulate 0, 1, and 0 and 1 at the same time. It's very comlicated. I suggest anyone interested read the explanation of quantum computing from the link.
 
Posted by Woodside Kid (Member # 699) on :
 
In TMP Spock made a reference to V'Ger's signal as being "a simple binary code," so apparently binary computing was antiquated even at that point.
 
Posted by Omega (Member # 91) on :
 
Or at least binary data transmission.
 
Posted by The Red Admiral (Member # 602) on :
 
True enough, I'd forgotten that reference, but it's uncertain how much we can read into that as Omega pointed out.
 
Posted by Proteus (Member # 212) on :
 
Dont computers use fuzzylogic now? 1s, 2s, and 0s?
 
Posted by Joshua Bell (Member # 327) on :
 
Quantum computing is actually quite simple.

Contemporary computing works on the basis of instructions causing operations to occur on registers. A register is a storage unit in the processor big enough to hold a single value - contemporary computers use 32 bit registers. Instructions describe simple actions - for example, the instruction "LOADA R5,00a020" might load the contents of memory address 00a020 into Register #5. "ADD R5,R6,R7" might add the values in R5 and R6, and place the results in R7. Given a handful of opcodes (load, store, add, multiply, compare, branch) you've ready to rock.

Quantum computing augments this by making your register out of "qbits" instead of just "bits" and adding an additional "opcode" to the toolkit. Let's call this "QLOAD" - it loads a second value into the same register. Let's say we had a 3-bit register and we LOAD the number 6 (binary: 110) into it; we could draw that as [<1>;<1>;<0>]. Now let's QLOAD the number 2 (binary: 010) into it: [<1,0>;<1,1>;<0,0>]. Okay, so what? It just looks like we have a register that holds twice as much! We could also call this "the superposition of 6 and 2".

Here's where it gets fun. Now let's use the DIV opcode and divide the result by 2: [<0,0>;<1,0>;<1,1>] - see how it was sneaky enough to divide the superposition and get "the superposition of 3 and 1" as an answer?

Wait - why is that a big deal?

Well, a multiplication operation takes time - the computer has to do the same steps to multiply numbers "in sillica" as you do on paper. Compared to that, everything else is fast. But imagine using QLOAD a million times, then doing a single multiply - you'd get a million multiplications for the price of one!

Great, now we just have to get the results out!

Not so fast - quantum operations don't work like that. You see, a superposition is the universe being in a truly "indeterminate" state, somewhat like a coin in mid-flip. By using tools like QLOAD you can very precisely limit the range and probabilities of outcomes, but there's only one answer you can get out. If you try and read the bits of the result from above, you might get 001 or 011 or even 010! Each bit just has an equal probability of being read (or observed) each way!

So what is this good for?

Well, if you're sneaky, you can perform operations that would "bubble" the results you're looking for to the top. When factoring large numbers there are thousands of wrong answers and only one right answer. You can mathematically describe the right answer and perform operations which make it stand out and the others cancel each other out. The same thing applies when doing a search over a database - you can load the entire record set into a quantum register and perform a lookups in a single step.

In computing terms this lets you take enormous shortcuts. Factor a 1024 bit number instantly instead of taking every computer on the planet and several years - great for breaking codes. Nearly any computation doesn't just get orders of magnitude faster, but in some cases, actually now possible within the lifetime of the universe (10^100 years).

(Practically, we've only made 3-qbit registers, using billions of copies of organic molecules that perform computations in a solution; the quantum states we're talking about are so fragile that 99.99..% of the registers flake out, but enough "right" answers stand out above the noise to prove the concept. Doing anything more will take some pretty amazing breakthroughs, but it seems feasible.)

What does that mean for Star Trek?

Not much: again, quantum computing - as we currently understand it - really just adds a couple of "opcodes" to our computer parlance and the promise of making certain operations much faster. Take a database lookup - if you had all of the database keys loaded into a quantum register you might get instantaneous lookups. But loading the register still takes linear time.

Anyway, the point is: when they say "bit" in Star Trek they might very well mean "qbit". But that, in itself, doesn't explain any of the technobabble or potential contradictions.
 


Posted by NightWing (Member # 4) on :
 
I remember something being said about a trinary syntax or something in 'Hope and Fear'.
 
Posted by The Red Admiral (Member # 602) on :
 
'Trinary syntax'? Hmm, I've have to check that one out. Joshua, also in Trek they use the term 'quads', as in quads of data, gigaquads, terraquads and so forth. Although this isn't much of a clue, it's at least something which possibly points to a system other than binary.
 
Posted by TSN (Member # 31) on :
 
In fact, it could stand for "quantum" something, as opposed to being related to "quad" meaning "four"...
 
Posted by The Red Admiral (Member # 602) on :
 
Possibly, but I'm not sure what 'quad' could be in reference to, so far it's just another technobabble line. But I still think this whole issue is a bit paradoxical, because consider the Bynars, remember them? Their entire civilization was based around binary, and they were presumed to be extremely advanced, and comfortably integrated their technology with the Enterprise. That, and the Voth's line from 'Distant Origin' I think suggests Starfleet are using the binary method in computing.

Kim's line I suppose has to be seen as an aberration.
 




© 1999-2024 Charles Capps

Powered by UBB.classic™ 6.7.3