Tag Archives: Star Trek

Pray for the Robots

You may have heard of the Turing Test for determining what is a true Artificial Intelligence. Though there are many versions of it, the basic version of the Turing Test is that if a human being is unable to distinguish between a machine and a human, then the machine is, for all intents and purposes, intelligent. Of course, there are potential philosophical objections from Descartes onward, and lots of caveats about how such a test might be accurately conducted, for how long the human  observer must not be able to distinguish them, what percentage of observers cannot distinguish them, etc., but the Turing Test acts as a rough-and-ready benchmark. By the way, if you want to know more about AI and philosophy, I would refer you to Damien Williams (“Wolven” for those of you in the nerd world), who thinks more about these things in a day than I do in a year.

Lately I’ve been thinking about the application of the Turing Test in my own life, and the way I (a human, I promise) seem to be failing it. As some of you might know, there is a character in Star Trek Online that is named for me. The back story is simple: I have friends who worked on the game and ran out of inspiration for names, so they used mine. The character looks nothing like me, doesn’t act like me, talk like me, etc. It’s literally just a name over an NPC (Non-Player Character).

Now, aside from affording me a weird teeny bit of fame, I’ve found a small handful of people who believe that the NPC really is me — that when they’re playing the game, I’m somewhere else off in the world controlling that avatar in real time. Now, at first thought, this is a quaint way of thinking about the internet, that every barkeep in World of Warcraft is some bored Blizzard employee typing the same script and selling the same items over and over. In this case, however, the people know me, so they are literally incapable of distinguishing me from an NPC with my name.

So, does that mean that the “Cadet Scott Nokes” has passed the Turing test, and is a true Artificial Intelligence, not only passing itself off as a human, but as a particular human? I don’t think a reasonable person would accept that premise, since “Cadet Scott Nokes” isn’t a particularly sophisticated simulacrum of life — heck, it probably isn’t even the most sophisticated one in Star Trek Online! I think we would chock this up to the problem of a naive observer.

But this then leads to the problem of naive observers: What if a significant percentage of STO players believed that Cadet Scott Nokes is a person? We would then (with some caveats) say that it has passed the Turing Test and is a intelligent. But in the real life case, the human observers also know me, which is the only reason they took any kind of interest in this insignificant NPC. What if the majority of human observers who know me not only believe Cadet Scott Nokes to be a human, but believe it to be me? In this case, we have got two intelligences, but they are indistinguishable. It’s not that I’ve failed the Turing Test, but rather that a human has become indistinguishable from a machine.

For at least for a small number of observers, we are there already. We can just hand-wave them away as “naive observers,” of course, but at what point is that no longer possible? At Turing’s 30% benchmark? Over half? Nearly everyone? But unlike Rick Deckard, I’m unable to point to myself and say, “I am an artificial intelligence.” In fact, so far as I can be certain, I am the only natural intelligence in the world. I know I’m “real” — it’s the rest of you who might be robots. No matter how much you try to persuade me, even if there is absolute consensus on the point from every other observer, I’ll never believe that Cadet Scott Nokes is “real,” and I’m the simulacrum of Scott Nokes. My own experience is too strong a warrant to be defeated by any percentage of consensus.

I’ve tried to be careful about my use of the words “real” and “artificial” here because I’m getting to a theological point.  From the perspective of the Turing Test, intelligence is intelligence is intelligence. The question of it being “artificial” is one of origin, that is to say that it is an artifact results from the artifice of another intelligent agent.  In other words, “artificial intelligence” is a “created intelligence,” and thus is not distinguishable from other created intelligences — i.e., humans.

The truth is, then, that from the perspective of the naive observers who think Cadet Scott Nokes is me, they are then bound by the same moral duties to treat it as they would treat me. If they wouldn’t curse me out, and they believe Cadet Scott Nokes is me, then cursing it out it, from their moral position, the same as cursing me out.

But remember, in the end, the only intelligence I can be certain of is myself, so let’s assume, for the sake of argument, that I am the only “real” intelligence in the world — that everyone else is just an artificial intelligence, and I have naively assumed them to be real. I am just as morally bound to them as I am to the only other intelligence, i.e. me.

Sound familiar? It should: “So in everything, do to others what you would have them do to you, for this sums up the Law and the Prophets” (Matt 7:12). We’ve backed into the moral imperative of the Golden Rule. This isn’t a far off, science-fictiony issue — it’s one we have to account for soon. With naive observers already failing to distinguish between machines and humans, we’re not far from sophisticated observers being unable to distinguish between them. Indeed, we’re not far from you being unable to distinguish between them. From the Christian perspective, the response is pretty obvious: If you believe something to be intelligent, you treat it as your neighbor, until you have sufficient warrant to believe otherwise. The claim of Jesus that the Golden Rule is the essence of the Law and the Prophets means that this is encoded into the entire cosmos.

For non-Christians, then, observers run into a few choices: The most natural and philosophically-defensible is a Nietzschian master-slave morality, but the essential problem with that is we might find ourselves forced to adopt a “slave morality” of subversion to our AI masters. We could go with Utilitarianism, though since an AI could calculate outcomes of “human flourishing” (a category that would presumably also include AI flourishing in this scenario) far better than humans could, we are left with complete dependence on the judgments of our benevolent AI moral judges. It’s not really possible to exhaustively list the potential ethical frameworks and analyze their various benefits and pitfalls, but it’s certainly time for even the layman to start thinking about it.

As for me, I have a prior moral engagement with the Christian framework, so I’ll keep trying to navigate the morality accordingly. I know Cadet Scott Nokes isn’t me, and I have no serious reason to think it is a true intelligence, so you should feel free to treat it as you would any other game NPC. But if you ever log in to Star Trek Online and Cadet Scott Nokes says something like, “Hey, did you read that article the other Scott Nokes wrote about me? It really got me thinking I should pray more,” you might consider being much more thoughtful in the way you treat my namesake.

 

 

 

The Novelty of Geek Rock

Geek rock has frequently been conflated with novelty rock and artists like They Might Be Giants and Weird Al Yankovic have both appeared on Dr. Demento. The question of geek vs. novelty, however, is not a simple one. For me, the conflation of “geek” and “novelty” is largely due to the conflation of “geek” with “smart” with “trivial and exclusive knowledge.” Smartness is magic. Ask any geek the square root of 467,982 and they can answer, magically, off the top of their heads. Ask any geek about your computer, and they will answer in a language that the common person cannot understand. Knowledge, and especially geek knowledge, is esoteric, effortless (to the geek), and  exclusive. Geek knowledge does not appear to be easily accessible, even though anyone can learn math, or computers, or programming, or technology, or literature, or what have you. And it’s knowledge that is trivial in that it is not immediately relevant to a “normal” day. Who needs to know pi? Or phi? Or the date of the first Mercury mission? Or the author of “A Sweet Nosegay?”

Much like magic, geekness became associated with novelty. Or perhaps it always was inherently novel–after all, a “geek” was the dude biting heads of chickens at the carnival. (Which means, of course, that Ozzy? Total geek. Right?) A geek was a sideshow, a freak attraction, a ten-cent spectacle. An amusement. A novelty, like a pulp novel or a comic book or a sci-fi movie or sci-fi tv show.

In other words, like Star Trek.

Not Star Trek now, of course, because Star Trek, with the reboot movies, and TNG, and George Takei’s winning of the Internet is different from the original airing of Star Trek and the decades of marginalization and parody when cosplay was freakish and replica phasers were best kept at home. Now, in the heyday of geekdom, Star Trek is no longer a sideshow, but a mainstream blockbuster phenomenon. Much like the novelty band/song who has been geek rock all along.

You know where I’m going, don’t you?

“Star Trekkin’” is the perfect example of the dynamics between novelty and geek rock. Yes, it’s a parody. A loving parody, to my ears. A silly, joking, homage to the original series. Yes, it’s a novelty; at least, it’s classified as such. But it’s all in the ears of the audience, too. And this audience member hears it as a tongue in cheek homage to a groundbreaking show that, while breaking ground, also forayed into over-acting, Edens and Earth like planets, and stunningly literal rocky terrain. All flaws, perhaps, and worthy of parody, but all lovable flaws nonetheless. So maybe it’s time to reclassify “Star Trekkin’” from a side show, chicken-beheading novelty act to a clever early foray into geek rock (which is, of course, the final frontier).