Disclosure (1994)

Our next 3D file browsing system is from the 1994 film Disclosure. Thanks to site reader Patrick H Lauke for the suggestion.

Like Jurassic Park, Disclosure is based on a Michael Crichton novel, although this time without any dinosaurs. (Would-be scriptwriters should compare the relative success of these two films when planning a study program.) The plot of the film is corporate infighting within Digicom, manufacturer of high tech CD-ROM drives—it was the 1990s—and also virtual reality systems. Tom Sanders, executive in charge of the CD-ROM production line, is being set up to take the blame for manufacturing failures that are really the fault of cost-cutting measures by rival executive Meredith Johnson.

The Corridor: Hardware Interface

The virtual reality system is introduced at about 40 minutes, using the narrative device of a product demonstration within the company to explain to the attendees what it does. The scene is nicely done, conveying all the important points we need to know in two minutes. (To be clear, some of the images used here come from a later scene in the film, but it’s the same system in both.)

The process of entangling yourself with the necessary hardware and software is quite distinct from interacting with the VR itself, so let’s discuss these separately, starting with the physical interface.

Tom wearing VR headset and one glove, being scanned. Disclosure (1994)

In Disclosure the virtual reality user wears a headset and one glove, all connected by cables to the computer system. Like most virtual reality systems, the headset is responsible for visual display, audio, and head movement tracking; the glove for hand movement and gesture tracking. 

There are two “laser scanners” on the walls. These are the planar blue lights, which scan the user’s body at startup. After that they track body motion, although since the user still has to wear a glove, the scanners presumably just track approximate body movement and orientation without fine detail.

Lastly, the user stands on a concave hexagonal plate covered in embedded white balls, which allows the user to “walk” on the spot.

Closeup of user standing on curved surface of white balls. Disclosure (1994)

Searching for Evidence

The scene we’re most interested in takes place later in the film, the evening before a vital presentation which will determine Tom’s future. He needs to search the company computer files for evidence against Meredith, but discovers that his normal account has been blocked from access.   He knows though that the virtual reality demonstrator is on display in a nearby hotel suite, and also knows about the demonstrator having unlimited access. He sneaks into the hotel suite to use The Corridor. Tom is under a certain amount of time pressure because a couple of company VIPs and their guests are downstairs in the hotel and might return at any time.

The first step for Tom is to launch the virtual reality system. This is done from an Indy workstation, using the regular Unix command line.

The command line to start the virtual reality system. Disclosure (1994)

Next he moves over to the VR space itself. He puts on the glove but not the headset, presses a key on the keyboard (of the VR computer, not the workstation), and stands still for a moment while he is scanned from top to bottom.

Real world Tom, wearing one VR glove, waits while the scanners map his body. Disclosure (1994)

On the left is the Indy workstation used to start the VR system. In the middle is the external monitor which will, in a moment, show the third person view of the VR user as seen earlier during the product demonstration.

Now that Tom has been scanned into the system, he puts on the headset and enters the virtual space.

The Corridor: Virtual Interface

“The Corridor,” as you’ve no doubt guessed, is a three dimensional file browsing program. It is so named because the user will walk down a corridor in a virtual building, the walls lined with “file cabinets” containing the actual computer files.

Three important aspects of The Corridor were mentioned during the product demonstration earlier in the film. They’ll help structure our tour of this interface, so let’s review them now, as they all come up in our discussion of the interfaces.

  1. There is a voice-activated help system, which will summon a virtual “Angel” assistant.
  2. Since the computers themselves are part of a multi-user network with shared storage, there can be more than one user “inside” The Corridor at a time.
    Users who do not have access to the virtual reality system will appear as wireframe body shapes with a 2D photo where the head should be.
  3. There are no access controls and so the virtual reality user, despite being a guest or demo account, has unlimited access to all the company files. This is spectacularly bad design, but necessary for the plot.

With those bits of system exposition complete, now we can switch to Tom’s own first person view of the virtual reality environment.

Virtual world Tom watches his hands rezzing up, right hand with glove. Disclosure (1994)

There isn’t a real background yet, just abstract streaks. The avatar hands are rezzing up, and note that the right hand wearing the glove has a different appearance to the left. This mimics the real world, so eases the transition for the user.

Overlaid on the virtual reality view is a Digicom label at the bottom and four corner brackets which are never explained, although they do resemble those used in cameras to indicate the preferred viewing area.

To the left is a small axis indicator, the three green lines labeled X, Y, and Z. These show up in many 3D applications because, silly though it sounds, it is easy in a 3D computer environment to lose track of directions or even which way is up. A common fix for the user being unable to see anything is just to turn 180 degrees around.

We then switch to a third person view of Tom’s avatar in the virtual world.

Tom is fully rezzed up, within cloud of visual static. Disclosure (1994)

This is an almost photographic-quality image. To remind the viewers that this is in the virtual world rather than real, the avatar follows the visual convention described in chapter 4 of Make It So for volumetric projections, with scan lines and occasional flickers. An interesting choice is that the avatar also wears a “headset”, but it is translucent so we can see the face.

Now that he’s in the virtual reality, Tom has one more action needed to enter The Corridor. He pushes a big button floating before him in space.

Tom presses one button on a floating control panel. Disclosure (1994)

This seems unnecessary, but we can assume that in the future of this platform, there will be more programs to choose from.

The Corridor rezzes up, the streaks assembling into wireframe components which then slide together as the surfaces are shaded. Tom doesn’t have to wait for the process to complete before he starts walking, which suggests that this is a Level Of Detail (LOD) implementation where parts of the building are not rendered in detail until the user is close enough for it to be worth doing.

Tom enters The Corridor. Nearby floor and walls are fully rendered, the more distant section is not complete. Disclosure (1994)

The architecture is classical, rendered with the slightly artificial-looking computer shading that is common in 3D computer environments because it needs much less computation than trying for full photorealism.

Instead of a corridor this is an entire multistory building. It is large and empty, and as Tom is walking bits of architecture reshape themselves, rather like the interior of Hogwarts in Harry Potter.

Although there are paintings on some of the walls, there aren’t any signs, labels, or even room numbers. Tom has to wander around looking for the files, at one point nearly “falling” off the edge of the floor down an internal air well. Finally he steps into one archway room entrance and file cabinets appear in the walls.

Tom enters a room full of cabinets. Disclosure (1994)

Unlike the classical architecture around him, these cabinets are very modern looking with glowing blue light lines. Tom has found what he is looking for, so now begins to manipulate files rather than browsing.

Virtual Filing Cabinets

The four nearest cabinets according to the titles above are

  1. Communications
  2. Operations
  3. System Control
  4. Research Data.

There are ten file drawers in each. The drawers are unmarked, but labels only appear when the user looks directly at it, so Tom has to move his head to centre each drawer in turn to find the one he wants.

Tom looks at one particular drawer to make the title appear. Disclosure (1994)

The fourth drawer Tom looks at is labeled “Malaysia”. He touches it with the gloved hand and it slides out from the wall.

Tom withdraws his hand as the drawer slides open. Disclosure (1994)

Inside are five “folders” which, again, are opened by touching. The folder slides up, and then three sheets, each looking like a printed document, slide up and fan out.

Axis indicator on left, pointing down. One document sliding up from a folder. Disclosure (1994)

Note the tilted axis indicator at the left. The Y axis, representing a line extending upwards from the top of Tom’s head, is now leaning towards the horizontal because Tom is looking down at the file drawer. In the shot below, both the folder and then the individual documents are moving up so Tom’s gaze is now back to more or less level.

Close up of three “pages” within a virtual document. Disclosure (1994)

At this point the film cuts away from Tom. Rival executive Meredith, having been foiled in her first attempt at discrediting Tom, has decided to cover her tracks by deleting all the incriminating files. Meredith enters her office and logs on to her Indy workstation. She is using a Command Line Interface (CLI) shell, not the standard SGI Unix shell but a custom Digicom program that also has a graphical menu. (Since it isn’t three dimensional it isn’t interesting enough to show here.)

Tom uses the gloved hand to push the sheets one by one to the side after scanning the content.

Tom scrolling through the pages of one folder by swiping with two fingers. Disclosure (1994)

Quick note: This is harder than it looks in virtual reality. In a 2D GUI moving the mouse over an interface element is obvious. In three dimensions the user also has to move their hand forwards or backwards to get their hand (or finger) in the right place, and unless there is some kind of haptic feedback it isn’t obvious to the user that they’ve made contact.

Tom now receives a nasty surprise.

The shot below shows Tom’s photorealistic avatar at the left, standing in front of the open file cabinet. The green shape on the right is the avatar of Meredith who is logged in to a regular workstation. Without the laser scanners and cameras her avatar is a generic wireframe female humanoid with a face photograph stuck on top. This is excellent design, making The Corridor usable across a range of different hardware capabilities.

Tom sees the Meredith avatar appear. Disclosure (1994)

Why does The Corridor system place her avatar here? A multiuser computer system, or even just a networked file server,  obviously has to know who is logged on. Unix systems in general and command line shells also track which directory the user is “in”, the current working directory. Meredith is using her CLI interface to delete files in a particular directory so The Corridor can position her avatar in the corresponding virtual reality location. Or rather, the avatar glides into position rather than suddenly popping into existence: Tom is only surprised because the documents blocked his virtual view.

Quick note: While this is plausible, there are technical complications. Command line users often open more than one shell at a time in different directories. In such a case, what would The Corridor do? Duplicate the wireframe avatar in each location? In the real world we can’t be in more than one place at a time, would doing so contradict the virtual reality metaphor?

There is an asymmetry here in that Tom knows Meredith is “in the system” but not vice versa. Meredith could in theory use CLI commands to find out who else is logged on and whether anyone was running The Corridor, but she would need to actively seek out that information and has no reason to do so. It didn’t occur to Tom either, but he doesn’t need to think about it,  the virtual reality environment conveys more information about the system by default.

We briefly cut away to Meredith confirming her CLI delete command. Tom sees this as the file drawer lid emitting beams of light which rotate down. These beams first erase the floating sheets, then the folders in the drawer. The drawer itself now has a red “DELETED” label and slides back into the wall.

Tom watches Meredith deleting the files in an open drawer. Disclosure (1994)

Tom steps further into the room. The same red labels appear on the other file drawers even though they are currently closed.

Tom watches Meredith deleting other, unopened, drawers. Disclosure (1994)

Talking to an Angel

Tom now switches to using the system voice interface, saying “Angel I need help” to bring up the virtual reality assistant. Like everything else we’ve seen in this VR system the “angel” rezzes up from a point cloud, although much more quickly than the architecture: people who need help tend to be more impatient and less interested in pausing to admire special effects.

The voice assistant as it appears within VR. Disclosure (1994)

Just in case the user is now looking in the wrong direction the angel also announces “Help is here” in a very natural sounding voice.

The angel is rendered with white robe, halo, harp, and rapidly beating wings. This is horribly clichéd, but a help system needs to be reassuring in appearance as well as function. An angel appearing as a winged flying serpent or wheel of fire would be more original and authentic (yes, really: ​​Biblically Accurate Angels) but users fleeing in terror would seriously impact the customer satisfaction scores.

Now Tom has a short but interesting conversation with the angel, beginning with a question:

  • Tom
  • Is there any way to stop these files from being deleted?
  • Angel
  • I’m sorry, you are not level five.
  • Tom
  • Angel, you’re supposed to protect the files!
  • Angel
  • Access control is restricted to level five.

Tom has made the mistake, as described in chapter 9 Anthropomorphism of the book, of ascribing more agency to this software program than it actually has. He thinks he is engaged in a conversational interface (chapter 6 Sonic Interfaces) with a fully autonomous system, which should therefore be interested in and care about the wellbeing of the entire system. Which it doesn’t, because this is just a limited-command voice interface to a guide.

Even though this is obviously scripted, rather than a genuine error I think this raises an interesting question for real world interface designers: do users expect that an interface with higher visual quality/fidelity will be more realistic in other aspects as well? If a voice interface assistant has a simple polyhedron with no attempt at photorealism (say, like Bit in Tron) or with zoomorphism (say, like the search bear in Until the End of the World) will users adjust their expectations for speech recognition downwards? I’m not aware of any research that might answer this question. Readers?

Despite Tom’s frustration, the angel has given an excellent answer – for a guide. A very simple help program would have recited the command(s) that could be used to protect files against deletion. Which would have frustrated Tom even more when he tried to use one and got some kind of permission denied error. This program has checked whether the user can actually use commands before responding.

This does contradict the earlier VR demonstration where we were told that the user had unlimited access. I would explain this as being “unlimited read access, not write”, but the presenter didn’t think it worthwhile to go into such detail for the mostly non-technical audience.

Tom is now aware that he is under even more time pressure as the Meredith avatar is still moving around the room. Realising his mistake, he uses the voice interface as a query language.

“Show me all communications with Malaysia.”
“Telephone or video?”
“Video.”

This brings up a more conventional looking GUI window because not everything in virtual reality needs to be three-dimensional. It’s always tempting for a 3D programmer to re-implement everything, but it’s also possible to embed 2D GUI applications into a virtual world.

Tom looks at a conventional 2D display of file icons inside VR. Disclosure (1994)

The window shows a thumbnail icon for each recorded video conference call. This isn’t very helpful, so Tom again decides that a voice query will be much faster than looking at each one in turn.

“Show me, uh, the last transmission involving Meredith.”

There’s a short 2D transition effect swapping the thumbnail icon display for the video call itself, which starts playing at just the right point for plot purposes.

Tom watches a previously recorded video call made by Meredith (right). Disclosure (1994)

While Tom is watching and listening, Meredith is still typing commands. The camera orbits around behind the video conference call window so we can see the Meredith avatar approach, which also shows us that this window is slightly three dimensional, the content floating a short distance in front of the frame. The film then cuts away briefly to show Meredith confirming her “kill all” command. The video conference recordings are deleted, including the one Tom is watching.

Tom is informed that Meredith (seen here in the background as a wireframe avatar) is deleting the video call. Disclosure (1994)

This is also the moment when the downstairs VIPs return to the hotel suite, so the scene ends with Tom managing to sneak out without being detected.

Virtual reality has saved the day for Tom. The documents and video conference calls have been deleted by Meredith, but he knows that they once existed and has a colleague retrieve the files he needs from the backup tapes. (Which is good writing: the majority of companies shown in film and TV never seem to have backups for files, no matter how vital.) Meredith doesn’t know that he knows, so he has the upper hand to expose her plot.

Analysis

How believable is the interface?

I won’t spend much time on the hardware, since our focus is on file browsing in three dimensions. From top to bottom, the virtual reality system starts as believable and becomes less so.

Hardware

The headset and glove look like real VR equipment, believable in 1994 and still so today. Having only one glove is unusual, and makes impossible some of the common gesture actions described in chapter 5 of Make It So, which require both hands.

The “laser scanners” that create the 3D geometry and texture maps for the 3D avatar and perform real time body tracking would more likely be cameras, but that would not sound as cool.

And lastly the walking platform apparently requires our user to stand on large marbles or ball bearings and stay balanced while wearing a headset. Uh…maybe…no. Apologetics fails me. To me it looks like it would be uncomfortable to walk on, almost like deterrent paving.

Software

The Corridor, unlike the 3D file browser used in Jurassic Park, is a special effect created for the film. It was a mostly-plausible, near future system in 1994, except for the photorealistic avatar. Usually this site doesn’t discuss historical context (the  “new criticism” stance), but I think in this case it helps to explain how this interface would have appeared to audiences almost two decades ago.

I’ll start with the 3D graphics of the virtual building. My initial impression was that The Corridor could have been created as an interactive program in 1994, but that was my memory compressing the decade. During the 1990s 3D computer graphics, both interactive and CGI, improved at a phenomenal rate. The virtual building would not have been interactive in 1994, was possible on the most powerful systems six years later in 2000, and looks rather old-fashioned compared to what the game consoles of the 21st C can achieve.

For the voice interface I made the opposite mistake. Voice interfaces on phones and home computing appliances have become common in the second decade of the 21st C, but in reality are much older. Apple Macintosh computers in 1994 had text-to-speech synthesis with natural sounding voices and limited vocabulary voice command recognition. (And without needing an Internet connection!) So the voice interface in the scene is believable.

The multi-user aspects of The Corridor were possible in 1994. The wireframe avatars for users not in virtual reality are unflattering or perhaps creepy, but not technically difficult. As a first iteration of a prototype system it’s a good attempt to span a range of hardware capabilities.

The virtual reality avatar, though, is not believable for the 1990s and would be difficult today. Photographs of the body, made during the startup scan, could be used as a texture map for the VR avatar. But live video of the face would be much more difficult, especially when the face is partly obscured by a headset.

How well does the interface inform the narrative of the story?

The virtual reality system in itself is useful to the overall narrative because it makes the Digicom company seem high tech. Even in 1994 CD-ROM drives weren’t very interesting.

The Corridor is essential to the tension of the scene where Tom uses it to find the files, because otherwise the scene would be much shorter and really boring. If we ignore the virtual reality these are the interface actions:

  • Tom reads an email.
  • Meredith deletes the folder containing those emails.
  • Tom finds a folder full of recorded video calls.
  • Tom watches one recorded video call.
  • Meredith deletes the folder containing the video calls.

Imagine how this would have looked if both were using a conventional 2D GUI, such as the Macintosh Finder or MS Windows Explorer. Double click, press and drag, double click…done.

The Corridor slows down Tom’s actions and makes them far more visible and understandable. Thanks to the virtual reality avatar we don’t have to watch an actor push a mouse around. We see him moving and swiping, be surprised and react; and the voice interface adds extra emotion and some useful exposition. It also helps with the plot, giving Tom awareness of what Meredith is doing without having to actively spy on her, or look at some kind of logs or recordings later on.

Meredith, though, can’t use the VR system because then she’d be aware of Tom as well. Using a conventional workstation visually distinguishes and separates Meredith from Tom in the scene.

So overall, though the “action” is pretty mundane, it’s crucial to the plot, and the VR interface helps make this interesting and more engaging.

How well does the interface equip the character to achieve their goals?

As described in the film itself, The Corridor is a prototype for demonstrating virtual reality. As a file browser it’s awful, but since Tom has lost all his normal privileges this is the only system available, and he does manage to eventually find the files he needs.

At the start of the scene, Tom spends quite some time wandering around a vast multi-storey building without a map, room numbers, or even coordinates overlaid on his virtual view. Which seems rather pointless because all the files are in one room anyway. As previously discussed for Johnny Mnemonic, walking or flying everywhere in your file system seems like a good idea at first, but often becomes tedious over time. Many actual and some fictional 3D worlds give users the ability to teleport directly to any desired location.

Then the file drawers in each cabinet have no labels either, so Tom has to look carefully at each one in turn. There is so much more the interface could be doing to help him with his task, and even help the users of the VR demo learn and explore its technology as well.

Contrast this with Meredith, who uses her command line interface and 2D GUI to go through files like a chainsaw.

Tom becomes much more efficient with the voice interface. Which is just as well, because if he hadn’t, Meredith would have deleted the video conference recordings while he was still staring at virtual filing cabinets. However neither the voice interface nor the corresponding file display need three dimensional graphics.

There is hope for version 2.0 of The Corridor, even restricting ourselves to 1994 capabilities. The first and most obvious is to copy 2D GUI file browsers, or the 3D file browser from Jurassic Park, and show the corresponding text name next to each graphical file or folder object. The voice interface is so good that it should be turned on by default without requiring the angel. And finally add some kind of map overlay with a you are here moving dot, like the maps that players in 3D games such as Doom could display with a keystroke.

Film making challenge: VR on screen

Virtual reality (or augmented reality systems such as Hololens) provide a better viewing experience for 3D graphics by creating the illusion of real three dimensional space rather than a 2D monitor. But it is always a first person view and unlike conventional 2D monitors nobody else can usually see what the VR user is seeing without a deliberate mirroring/debugging display. This is an important difference from other advanced or speculative technologies that film makers might choose to include. Showing a character wielding a laser pistol instead of a revolver or driving a hover car instead of a wheeled car hardly changes how to stage a scene, but VR does.

So, how can we show virtual reality in film?

There’s the first-person view corresponding to what the virtual reality user is seeing themselves. (Well, half of what they see since it’s not stereographic, but it’s cinema VR, so close enough.) This is like watching a screencast of someone else playing a first person computer game, the original active experience of the user becoming passive viewing by the audience. Most people can imagine themselves in the driving seat of a car and thus make sense of the turns and changes of speed in a first person car chase, but the film audience probably won’t be familiar with the VR system depicted and will therefore have trouble understanding what is happening. There’s also the problem that viewing someone else’s first-person view, shifting and changing in response to their movements rather than your own, can make people disoriented or nauseated.

A third-person view is better for showing the audience the character and the context in which they act. But not the diegetic real-world third-person view, which would be the character wearing a geeky headset and poking at invisible objects. As seen in Disclosure, the third person view should be within the virtual reality.

But in doing that, now there is a new problem: the avatar in virtual reality representing the real character. If the avatar is too simple the audience may not identify it with the real world character and it will be difficult to show body language and emotion. More realistic CGI avatars are increasingly expensive and risk falling into the Uncanny Valley. Since these films are science fiction rather than factual, the easy solution is to declare that virtual reality has achieved the goal of being entirely photorealistic and just film real actors and sets. Adding the occasional ripple or blur to the real world footage to remind the audience that it’s meant to be virtual reality, again as seen in Disclosure, is relatively cheap and quick.
So, solving all these problems results in the cinematic trope we can call Extradiegetic Avatars, which are third-person, highly-lifelike “renderings” of characters, with a telltale Hologram Projection Imperfection for audience readability, that may or may not be possible within the world of the film itself.

IMDB: https://www.imdb.com/title/tt0109635/Currently streaming on:

High Tech Binoculars

In Johnny Mnemonic we see two different types of binoculars with augmented reality overlays and other enhancements: Yakuz-oculars, and LoTek-oculars.

Yakuz-oculars

The Yakuza are the last to be seen but also the simpler of the two. They look just like a pair of current day binoculars, but this is the view when the leader surveys the LoTek bridge.

jm-25-yakuza-binocs-adjusted

I assume that the characters here are Japanese? Anyone?

In the centre is a fixed-size green reticule. At the bottom right is what looks like the magnification factor. At the top left and bottom left are numbers, using Western digits, that change as the binoculars move. Without knowing what the labels are I can only guess that they could be azimuth and elevation angles, or distance and height to the centre of the reticule. (The latter implies some sort of rangefinder.)

So far, this is a simple uncluttered display. But why is there a brightly glowing Pharmakom logo at the top right? It blocks part of the view, and probably doesn’t help anyone trying to keep their eyes adapted for night vision.

LoTek-oculars

The LoTeks, despite their name, have more impressive binoculars. They’re first used when Johnny gets out of his airport taxi.

jm-11-lotek-binocs-a-adjusted

There’s a third tube above the optics, a rectangular inlet, and an antenna.

In these binoculars, the augmented reality overlay is much more dynamic. Instead of a fixed circle, green lines converge in a bounding box around the image of Johnny. Text slides onto the display from left to right, the last line turning yellow.

jm-11-loteks-animated

Zoomrect

The animated transition of the bounding box resembles what Classic MacOS programmers of the 1990s called “zoomrects” used for showing windows opening or closing. It’s a very effective technique to draw attention to a particular area of an image.

Animated text

Text appearing character by character is ubiquitous in film interfaces. In the 1960s and 1970s mainframe and minicomputer terminals really did display incrementally, as the characters arrived one by one over slow serial port links. On any more recent computer it actually takes extra programming to achieve this effect, as the normal display of text is so fast that we would perceive it as instantaneous. But people like to see incremental text, or have been conditioned by film to expect it, so why not?

Bioscanning

The binoculars detect Johnny’s implant. It might just be possible to detect this passively from infrared or electronic signals, but more likely the binoculars include a high resolution microwave radar as well. If there had been more than one person in view, the bounding box would indicate which one the text refers to. And note that the last line of text is a different color. What that means is unclear here, but it becomes clear (and I’ll discuss it) later.

The second time we see the LoTek binoculars is when a lookout spots Street Preacher, a very bad guy and another who wants to remove Johnny’s head. Once again the binoculars have performed more than just a visual scan.

jm-17-lotek-binocs-a-adjusted

The binocular view and overlay are being relayed to another character, the LoTek leader J-Bone who can watch on a monitor. Here the film anticipates the WiFi webcam.

jm-17-lotek-binocs-b-adjusted

The overlay text now changes.

jm-17-lotek-binocs-c-adjusted

Narrow AI?

This is interesting, because the binoculars can not only detect implants and other cyborg modifications, but are apparently able to evaluate and offer advice. It appears that the green text is used for the factual (more or less) information about what has been detected, while yellow text is uncertain or or speculative.

Does this imply a general artificial intelligence? Not necessarily. This warning could be based solely on the detected signature, in the same way that current day military passive sonars and radar warning receivers can identify threats based on identifying characteristics of a received signal. In the world of Johnny Mnemonic it would make sense to assume that anyone with full custom biomechanics is extremely dangerous. Or, since Street Preacher is a resident rather than a stranger and already feared by others, his appearance and the warning could have been entered into a LoTek facial recognition database that the binocular system uses as a reference.

These textual overlays are an excellent interface, not interfering with normal vision and providing a fast and easy-to-understand analysis. But, the user must have faith that the computer analysis is accurate. There’s no reason given as to why any of the text is displayed. If Johnny was carrying an implant in his pocket instead of his brain, would the computer know the difference?

An alternative approach would be some kind of sensor fusion or false spectrum display, with the raw infrared or radar image overlaid over the visuals and the viewer responsible for interpreting the data. The problem with such systems is that our visual system didn’t evolve to interpret such imagery, so a lot of training and practice is required to be both fast and accurate. And the overlay itself interferes with our normal visual recognition and processing. If the computer can do a better job of deciphering the meaning of non-visual data, it should do so and summarise for the human viewer.

Further advantages of this interface are that even a novice sentry will benefit from the built-in scanning and threat analysis, and the wireless transmission ensures that the information is shared rather than being limited to the person on watch. 

Brain Upload

Once Johnny has installed his motion detector on the door, the brain upload can begin.

3. Building it

Johnny starts by opening his briefcase and removing various components, which he connects together into the complete upload system. Some of the parts are disguised, and the whole sequence is similar to an assassin in a thriller film assembling a gun out of harmless looking pieces.

jm-6-uploader-kit-a

It looks strange today to see a computer system with so many external devices connected by cables. We’ve become accustomed to one piece computing devices with integrated functionality, and keyboards, mice, cameras, printers, and headphones that connect wirelessly.

Cables and other connections are not always considered as interfaces, but “all parts of a thing which enable its use” is the definition according to Chris. In the early to mid 1990s most computer user were well aware of the potential for confusion and frustration in such interfaces. A personal computer could have connections to monitor, keyboard, mouse, modem, CD drive, and joystick – and every single device would use a different type of cable. USB, while not perfect, is one of the greatest ever improvements in user interfaces.

Why not go wireless? Wireless devices remove the need for a physical connection, but this means that anyone, not just you, could potentially connect. So instead of worrying about whether we have the right kind of cable, we now worry about the right kind of Bluetooth pairing and WiFi encryption password scheme. Mobile wireless devices also need their own batteries, which have to be charged. So wireless may seem visually cleaner, but comes with its own set of problems.

As of early 2016 we have two new standards, Lightning and USB-C, that are orientation-independent (only fifty years after audio cables), high bandwidth, and able to transmit power to peripherals as well. Perhaps by 2021 cables will have made a comeback as the usual way to connect devices.

2. Explaining it

Johnny explains the process to the scientists. He needs them to begin the upload by pushing a button, helpfully labelled “start”, on the gadget that resembles an optical disk drive. There’s a big red button as well, which is not explained but would make an excellent “cancel” button.

jm-6-uploader-kit-b

It would be simpler if Johnny just did this himself. But we will shortly discover that the upload process is apparently very painful. If Johnny had his hands near the system, he might involuntarily push another button or disturb a cable. So for them, having a single, easily differentiated button to press minimizes their chance of messing it up.

1. Making codes

He also sticks a small black disk on the hotel room’s silver remote control. The small disk is evidently is a wireless controller or camera of some kind. The scientists must watch the upload progress counter, and as it approaches the end, use this modified remote to grab three frames from the TV display, which will become the “access code” for the data. (More on this below.)

jm-6-uploader-kit-x

None of the buttons on this remote have markings or labels, but neither Johnny nor the scientist who will be using it are bothered. Perhaps this hotel chain tries to please every possible guest by not favouring any particular language? But even in that case, I’d expect there to be some kind of symbols on the buttons and a multilingual manual to explain the meaning of each. Maybe Johnny spends so much time in hotel suites that he has memorised the button layout?

Short of a mind reading remote that can translate any button press into “what the user intended”, I have to admit this is a terrible interface.

(There is a label on the black disk, but I have no idea what it means or even which script that is. Anyone?)

0. Go go go

Johnny plugs in his implant, puts on a headset with more cables, and bites down on a mouthguard. He’s ready.

jm-6-uploader-kit-d

The scientist pushes the start button and the upload begins. Johnny sees the data stream in his headset as a flood of graphics and text.

jm-7-uploading-a

Why does he need the headset when there is a direct cable connection to the implant? The movie doesn’t make it explicit. It could be related to the images used as the access code. (More on this below.) Perhaps the images need to be processed by the recipient’s own optic nerve system for more reliable storage?

Still, in the spirit of apologetics we should try to find a better explanation than “an opportunity for 1995 cutting edge computer generated graphics.” Perhaps it is a very flashy progress indicator? Older computer systems had blinking lights on disk drives to indicate activity, copied on some of today’s USB sticks. Current-day file upload or download GUIs have progress bars. As processing and graphics capabilities increase, it will be possible for software to display thumbnails or previews of the actual data being transferred without slowing down.

Unfortunately there is an argument against this, which is that the obvious upload progress indicator is a numeric display counting gigabytes down to zero, and it makes a fast chirping sound as a sonic indicator as well. The counter shows the data flowing at gigabytes per second, the entire upload lasting about a minute. There’s also the problem that it’s not Johnny who is interested in knowing whether the upload is scientific data rather than, say, a video collection; but the scientists, and they can’t see it.

jm-7-uploading-b

As the counter drops below one hundred, the scientist points the remote with black disk at the TV display, currently showing a cartoon, and presses the middle button. The image from the TV appears overlaid on the data stream to Johnny. This is a little odd, because Johnny assured the scientists that he wouldn’t know what the access codes were himself. Maybe these brief flashes are not enough time for him to remember these particular images among the gigabytes of visual content. But the way they’re shown to us, I’ll bet you can remember them when they come up again later in the plot.

jm-7-uploading-d

Two more images are grabbed before the counter stops. When the upload finishes, the three images are printed out. (In the original film this is shown upside down, so I have rotated the image.)

jm-7-uploading-f-rotated

Tagged

So what are the images for? The script isn’t clear. I suggest that the images are being used as the equivalents of very large random numbers for whatever cryptography scheme protects the data against unauthorised access. Some current day systems use the timing of key presses and mouse movements as a source of randomness because humans simply can’t move their fingers with microsecond precision. Here, the human element makes it impossible to predict exactly which frame is chosen.

Humans also find images much easier to recognise than hundred digit numbers. Anyone who has seen the printout will be able to say whether a particular image is part of the access code or not with a high degree of confidence. In computer systems today, Secure Shell, or ssh, is a widely used encrypted terminal program for secure access to servers. Recent versions of ssh have a ‘randomart’ capability which shows a small ASCII icon generated from the current cryptographic key to everyone who logs on. If this ASCII icon appears different, this alerts everyone that the server key has been changed.

There’s one potential usability problem with the whole “pick three random images” mechanism. The last frame was grabbed when the counter was very close to zero. What would have happened if he had been too slow and missed altogether? Wouldn’t it be more reliable to have the upload system automatically grab the images rather than rely on a human? Chris suggests that maybe it secretly did grab three images that could have used without human input, but privileged the human input since it was more reliably random.

Quick aside: You may be asking, if images would be so wonderful, why aren’t we using them in this way already? It’s because our current security systems need not just very large random numbers, but very large random numbers with particular mathematical properties such as being prime. But let’s cut Johnny Mnemonic some slack,  saying that by 2021 we may have new algorithms.

OK, back to the plot.

-1. Sharing the codes

The access codes are to be faxed from Beijing to Newark, although this gets interrupted by the Yakuza intruders. This is yet another device with unmarked buttons.

jm-7-uploading-g

This device makes the same beeps and screeches as a 1990s analog fax machine. Since we’ll later learn that all the fax messages and phone calls are stored digitally in cyberspace, this must be a skeuomorphism, the old familiar audio tones now serving just as progress indicators.

As with other audio output, the tones allow the user to know that the transmission is proceeding and when it ends without having to pay full attention to the device. On the other hand, there is potential for confusion here as the digital upload is (presumably) much faster. Most current day computer systems could upload three photos, even in high resolution, well before the sequence of tones would complete. Users would most likely wait longer than actually necessary before moving on to their next task.

-2. Washing up

During the upload Johnny clenches his fists and bites his mouthguard. When the upload finishes, he retreats to the bathroom in considerable pain. At one point blood flows from his nose, and he swipes his hand over the tap to wash it down the drain. The bathroom announces that the water temperature is 17 degrees. We’ll come back to this later.

jm-8-bathroom-tap

As Make It So emphasises in the chapter on brain interfaces, there is nothing in our current knowledge to suggest that writing or reading memories to or from a human brain would be painful. On the other hand, we know that information in the brain isthe shape of the neurons in the brain. Who knows what side effects will happen as those neurons are disconnected and reconnected as they need to be? We don’t know, so can’t really say whether it would hurt or not.

-3. Escaping the Yakuza

As mentioned in a prior post, while he is in the bathroom, the motion detector Johnny installed on the hotel door isn’t very effective and the Yakuza break in, kill everyone else, and acquire the second of the three access code images. Johnny escapes with the first image and flies to Newark, North America. 

Starnav

STARNAV

To travel to Jupiter, navigator Zander must engage the Star Drive, a faster than light travel mechanism. Sadly, we only see the output screens and not his input mechanism.

Captain Deladier tells Ibanez, "Steady as she goes, Number 2. Prepare for warp."
She dutifully replies, "Yes m’am."
Deladier turns to Barcalow and tells him, "Number 1, design for Jupiter orbit."

In response, he turns to his interface. We hear some soft bleeping as he does something off screen, and then we see his display. It’s a plan view of the Solar system with orbits of the planets described with blue circles. A slow-blink yellow legend at the top reads DESIGNATING INTRASYSTEM ORBITAL, with a purple highlight ring around Earth. As he accesses "STARNAV" (below) the display zooms slowly in to frame just Jupiter and Earth.

StarshipT-STARNAV07

STARNAV

As the zoom starts, a small box in the lower right hand corner displays a still image of Mars with a label LOCAL PRESET. In the lower left hand corner text reads STARNAV-0031 / ATLAS, MARS. After a moment these disappear replaced with STARNAV-3490 / ATLAS, NEPTUNE, STARNAV-149.58 / ATLAS URANUS, STARNAV-498.48 / ATLAS, SATURN, and finally STARNAV-4910.43 / ATLAS JUPITER. The Jupiter information blinks furiously for a bit confirming a selection just as the zoom completes, and DESIGNATING INTRASYSTEM ORBIT is replaced with the simpler legend COURSE. Jupiter has a yellow/orange ring focus in on it as part of the confirmation.

Some things that may be obvious, but ought to be said:

  1. How about "Destination" instead of "Local preset"? The latter is an implementation model. The former matches the navigator’s goals.
  2. Serial options are a waste here. Why force him to move through each one, read it to see if that’s the right one, and then move on? Wouldn’t an eight-part selection menu be much, much faster?
  3. The serial presentation is made worse in that the list is in some arbitrary order. It’s not alphabetical: MNUSJ? It’s not distance-order either. He starts at 4, he jumps to 8, 7, and 6 before reaching 5, which is Jupiter. Better for most default navigation purposes would be distance order. Sure, that would have meant only one stop between Earth and Jupiter. If you really needed more stops for the time, start at Mercury.
  4. planets_iau (1)

  5. What are those numbers after "STARNAV-"? It’s not planet size, since Uranus and Neptune should be similar, as should Saturn and Jupiter. And it’s not distance, since Jupiter has the largest number but is not the fathest out. Of course it could be some arbitrary file number, but it’s really unclear why the navigator would need to know this when using the screen. If a number had to be there, perhaps a ranking like Sol-V Best would be to get rid of any information that didn’t help him with the microinteraction.
  6. How about showing the course when the system has determined the course?
  7. NUI would be better. When he looks at that first screen, he should be able to touch Jupiter or its orbit ring.
  8. Agentive would be best. For instance, if the system monitors the conversation on the bridge, when it heard "design for Jupiter," it could prepare that course, and let the navigator confirm it.

StarshipT-STARNAV08

Sneakily agentive?

Regular readers of my writing know that agentive tech is a favorite of mine, but in this case there is some clue that this is actually what happened. Note that the zoom to frame Earth and Jupiter happens at the same time as he’s selecting Jupiter. How did it know ahead of time that he wanted Jupiter? He hadn’t selected it yet. How did it know to go and frame these two planets? Should he select first and this zoom happen afterward? Did it actually listen to Deladier and start heading there anyway?

No.

It would be prescient if this throwaway interface was some secret agentive thing, but sadly, given that the rest of the interfaces in the film are ofttimes goofy, powered controls, it’s quite likely that the cause and effect were mashed together to save time.

STARNAV fuigetry

Though I can’t quite make sense of them (and they don’t change in the sequence), for the sake of completeness, I should list the tabs that fill the top and bottom of the screen, in case its meaning becomes clear later. Along the top they have green tab strokes, and read from left to right POS, ROLL, LINE, NOR, PIVOT, LAY. Tabs at the bottom have orange and purple strokes and read SCAN M, PLACE, ANALYZE, PREF, DIAG-1 on the first row. The second row reads SERIAL [fitting -Ed.], CHART, DECODE, OVER-M, and DIAG-2.