Dan Q found GLVVN5PH Pub Quiz

This checkin to GLVVN5PH Pub Quiz reflects a geocaching.com log entry. See more of Dan's cache logs.

Took me longer than it might’ve to look at this one because I’m feeling a bit saturated with “city centre” mystery caches: why are there so many these days? Pub quiz questions were right up my alley and it was only having forgotten about the existence of a particular piece of filming history that slowed me down – luckily I was able to derive from the maths where my error was likely to be and correct it, and geochecker confirmed where I was headed.

The first place I searched gave me a false sense of satisfaction when I felt something small and smooth stuck-on to something near the GZ… but it turned out to be a snail.

The second place a searched resulted in a knuckle full of stinging nettles. Ouch!

But the third place I searched put the cache right into my hand. Delightful.

Dan Q found GLVVKH3G Oxford-we-go – Geolympix legacy

This checkin to GLVVKH3G Oxford-we-go - Geolympix legacy reflects a geocaching.com log entry. See more of Dan's cache logs.

I’ve had two previous trips to the GZ to look for this cache: turns out that the hint in the Wherigo cartridge is completely inaccurate and mislead me quite a bit! To reiterate what others have said (I should’ve looked at the past logs sooner!), you’re looking for the kind of thing you’d normally expect to find given the cache size and location, NOT a combination lock.

Enjoyed the Oxford tour but ending on the edge of an industrial estate-turned-building site puts a dampener on the whole experience! Perhaps this final stage could enjoy being moved elsewhere? It’s possible it’ll have to anyway, once construction is complete here.

TFTC.

Dan Q note OK0231 St Giles Webcam Cache, Oxford

This checkin to OK0231 St Giles Webcam Cache, Oxford reflects an opencache.uk log entry. See more of Dan's cache logs.

The Oxford Internet Institute webcams are currently offline for an upgrade, so this webcam cache currently can’t be logged. Hopefully they’ll be back up and running soon, but if they’re being replaced with new ones then there’s a possibility that the viewing angle, position/facing or other characteristics might change: cachers take care!

Edit (2018-06-08): I’ve just spoken to the Oxford Internet Institute and they don’t know exactly when their webcams are coming back because they’re dependent upon the completion of some redecoration in the office from which they were based.

Review of City Mazes Oxford, Escape Rooms

This review of City Mazes Oxford, Escape Rooms originally appeared on Google Maps. See more reviews by Dan.

Best of the three escape room centres we’ve visited: we undertook a well-constructed and beautiful puzzle with some genuine “eureka” moments (thanks in part to a few hints which were well-timed and just-cryptic-enough to make us feel like we were still working for it…). We’ll be back!

Review of CRAFT Burger

This review of CRAFT Burger originally appeared on Google Maps. See more reviews by Dan.

Great burgers and drinks (often with well-chosen names with enjoyable wordplay) served in a fabulously relaxed environment. Sides/starters nothing to write home about, but that’s not why you’re here, is it?

Intermediary Protocols and Google Duplex

There’s a story that young network engineers are sometimes told to help them understand network stacks and/or the OSI model, and it goes something like this:

You overhear a conversation between two scientists on the subject of some topic relevant to their field of interest. But as you listen more-closely, you realise that the scientists aren’t in the same place at all but are talking to one another over the telephone (presumably on speakerphone, given that you can hear them both, I guess). As you pay more attention still, you realise that it isn’t the scientists on the phone call at all but their translators: each scientist speaks to their translator in the scientist’s own language, and the translators are translating what they say into a neutral language shared with the other translator who translate it into the language spoken by the other scientist. Ultimately, the two scientists are communicating with one another, but they’re doing so via a “stack” at their end which only needs to be conceptually the same as the “stack” at the other end as far up as the step-below-them (the “first link” in their communication, with the translator). Below this point, they’re entrusting the lower protocols (the languages, the telephone system, etc.), in which they have no interest, to handle the nitty-gritty on their behalf.

The OSI model reflected using the "scientists conversation" metaphor. Based upon original art by Yuki Fujimura, used under a Creative Commons License.
The two scientists are able to communicate with one another, but that communication is not direct.

This kind of delegation to shared intermediary protocols is common in networking and telecommunications. The reason relates to opportunity cost, or – for those of you who are Discworld fans – the Sam Vimes’ “Boots” Theory. Obviously an efficiency could be gained here if all scientists learned a lingua franca, a universal shared second language for their purposes… but most-often, we’re looking for a short-term solution to solve a problem today, and the short-term solution is to find a work-around that fits with what we’ve already got: in the case above, that’s translators who share a common language. For any given pair of people communicating, it’s more-efficient to use a translator, even though solving the global problem might be better accomplished by a universal second language (perhaps Esperanto, for valid if Eurocentric reasons!).

1950s illustration of "driverless cars of the future". The car follows a series of electronic markers down the middle of the highway.
In the 1950s and 1960s, the concept of a self-driving car was already well-established… but the proposed mechanism for action was quite different to that which we see today.

The phenomenon isn’t limited to communications, though. Consider self-driving cars. If you look back to autonomous vehicle designs of the 1950s (because yes, we’ve been talking about how cool self-driving cars would be for a long, long time), they’re distinctly different from the ideas we see today. Futurism of the 1950s focussed on adapting the roads themselves to make them more-suitable for self-driving vehicles, typically by implanting magnets or electronics into the road surface itself or by installing radio beacons alongside highways to allow the car to understand its position and surroundings. The modern approach, on the other hand, sees self-driving cars use LiDAR and/or digital cameras to survey their surroundings and complex computer hardware to interpret the data.

This difference isn’t just a matter of the available technology (although technological developments certainly inspired the new approach): it’s a fundamentally-different outlook! Early proposals for self-driving cars aimed to overhaul the infrastructure of the road network: a “big solution” on the scale of teaching everybody a shared second language. But nowadays we instead say “let’s leave the roads as they are and teach cars to understand them in the same way that people do.” The “big solution” is too big, too hard, and asking everybody to chip in a little towards outfitting every road with a standardised machine-readable marking is a harder idea to swallow than just asking each person who wants to become an early adopter of self-driving technology to pay a lot to implement a more-complex solution that works on the roads we already have.

LiDAR unit on a Google Self-Driving Car
In real life, these things spin much faster.

This week, Google showed off Duplex, a technology that they claim can perform the same kind of delegated-integration for our existing telephone lives. Let’s ignore for a moment the fact that this is clearly going to be overhyped and focus on the theoretical potential of this technology, which (even if it’s not truly possible today) is probably inevitable as chatbot technology improves: what does this mean for us? Instead of calling up the hairdresser to make an appointment, Google claim, you’ll be able to ask Google Assistant to do it for you. The robot will call the hairdresser and make an appointment on your behalf, presumably being mindful of your availability (which it knows, thanks to your calendar) and travel distance. Effectively, Google Assistant becomes your personal concierge, making all of those boring phone calls so that you don’t have to. Personally, I’d be more than happy to outsource to a computer every time I’ve had to sit in a telephone queue, giving the machine a summary of my query and asking it to start going through a summary of it to the human agent at the other end while I make my way back to the phone. There are obviously ethical considerations here too: I don’t like being hounded by robot callers and so I wouldn’t want to inflict that upon service providers… and I genuinely don’t know if it’s better or worse if they can’t tell whether they’re talking to a machine or not.

Process of having Google Assistant order a pizza, by phone, on your behalf.
I, for one, welcome our pizza-ordering overlords.

But ignoring the technology and the hype and the ethics, there’s still another question that this kind of technology raises for me: what will our society look like when this kind of technology is widely-available? As chatbots become increasingly human-like, smarter, and cheaper, what kinds of ways can we expect to interact with them and with one another? By the time I’m able to ask my digital concierge to order me a pizza (safe in the knowledge that it knows what I like and will ask me if it’s unsure, has my credit card details, and is happy to make decisions about special offers on my behalf where it has a high degree of confidence), we’ll probably already be at a point at which my local takeaway also has a chatbot on-staff, answering queries by Internet and telephone. So in the end, my chatbot will talk to their chatbot… in English… and work it out between the two of them.

Let that sink in for a moment: because we’ve a tendency to solve small problems often rather than big problems rarely and we’ve an affinity for backwards-compatibility, we will probably reach the point within the lifetimes of people alive today that a human might ask a chatbot to call another chatbot: a colossally-inefficient way to exchange information built by instalments on that which came before. If you’re still sceptical that the technology could evolve this way, I’d urge you to take a look at how the technologies underpinning the Internet work and you’ll see that this is exactly the kind of evolution we already see in our communications technology: everything gets stacked on top of a popular existing protocol, even if it’s not-quite the right tool for the job, because it makes one fewer problem to solve today.

Hacky solutions on top of hacky solutions work: the most believable thing about Max Headroom’s appearance in Ready Player One (the book, not the film: the latter presumably couldn’t get the rights to the character) as a digital assistant was the versatility of his conversational interface.

A man and a woman look at a laptop screen in a cafe/bar.
“See? My laptop says we should hook up.”

By the time we’re talking about a “digital concierge” that knows you better than anyone, there’s no reason that it couldn’t be acting on your behalf in other matters. Perhaps in the future your assistant, imbued with intimate knowledge about your needs and interests and empowered to negotiate on your behalf, will be sent out on virtual “dates” with other people’s assistants! Only if it and the other assistant agree that their owners would probably get along, it’ll suggest that you and the other human meet in the real world. Or you could have your virtual assistant go job-hunting for you, keeping an eye out for positions you might be interested in and applying on your behalf… after contacting the employer to ask the kinds of questions that it anticipates that you’d like to know: about compensation, work/life balance, training and advancement opportunities, or whatever it thinks matter to you.

We quickly find ourselves colliding with ethical questions again, of course: is it okay that those who have access to more-sophisticated digital assistants will have an advantage? Should a robot be required to identify itself as a robot when acting on behalf of a human? I don’t have the answers.

But one thing I think we can say, based on our history of putting hacky solutions atop our existing ways of working and the direction in which digital assistants are headed, is that voice interfaces are going to dominate chatbot development a while… even where the machines end up talking to one another!

× × × × ×

The woman who watched 300 executions in Texas

This is a repost promoting content originally published elsewhere. See more things Dan's reposted.

The woman who watched 300 executions (BBC News)

A former employee of the state of Texas describes the profound effect the executions had on her.

Texas has executed far more people than any other US state, and one former employee of the state has watched hundreds of executions unfold. She speaks to Ben Dirs about the profound effect that had on her.

It is 18 years since Michelle Lyons watched Ricky McGinn die. But it still makes her cry.

When she least expects it, she’ll see McGinn’s mother, in her Sunday best, her hands pressed against the glass of the death chamber. Dressed to the nines to watch her son get executed. Some farewell party.

For 12 years – first as a newspaper reporter, then as a spokesperson for the Texas Department of Criminal Justice (TDCJ) – it was part of Lyons’ job to witness every execution carried out by the state.

Between 2000 and 2012, Lyons saw almost 300 men and women die on the gurney, violent lives being brought to a peaceful conclusion, two needles trumping the damage done.

Lyons witnessed her first execution when she was 22. After seeing Javier Cruz die, she wrote in her journal: “I was completely fine with it. Am I supposed to be upset?”

The man with the golden blood

This is a repost promoting content originally published elsewhere. See more things Dan's reposted.

The man with the golden blood (The man with the golden blood)

Meet the donors, patients, doctors and scientists involved in the complex global network of rare – and very rare – blood. By Penny Bailey.

His doctor drove him over the border. It was quicker that way: if the man donated in Switzerland, his blood would be delayed while paperwork was filled out and authorisations sought.

The nurse in Annemasse, France, could tell from the label on the blood bag destined for Paris that this blood was pretty unusual. But when she read the details closely, her eyes widened. Surely it was impossible for this man seated beside her to be alive, let alone apparently healthy?

Thomas smiled to himself. Very few people in the world knew his blood type did – could – exist. And even fewer shared it. In 50 years, researchers have turned up only 40 or so other people on the planet with the same precious, life-saving blood in their veins.

I only have pretty regular (ABO-type O, Rh D+) blood, but it pleased me to discover this article right after I’ve started donating blood again after a long dry spell.