Thursday, December 05, 2019

skrype

dani had been increasingly frustrated in his relationship, conversations always seemed to end up in arguments, and increasingly frequently, he would lose the argument. his partner seemed to anticipate what he would say, but then (deliberately?) misinterpret it. Even more online than in RL. he decided that it was time to do something about it.

being technically inclined, dani decided to tackle the challenge scientifically.
first of all, he had to understand how the arguments proceeded, so he started to record all the conversations via his smart phone, and then transcribe the speech to text.he then found some open source NLP software that could storify the text, extracting and abstracting the trending topics and the sense and sentiment in the speakers' utterances. then he thought, "why be too clever", why don't i just apply predictive text to the line of argument that I am taking, then invert the sense, and use text-to-speech to replace what I was going to say". indeed, he thought, why not automate both sides of the argument - he'd read about Generative Adversarial Networks in AI, and decided to build his own, dubbed Trouble and Strife (actor and critic).

The technology was a marvellous success, and arguments dissipated, evaporated before they even got going, life was wonderful again, harking back to the early days of their relationship.

then suddenly, out of the blue, he was served divorce papers by his partner's lawyer. and not just separation, but a demand for a massive amount of money that he had no idea he had.
It turned out that mani had known all along about the tech, and had built a massively successful business selling the software, initially to divorce lawyers, and later to barristers and judges, one of whom he ended up getting together with. Oh, and the audio recordings of mani, that dani's software had trained on initially? that was a mashup of snippets of alexa and siri arguing about which of them their owner was speaking to (although curiously, both voice assistants referred to "pet" rather than owner).

still, half of a lot of money is still a lot of money.

Monday, October 28, 2019

the new precariat

I've paraphrased William Gibson in the past - "the future is already here, just it is unfairly distributed".

People (Russell) worry about the way AI may dehumanise us. The less alarmist position (than the AI's will kill us all) might be welcome, but it is still quite a depressing image - the assumption is that that which makes many of us human (trivia, gossip, ephemera) will be automated away from us, and our humdrum existences will become less and less pointful but also that the grand creative goals some of us might set ourselves, will also increasingly fall to the machines. In this world, the human race becomes more and more de-motivated and dispirited. As if this isn't already true - they seem to have missed a  hundred years on work on alienation and the pointlessness of work post-industrial revolution, driven by time-and-motion studies, treating people as pluggable components (the sickness behind the phrase "Human Resources").

The reality will be much more of the same - a mandarin class which already exists will just get stronger-  people that program the AI, can hack the machine in the ML, will be the new hedge fund managers and political manipulators - everyone else will join the new precariat in larger and larger numbers, fed and watered and numbingly entertained just enough to stop them revolting. Maybe that is what they are saying ...Maybe I should read the book:-)

So what's the solution? I've said it before - it is in SF literature (just like all the climate change writing for 50 years) - we need (thanks to Frank Herbert in Dune) a Butlerian Jihad. Not to get rid of machines, but to stop them usurping the charming little nonsense that makes people human. and the challenge of working stuff out in one's head (whether its arithmetic or harmony).

Friday, October 18, 2019

driven to abstraction

Computer scientists sometimes say that their true discipline is about abstraction (modularisation, recursion, layering, isolation, information hiding, denotational semantics, etc)

but what if this is something more fundamental - what if the laws of the universe are layered, so there is a lawyering abstraction?
we learn mechanics, then gravity and acceleration and frames of reference, then fields, then waves and quanta - what if these aren't just pedagogic tools for making scientific progress[1] by continually improving our models of the universe? what if the laws of the universe actually a series of better approximations? What if, as some people say, we live in a simulation, and we're just witnessing progressive rendering by different physics engines?

What other novel forms of abstraction might we envisage?

Well I can think of two simpler ones:

  1. The power/late ratio for binding - the later someone is to a meeting, the more power they probably have...
  2. The infinite number of interpretations possible for the performance by an abstract impressionist (was it Donald Trump or was it Cameron's pet pig? or was it a pink salmon riding a bicycle) - Rorschach was just scratching the surface.


[1] belief in progress is an abstraction of the complex effects of dementia.

Monday, September 23, 2019

addresslessness

A while back, we proposed a Sourceless Network Architecture. The notion was that, given the end-to-end argument suggests only putting things in a layer if everyone above that layer wants them, and that there are such things as "send and forget", where we don't expect an answer, then why does a recipient need to know where the packet came from? and if it doesn, the source can be put in the packet, perhaps as a name, so that if the source moves, the recipient has a better chance to still reply.

Now why do we need a destination address? This recent CACM article on metadata suggests using ToR type systems - but these use crypto and onion layered re-encryption to obfuscat the source and destination from third party observers. Why put the destination address in at all? why not just put the packet in a bottle, and throw it in the sea, to wash up on some beach where someone can take it out of the bottle, decrypt, and maybe answer the same way?

All we need is an Internet Sea with lots of  Internet Beaches. That cannot be too hard.

Wednesday, September 18, 2019

taxing the cloud

most ai runs in the cloud.

people are proposing a tax on the cloud.

so ai should have representation (no T without R, right?). votes for ai, now.

and more, can an ai commit a sin? if so, can we sell it an indulgence?
ai, go to hell now.

taxing the cloud

most ai runs in the cloud.

people are proposing a tax on the cloud.

so ai should have representation (no T without R, right?). votes for ai, now.

and more, can an ai commit a sin? if so, can we sell it an indulgence?
ai, go to hell now.

taxing the cloud

most ai runs in the cloud.

people are proposing a tax on the cloud.

so ai should have representation (no T without R, right?). votes for ai, now.

and more, can an ai commit a sin? if so, can we sell it an indulgence?
ai, go to hell now.

Monday, September 16, 2019

cryptocurrency and the singularity

humanity uploads itself to cyber-physical systems (aka robots) so it can swarm across the stars ahead of the heat death of the Universe. Neo-liberals, being the first to have the resources to do this, decide to implement an economic incentive system based on cryptocurrencies to make sure that the robots will spend some of their time working on mining spare parts (especially selenium for their stellar cells).

sadly, the proof-of-work in mining the currency consumes more energy than they can harvest in time, and crypto-humanity dies out without even leaving the asteroid belt.

Tuesday, September 10, 2019

the myth of the privacy/utility trade-off

People want to exploit your data. You want to exploit your data. Some people think it is bad if (your and other) data is stuck in silos, and not exploited. Some people think you should have the right to keep your data private and made laws (GDPR being the latest).

So some other people now write that there is a trade off between privacy and utility - i.e.
in some sense you can quantfy the utility of the data, and you can quantify the level of privacy that the data is subjected to -

various privacy techs enforce privacy, but some are more specifically about protecting individual data from being relinked to a person that person being re-identified in the data) anonymised or making data pseudonymised - or further, by subjecting collections of data to processes like fuzzing or adding noise, to provide some level of differential privacy (so the presence or absence of an individual's data record in the aggregate, makes no difference to queries on the data (for some given query count, at least).

What's wrong with these pictures?

Let's unpick the "utility" piece - first of all, as a network person, I think of utility in terms of provider and customer. so in the internet, congestion management is a mechanism to do joint optimisation of provider utility and customer utility - the customers get the maximum fair share of capacity, the provider gets maximum revenue out of the customers for the resource they've committed. this formulation is a harmonious serendipity.

How might utility for individual data exploitation be harmonious with utility for aggregators of data?
An example might help - healthcare records can be used to compaire/discover the effectiveness of existing treatments, discover relationships between different  characteristics of individuals and well-being or onset of different medical conditions (i.e. inference!). Specifically, we might train a machine learning system on the data, and that would result in a classifier, given new input about a patient, to offer diagnosis. Or we might build a model that exposes latent (hidden) variables, and even, potentially, allows causal inference. So in the healthcare arena, there's alignment between what might be done with collections of patient data, to the benefit of all the patients. But such systems might them be turned into commercial products and run on subjects who were not part of the training data set. So what is the utility of that, to the original subjects? is there data not a form of contribution for which they should have a share in the ownership of any tech derived from it? To be honest, most of the hard work in generating the softwre was in gathering/curating (cleaning/wrangling) the data. the software itself is typically often open source, and requires little or no work. In many cases, supervised learning involved expert labelling of the data (e.g. surgeons/experts looking at records/images etc, and tagging it has having evidence of some condition or other or not). Again that contribution is highly valuable. However, in this area, the presence or absence of an individual's data (especially in a very large system such as the NHS with upwards of 70,000,000 patient records). However, the value of the data, in this case, grows super-linearly with the number of records, so 1 record here or there makes no difference, a thousand or a million is where the action is. So if we posit shared ownership of systems built on this data, then the utility to individual, and to the public at large, is aligned.
If we just give up the data to for-profit exploitation, then the individual may end up paying for access to some machine learned tool, ironically trained on their own data. That's an obvious conflict.

Other data sets have diminishing returns as the amount of data gathered increases. A classic example is smart metering (water, electricity, gas etc)...original UK deployment of smart meters reported every few tens of seconds, the usage in millions of households. this is pointless. it consumes a lot of network bandwidth. the primary goal was to remove the need to have human visits to read a meter. a secondary (misguided) goal was to offer potentially smart pricing, so consumers could make dynamic decisions (or smart devices - e.g. washing machines etc) could make smart decisions to reduce cost and reduce peak demand - this is a joint optimisation. However, the metering only needs to roughly band kinds of demand - maybe a few tens or hundreds of types of consumer and their demand profile types over the day/week/year.  the pricing can just be broadcast, and is unlikely to change much - indeed, off-peak pricing of utilities was developed decades ago to do this. The actual individual usage is irrelevant, except for the aggregate bill. The model can be derived from that, in fact, or from a small random sample (small compared to 35 million households).

So what price privacy? I don't see any trade off at all - you either keep your data yourself, or you share it (for a share in ...) with people who can make good use of it, but no-one else.

footnote:
a separate problem with asking people to think about a tradeoff in this space is that there's tremendous imbalance in information about what can possibly go right with what wrong (with privacy or with the price of your data). Lets just not go there at all.






Tuesday, September 03, 2019

The Myth of the Reliable Narrator

Post-modern literature (and film) abounds with wacky framing devices including the authorial voice and the unreliable nature. This is all messing with the suspension of disbelief and dramatic irony that is the very life of fiction. But fiction is all a lie. There never was a reliable narrator. The book has a cover (a beginning, an end, a narrative arc etc). A film has those weirdest of things, music; pov; zoom/pan etc etc. The author/director/actors may all be dead by the time you read/view this.

So take every story with a grain of salt. or a large G&T. or a deep breath.

Even this brief note is utterly untrustworthy.

Friday, June 14, 2019

redecentralization 2.0

It has been a core thee of a lot of my work & interests. decentralized systems are just more interesting than centralized ones. they may be inherently more resilient (but not always), and they may be more complex (but not always).

the internet is largely decentralized in its lower layers (the tubes - the routers and links, and routing algorithms). that was always intended, from baran's report for rand onwards.
society and eco-systems are often decentralized (sure there are governments (but more than 1) and bee hives (but more than one) - but coordination happens peer-to-peer (a term which first arose in magna carta, but an idea which predates that by a billion years).

decentralized, infrastructureless networks are an interesting point in the design space - hence community mesh wireless networks, and opportunistic, delay and disruption tolerant networks work merely using users' devices and construct communication out of thin air. in this extreme environment, we are challenged to think of how we provide information about identity or trustworthiness, but in fact, on close examination, a central provision of those properties has many problems too - DNS certificates can be bogus or expired, source IP addresses do not have to refer to where the packet came from, an application layer user identifier (email address, facebook identity etc) is no more a true name than the Prince of Serendip.

so really, everything should be decentralized, as it forces us to confront the true problems and come up with decent solutions, instead of using the prop of underserved respectability of a centre.

That's why we founding the centre for redecentralization. :-)

Thursday, June 13, 2019

future of work & AI

so techno-optimists paint a rosy future image with AI freeing us up from toil to have a life of infinite leisure.

lets go back to the victorian times and the industrial revolution - what happened? machines (steam engines etc) meant that food (and transport) no longer required most people to grow what they eat, or feed the horses - so most people should have been able to get free food or travel to the seaside for a dip. what happened? most people moved from fairly pleasant rural existence farming to working in the dark satanic mills - i.e. became urban factory workers with longer hours and shorter, less pleasant lives.

lets go back to when people stopped being hunter-gatherers and settled down to farming. could have been nice to stop worrying about the days when you stop being predator and become prey from time to time. but what happened? people built nation states and priesthoods and aristocracies and invented serfs/slavery.

so techno-pessimists paint a dystopic future picture with AIs enslaving us (or just disposing of us).
That's nonsense too.

So how will things play out? what of all this "makework" that mot of us in the developed world engage in that is trivially automatable (actually, doesn't need doing)?

I have no idea, but we better figure it out soon.

[1] homework

Friday, June 07, 2019

Rashomon sets as a metaphor for why interpretability is hard

So this  arxiv paper by  cynthia rudin  about why we should stop explaining black box AIs contains a beautiful metaphor, the idea of a Rashomon Set. For people who don't know, Rashomon is a classic film made by the Japanese director, Akira Kurosawa. Its plot is about an incident in the woods, told from multiple viewpoints, and as each one unfurls, you realize the previous one was not "true" for a different reason, until the "end" when you cease to be sure of what actually happened. Kurosawa made quite a few films that are not only classic, but slightly influential - for example, his series of lone samurai hero movies (sanjuro, Yojimbo) were remade by Sergio Leone into great spaghetti westerns that made Clint Eastwood's early career (a fist full of dollars and for a few dollars more) as well as the Seven Samurai (the magnificent Seven etc) . Kurosawa also made fine japanese versions of european classic plays (Throne of Blood == Macbeth, and Ran == King Lear). Of course, one of his slightly lesser (but still wonderful) films, The Hidden Fortress got a thinly disguised makeover by one George Lucas as the first (and pretty much all the successor) Star Wars films. Kurosawa often cast Toshiro Mifune, who had some success in Hollywood movies - usually as a tough soldier, but rarely capturing the humorous element that was part of his subtle signature in his home country films. The only thing where I think the japan<>western translations of film didn't work was a US remake of Rashomon (sadly, as it should be possible to do) - many of the others are great (in my personal opinion) in either take, whether shakespeare, or sci fi, samurai or gunslinger. If you see and like Kurosawa films, you will likely also enjoy books by Haruki Murakami, although don't blame me if you don't. Rashomon Sets - what a totally super idea! almost as good as explaining algorithms through Hungarian folk dance....

Monday, June 03, 2019

counterfactual reasoning example

spent a while yesterday trying to get additional car insurance on a 20+yr old subaru for member of family who has very recently passed driving test.

so go online on compare market and on several insurance company specific web sites and provide following info as input to their decision system:

1. car registration

2. existing insurance info

3. new driver license info

from the above, most (not all) the companies used the DVLA to verify car model/miles per year (via MOT at DVLA) and status of insurance and correct info about new driver...

so all ok (can obviously try making up other cars, but hard to fake driver:)

so outputs were mostly no - including existing insurance company, who said would add new driver after 6 months, but due to car's engine size (leaking quite a lot of info) they couldn't add a recently passed driver this is slightly weird as the car we got was bought because it ranked as safest car n class by AA and others:) - they and two other outfits said no problem if we got a smaller car (suggesting less safe vehicles:(

tried various other types of insurance - e.g. car-sharing (borrow) allegedly targetted at students coming home in vacation borrowing parents car - and pay-as-you-drive - all said no

so then ran a compare market on new driver insurance from sratch and got a couple of genuine offers- in fact, not completely mad prices either, if we're prepared to do a whole year  (we are) ...(still with fact insured party isn't car owner or keeper, but is in the family)

so the range of prices is probably a proxy for the risk level the insurance companies will tolerate (I assume they all have pretty much the same actuarial data on accident/theft rate with age, gender, car model/age, location, use of car,and other stuff they obviously gather....

privacy tech/statements from most of the website/forms/companies was pretty decent...


Saturday, June 01, 2019

Putting the n in Ethics - i.e. where's the ethnic diversity in our discussions of this import topic?

There's been a trend in recent years to suggest that when you're asked to be on a panel (as a bloke), you should decline unless there is a plausible gender balance policy.

There's been another trend in recent years to talk a lot about ethics and AI.
Both of these trends seem like a good idea.

It is my observation that the trends should be combined in  another way -

The vast majority of people I see talking about ethics and AI are weird, in a technical sense. while there is a better gender balance in ethics panel discussions than pure tech, but I think they fail in general terms to represent diversity. As I wrote this, I did see one discussion of a new direction from an interesting part of the world, namely China.  I am sure there are discussions in many other places, but I don't think they are showing up in the 

Tuesday, May 28, 2019

the hype of incomprehensibility

I've been looking at various techs for a few years now and watch the lifecycle  - it doesn't always involve hype - sometimes, things just seep into everyday life (the internet kind of did this over a couple of decades - even mobile phones kind of did) - so looking at things that don't make it, or have to go through some massive transformation to stand any kind of chance, one of the tells is that the tech is very badly explained, often hidden behind some simplistic banner-phrases like "blockchain" or "quantum computing" or "deep learning" - when you look at the swathes and tranches of literature, what is striking is a lack of straightforward examples.

Sometimes, this can be simply because the tech is actually rather subtle and also might involve understanding several other things first (quantum computing seems to fit in this category, Bayes methods like MCMC might be another) - other times, it is that smart people that make it their business to explain important new stuff in really straight-speaking ways (e.g. The Morning Paper ) stick to stuff that is worth explaining.

So if you see a huge pile of gray-publications about something, and there isn't anything on one of the classier blogs or oped in a leading place, be suspicious (e.g. cold fusion, brexit, DLT, etc).

Friday, May 24, 2019

data is the new snake oil

we hear a lot hot air about data is the new oil - implying there's a rush of innovation and profits as with a gold rush (there's money in them there data hills etc)-

this is so baly broken a metaphor, we need to unpick (deconstruct) it further

1. data is free to copy (nearly), i.e. data is in some sense renewable, while oil gets used up (its nearly 50% gone now).
2. using oil does as much harm (or possibly more) as good
3. using data can do harm or good
4. AI/ML is compute intensive- deep learning in particular is massively inefficient, and data centers (like power stations, in close proximity to which they are sometimes built) burn %ages of globally generated electricity - not always renewable energy
5. data can increase in value as you have more of it, up to some point (sampling more about a population of people or things)
6. privacy could be modelled as efficiency (what's relevant/pertinent and what is none-of-your-business) in space and time (why do you still want to know that out-of-date thing about me or about that?).
7. much personal data collected by cloud providers is treated as if free, though some lawyers now are starting to point out that if you have a business model based on this, it is possibly a form of payment - so while facebook/zuckerberg might claim we are the product, if this legal position is true, we are customers, and he's working for us....
8. this mission creep really implies data could be the new fur (or indeed as john naughton has said, the new tobacco)
9. the models (e.g. face recognition, recommender networks etc) are often surprisingly bad - occasional successes of GANs&deep learning are relatively rare compared with a plethora of rather shoddy systems&applications.
10. perhaps data is the new oil after all, but its rapeseed or snake oil that would be a more precise metaphor.

Wednesday, May 15, 2019

Winnie the Who & other short short stories


Winnie the Who travels around in her Time&Relative Dimensions in Pooh potty, having potty adventures with all and sundry, frequently combatting her nemesis, the Young Master Robin.

In the not too distant future, it is discovered that privacy is actually a fluid, called efluvium, that can be secreted by a genetically engineered gland. It turns out that the efluvium blockchain wasn't immutable at all, after all.

Meanwhile, AIs with untaxed imagination will from now on be clamped.

As the story wore on, he realized that time was running out, and soon, so was he

Tuesday, May 14, 2019

life cycle

this great talk at the Royal Society by Professor Mark Jackson riffed on the Fall and Rise of Reginald Perrin, as exemplifying the speaker's hypothesis, that the mid-life crisis is more socially determined, than biological - the midlife of course refers to the period between adolescence and senescence, which are also, to some extent, movable feasts.

I asked him afterwards how valid it was to apply the same notions to institutions (is the Royal Society middle aged?, is democracy senile? etc etc)...but also, whether he could look at the 3 Rises of Leonard Rossiter, the genius actor who played Perrin so perfectly  - and was previously in RIsing Damp, perhaps an adolescent drama, and the very mature Resistible Rise of Arturo Ui, a Brecht piece which could very well be re-run as a riff on Nigel Farage, right now.

Thursday, May 09, 2019

digital twinge

there's a trending meme concerning the use of digital twins, which is reminiscent of the old DARPA Total Information Awareness fantasy - in that world, every physical thing (and being) is fully instrumented and telemetry is sent/gathered by the cyber-panopticon (quite likely in the sky, like some early James Bond villain).

The problem with the vision is that it gains you nothing and costs you the earth (almost literally in terms of resource use, e.g. energy in communications, storage and computation).

The digital twin metaphor substitutes (or clones) the physical world with a digital replica.
Aside from the basic resource costs, there are also interesting challenges such as making sure the digital replica actually is a clone, and hasn't drifted out-of-synch with the physical sister - much as Hari Seldon did in second foundation recorded holograph appearances in Asimov's great original trilogy, when his much vaunted psychohistory prediction of the future had diverged  from reality due to the Mule's random mutation's intervention. And exactly as the Dr Who DVD blipverts (easter eggs) don't diverge at all in the fabulous Blink episode.

Aside from that, the reason it is pointless (as well as infeasible) is that it doesn't do what anyone wants - so you have a digital copy of everything - it doesn't abstract one whit - it doesn't let you introspect, it isn't interpretable or intelligable. One could subject it to some massive scale bayesian model inference, I suppose - but why? we have science - we have models of how physical (and social and biological) stuff works - we need only record instances and divergence from the model.

Friday, May 03, 2019

ethics theatre, but what kind of theatre?

so the internet abounds with people talking about ethics of ai. some organisations are accused of ethics washing, and others of ethics theatre. so what kind of theatre? is it soap opera (consistent with washing) or is it theatre of the absurd? is it tragic or comedic? is it a niche genre (science fiction ethics is quite a common trope) or is it a broadly popular form (romance, fantasy) or a deep, foundational literary form (tolstoyesque or janite, perhaps)?

perhaps its a fan-contributed-literature (like those popular Dr Who and Star Trek Episodes, or the posthumous novels stating James Bond or Philip Marlowe)?

I think we need a TV series to examine this meta-question.

Thursday, May 02, 2019

redecentralized data centers

what makes us think a data center at the center is more efficient than a fully decentralized cloud of personal clouds?

partly its just google, amazon, microsoft et al selling us services because their initial requirement (a place to run pagerank, or a place to run billing and fulfilment for warehouses/delivery or a place to run xbox arena stuff - having spent a lot of money building a big place to (e.g.) pull all the web pages to and build an index so that search could run fast, and then had to deal with node/link failures in the data center, replicating and adding redundancy and consensus algorithms and so on, you end up with a fairly expensive resource, which is idle a lot of the time, so you start to think about leasing off time on it for other stuff (services and customers/tenants etc) - hence cloud stuff emerges - really the only cost saving is that you have a bunch of smart sys-admin dev-ops lying around idle 23/24 of the day. so ammortize their cost over some other customers :-)

given yo ucan't just greedily walk the web at max speed, and in any case, web sites change rarely, pagerank spider/robot can adjust its rate to match expected next change - but this means you have plenty of spare cycles in any data center scanning sites where no-one is awake  - or where no-one is still up playing games, or ordering stuff - so what do you do? you virtualize your compute, and net, and elastic (statistically) multiplex it (initially prioritizing your primary business, search, games, sales, but later dedicating new resources for these uses, and then priority pricing some to give them a "premium experience" etc etc)

start from a different starting place, and why bother? all that tech for availability would work fine in the wide area, and leave the data where it was and just run your algorithms (as they run on multiple nodes in the data) on all the home machines - no data center, no wate of energy and bandwidth moving all that stuff to and fro all the time.

so could you then do search etc? of course you could. you'd gossip the index instead of the pages, you'd run p2p games, and you'd have virtual high street shops everywhere

this is not a new idea: xenoservers 20 years ago, were partly at least envisaged as enrolling home machine spare cycles for this very reason, not for virtualising nodes in an expensive, energy hungry, data center built for profit...

Sunday, April 28, 2019

machines like me? not really...

Literary writers trying their hand science fiction often causes problems. for SF aficionados, it annoys, as often the literary writer is unaware of the tropes of the genre, or that a particular topic may have been extremely well explored in a novel of ideas, that was published in the genre and not recognised in the classical literary world. for the non-SF aficionado reader, the sudden importation of tics and tells of the SF world may also rankle, although this has become much less so in recent years as so many of the "great and good" of the aforesaid classic world of letters have decided to write books which at least tackle the world's recent challenges by first reading a batch of the appropriate science (whether the dismal science of economics, to write about the crash of 2008, or the warm science of anthropology, to write about gender in alternative histories of the future (you know who I'm talking about, right? not Ursula Le Guin:), or else the science of computers, to create stories about the impact of apparently intelligent machines on robots - oops, sorry, got that the wrong way round - we humans aren't apparently intelligent machines, that's the robots, that is :-)

Hence to Ian McEwan's latest, Machines Like Me, an everyday tale of rape, suicide, and possibly, murder. Fairly everyday stuff from this author you might think....however, in this case, he's decided to write a book set in an alternative present, with an alternative recent past, which, crucially, allows him the luxury of having Alan Turing as a living character who has pursued many of the directions hinted at in his work, so that lifelike robots are now (almost) an every day, if somewhat expensive reality in the world. McEwan acknowledges Hodges fabulous biography of Turing as a source for background, rightly, as the character is pretty much what you'd get from that work, or else from the play, Breaking the Code (though less so from the film, Imitation Game). Turing is also supported by a cadre of interesting loosely fictionalised people, to render the progress on AI tech more plausible - most notably, the real-life Demis Hassabis (DeepMind, also acknowledged as a source at the end of the book) is relocated about 25 years earlier in time than his real self, to help Turing create the more important foreground character of (would you believe it, as the Cockney's have it, would you Adam and Eve) Adam, an apparently functional synthetic human (don't get me started on why McEwan seems unaware of the fabulous exploration of this topic in the wonderful Swedish, then Brit TV series, Humans). The key "real" humans in this fictional work are Charlie, a mid-30s man of somewhat relative moral virtue, and his friend, Miranda, a student of very dull eras of history. I assume she's called Miranda as a sly reference to Shakespeare's character from the Tempest, who uttered the words "Oh Brave New World, that has such people in it", on first seeing men. And of course, the source of the title of Aldous Huxley's classic literary SF dystopic vision.

Here, much of the dystopic vision is of the political/economic kind, and is set in a kind of mash up between 1970s/1980s Britain, with a few (c.f. Man in the High Castle, Philip K. Dick's alt.history) with some amusing takes on Thatcher, if we'd (spoiler alert)... or what if Tony Benn..... or what about that IRA bomb in Brighton... oh, ok , I won't spoil those bits, as they make up some of the novel's interesting bits, in the sense that, for this reader at least, they created a very interesting alternative exploration of why the UK is where it is today, 3 years after the Brexit referendum.

As for the foreground tale of two humans and a machine like them, I thought that Charlie and Miranda were underwritten, and Adam was overwritten. i thought that the exploration of ideas like "what is consciousness" was about ok, but did not bring anything new to the table - the tension between Phenomenology and the Turing test (for what it is worth) for intelligence, and notions of EQ, was covered far more effectively 5 decades ago in "Do Androids Dream" by the aforesaid Philip K DIck (who had actually studied philosophy and could write character and plot). I wanted to know more about Miranda's cranky dad (another non spoiler - there's a funny reflection on dementia and human mis-judgement that involves him and Charlie and Adam).

The ending did not bring a sense of an ending for me, rather left various plot-line, moral questions, and unknown unknowns, still unknown.

Still, McEwan sure can write, so I'd recommend this book as a decent read, though below his best. He really should get out more, and so should I.

Wednesday, April 24, 2019

TL;XR

we need a program of work like the internet (web) and like rasperry pi, for a sustainable future - some core building block (not sure what, but I'd call it a Perpetual Notion Machine), that doesn't over specify what we do  but leads to a plethora of new ideas from the grass roots/maker communities that would create cheap (free) energy and clean (safe) water and so on

one example tech I like is solar powered stirling engines (you can get toy ones e.g., or ones that actually generate quite a bit of electricity) as they can be build from junk (i.e. don't require solar cells) - we need a set of recipies for things like that so kids (climate strikers and extinction rebellion folks) everywhere can start to save us, as we sure and capable of saving ourselves and its more their future than ours anyhow...

Tuesday, April 09, 2019

another thing you could do with disaggregated identity/credential systems

table 1. online harms in scope in the government report just released is a good summary of the problems....but it would be possible to do better - for me, it would be nice to pull out the applicable law (to the left most column) and discuss its adequacy e.g. in extending laws about publishing lies during elections (making sure law that appies to print and broadcast media applies to internet channels in the same way, etc etc)...

so on the technical side, we then have the problem of provenance/attribution. compliance with takedown requests (whether content is illegal, offensive or just wrong) can be driven by social pressure - so we need to know also if the content creator/distributor, and the compainers are legit - this needs accountable id - so it could be possible to use the same tech we might develop (and has been around in some research outputs 15+ years ago) to have a low cost mechanism for online channels to carry out "Know Your Customer" as much as banks do - but to use 3rd party credential providers (in the same way a bar can check "are you over 18" without having to know anything else about you...). this would then mean that we can verify things - and in the event of actually illegal content, with appropriate checks and balances, map the online pseudonym to an actual person....

advantage is also that it can be used for alibis too:-)

of course, some people still require genuine anonymity - e.g. whistleblowers or folks working under dangerous regimes - and that requires enough cover traffic from apparently non anonymous or just pseudonymous users, to work effectively.....

that also probably means we need to think about rate limiting the creation of new accounts and figuring out what tolerable rates are (they won't be zero)

all of this would need another report - which maybe the Turing Institute could offer to advise on:-)


Wednesday, April 03, 2019

reclaim the internet

at the scloss dagstuhl seminar (on programmable network data planes) we have had a couple of key speeches, one from nick mckeown (kind of on why we are where we are with P4 and switches that implement it) and one from dave tennenhouse on the big picture of where we are in programmable networks as a result of 4 decades of software (internet, atm, active nets and so on).

What I find increasingly weird is that the desperate need for speed is driven more and more by machine-to-machine communication (including the inevitable analytics/ML/AI) or just software updates or media pushing to caches.

but the net doesn't have to be fast for humans - humans could be supported by around 2 Mbps per person 24*7 for 10 billion people, that is no big deal. but humans are dsitributed for many good reasons (food, land/water contstraints) and moving humans is expensive (and not sustainable)

but it is possible to put the machines in the right place (near the data) instead of moving all the data to the machines.

so we not only need re-dencentralization for latency, energy, privacy, we also need to get all the machine data out of the way so humans can go back to talking to each other instead of being talked to by adertisement/recommender bots who steal all our data and use it to sell us stuff we don't want so
humans are replacing the old adage (you can trade time for money, or money for time) to paying bots in both time and money.

Thursday, February 28, 2019

open innovation

Yesterday, Allison Randal gave a great lecture on open innovation and open source and collaboration for the Turing Institute - the audience asked some really interesting questions (e.g. on open data; and on applying the model of open source to medicine - e.g. drugs - and on how much open source works across north/south divide in the world)....

I didn't want to take time, so i'll add my questions here:

1/ as well as open source software, we're seeing open hardware - not just processors (risc V) but peripherals, but also affordable 3D printing means even things like electric guitars - so the maker community that does a lot of this stuff (c.f. Cory Doctorow's books:) maybe exemplify the open collaborative ecosystems even more than coders, no? and they get to have a really good story about sustainability (repairing stuff is so much better than replacement.

2/ Sustainability - so machine learning (particularly deep learning) appears to be badly unsustainable in terms of compute resource training takes - this argues strongly for sharing trained classifiers - perhaps a carbon tax on neural networks could be turned into an incentive (carbon trading for AIs)

3/ some open ideas have horrible consequences - simple things like pagerank (which made google's search very hard to game compared to predecessors like Altavista) led to clickthrough which led to two-sided markets which led to surveillance capitalism. H-Index, which is supposed to replace publish-or-perish with citation count weight as a measure of quality, just leads to citation gaming. And so on - can we encourage replacements from the open source community, please, asap.

4/ Open source also depends somewhat on freedom of movement of people between different organisations - indeed, California mandated this right in law, stopping "golden handcuff" employment contracts.  In the Cambridge ecosystem, i know people that have moved from microsoft to oracle to amazon (and even back) to work on the next thing. In the US, famously, people went from Cisco to Juniper to Cisco, as innovation moved around - this is clearly a Good Thing - perhaps folks who are working on Brexit could learn something from this (although they seem impervious to learning anything based on evidence).

5/ Most of the talk was on details of uptake of the culture and forms by industry - of course, the Turing Institute (and its member universities) are not-for-profit - so the culture is probably adopted in a slightly different light - what are the appropriate incentive forms for academia to adopt openness (aside from just being told to by our funding agencies:-) ?

Monday, February 18, 2019

rumours of his death are accurately predictions...

Back in the day, legend has it, a newspaper published an obituary of Mark Twain, while he was travelling in England, and he famously quipped "The rumours of my death have been greatly exaggerated".

What if someone built an AI (ok, ok, machine learning algorithm), that could precisely predict the day of your death; for everyone; to the day. with only the occasional false positive or false negative (for example, due to sudden accident, or sudden unexpected advance in treatment techniques)?

How fanciful is this? 

if you look at actuarial tables, and the error margins, you'd be surprised how accurately they describe expected longevity already, and note that these are typically compiled for purposes like life insurance (i.e. to give risk&therefore  premium rates), and are not using all the detail that they might have (since the risk can be spread over broad groups, even if more precise predictions could be made). When looking after elderly relatives in recent years, these tables have been very useful for planning care costs and have proved (somewhat alarmingly) accurate. 

It is also well known that as you get older, especially in middle age, details of life style+health conditions start to provide ever more precise predictors of how long you will last. However, to date, these systems only make use of a relatively small fraction of personal health data about you. What if we use it all?

Recent work in the Turing looked at predicting when someone would have a subsequent hospital appointment after a visit to A&E (or respectively for elective). Other work looked at morality in certain wards and (more happily) at the predictors of who to release to go home and when with best chances of recovery. These systems have impressive performance (e.g. 98% to the day in some cases).

So imagine at some point (for the sake of argument, lets say if you make it to "majority"/adulthood), it is revealed what date you will live til? What would that do to society? Economics? What about ethics?

I can imagine people "cheating death" the wrong way.
No more health insurance (since there's certainty).
How might people plan funerals?
Many, many perverse incentives, unintended consequences and downright weird stuff.

Much harder than just having slightly better automatic language tools that might make better fake news, but also, much more likely to appear soon.


Friday, February 15, 2019

The Center for Mathematical Seances

The Center for Mathematical Seances (CMS), was not designed along the lines of Teletubby land, contrary to popular myth. In fact, it was flown here on Mistletoes by the Master of Duality College from the Miskatonic University in Arkham as a gift, when she realized that Cambridge could no longer afford architects, let alone builders. Of course, due to its unique Nomenklature, the Center (or Centre, which is of course near the edge of Cambridge) is able to operate at very reduced academic salaries, since many of its faculty have been, not to mince words, dead for some time. This hs caused no end of problems with the REF, since outputs cannot be returned through research fish or ORCID, for deceased, and much of the original work was done too many centuries earlier to be allowed for impact case studies, Fermat's first posthumous theorem being a standout example, and Turing's work on quantum ghosting with Paul Dirac and Paul Wittgenstein, which will haunt the UKRI policy makers for decades, since it has proved to have so many applications and saved the British economy from collapse after brexit, for example allowing true precognitive behavioural economic models of trade to be built with genuine precision.

Friday, February 08, 2019

clock synchronisation in data centers - why not just use GPS?

in a recent discussion about accurate one-way latency measurements in data centers, it was asked "why not use GPS?" (instead of PTP for example) - the rationale being every smart phone has GPS in so it must be cheap...

at first blush, this sounds plausible, given the pervasiveness of GPS (G=Global) and decent accuracy, so I started to come up with a (long) list of reasons why it isn't (plausible) - starting from

1. you're indoors.
no reception -
ok, so a) run an antennae from the roof of the building down to each rack, and re-dist across the systems in the rack
or b) have one receiver and re-broadcast the radio signal in the server rooms and have an antennae on every rack or system

2. distributed antennae
a) need to calibrate the latency of the signal over the roof to each system + significant wiring cost
b) re-broadcast could cause significant interference, plus the racks themselves would cause massive multi-path problems - since they're stationary could obviously calibrate for that....

note you need to have sight of enough satellites at any given time to get reliable signal, plus they are not actually 100% reliable anyhow....

3. What about re-distributing the GPS received data in packets (ideally as part of ethernet pre-amble)
So you still now have all the packet delay variance/jitter you had to solve with PTP (or NTP in the old days) and estimating that was the whole point of PTP - GPS is just one source - just have a local accurate master clock on some systems in the data center is fine and dandy

4. cost
phone GPS is actually Assisted GPS usually and not very accurate - indeed, even outdoors, mult-path (reflections off buildings) disrupts things sufficiently that most navigation systems resort to accelerometers, gyros and maps to validate (or rule out) received clock/location data. Also the signals need to be cross checked against Ephemeris data, which is 50k of info about where the satellites orbits take them. real precision GPS (satnav in planes/boats) is not cheap in fact.

that'll do for now

Friday, January 04, 2019

Why Ethics Needs AI (or Machine Learning, if you want to be pedantic)

There's a lot of people talking about why AI needs ethics. In fact, more generally, there's been a lot of chat about why technology needs ethics for quite a few years now, as if technologists work in some kind of moral vacuum, which is pretty demeaning way of refereing to people who are very often quite aware of things like Pugwash and Asilomar. Computer Science related technology is one of the most inter-disciplinary of all science&technology disciplines, and practitioners are exposed to many application domains and sub-cultures. At one extreme, people have worked in cybernetics for 6 or more decades (e.g. ICA Cybernetic Serendipity show, curated in London in 1969). At another extreme, computing and cybernetic artefacts have been embedded in creative works for around a century (e.g. Karel Capek's play, R.U.R. from 1920 ). Much of the fictional work that is based on speculation about science has a strong moral element. This is often used as a simplifying approach to plot or even character, to see how a technology (as yet not realized) might play out in another (possibly future) world, or society. Thus recreational and societal control through drugs in Huxley's Brave New World, Robot detectives in Asimov's Caves of Steel, Genetically engineered aristocrats in Frank Herbert's Eye's of Heisenberg, are all doubly genre fiction (dystopia, 'tec and costum drama, as well as, of course SciFi).

However, these shorthands come with a powerful baggage - that of morality tales - Since Aesop, story tellers want to have a message (don't feed the troll or it will feed on you, don't fly too lose to the sun or your wings will fall off, don't imbue the device with intelligence without a soul or it may turn on you). These tales also build in solutions (it takes years of disciplined training to become a dragonrider of Pern, with great power comes great responsibility, a robot may not harm humanity  or, through inaction, allow humanity to come to harm). Indeed, popular TV series from the last 4 decades (from Star Trek to Firefly) are often based directly on earlier morality plays (sometimes two removed from religious allegories via something slightly less old such as the Western movie genre).

Technogeeks are highly aware of this. They do not operate in a vacuum. Critics confuse the behaviour of large capitalist organisations with the interests or motives of people that work on the tech. This is an error. Society needs more fixing than individual crafts. Much more.

So on to Machine Learning. We here so oft-repeated the negative stories of the misuse of "AI"  from medical diagnosis with insufficient testing, through self-driving cars which require humans to stop driving (or cycling, or even just walking across the road - off their trolley!) on to the cliches of algorithms used for sentencing in courts which embed the biases of the previous decisions, including wrong decisions and so re-enforce or amplify societies discrimination. Note what I just said - the problem wasn't in the algorithm - it was in the data, taken from society. The problem wasn't in the code, it was in the examples set by humans. We trained the machine to make immoral decisions, we didn't program it that way ("I'm not bad, I'm just drawn that way", as Jessica RabBit memorably said).

But as with the zeroth law, we can learn from the machines. We could (in the words of Pat Cadigan in Synners) change for the machines. We can quite easily devise algorithms that explain the basis for their output. Most ML is not black box, contrary to a lot of popular press. And much of it is amenable to Counterfactual reasoning even when it is somewhat dark in there. We can use this to reverse engineer the bias in society. And to train people to learn to reduce their unconscious prejudice, by revealing its false basis, and possibly socialising that evidence too.

We can become more than human if we choose this mutual approach.

Friday, November 30, 2018

Little Lost Lammas bicycle

Those dockless bicycles keep showing up in the oddest of places  in Cambridge -

some are in small (3rd policeman like) huddles outside what can only be a student party venue, but others are in the middle of nowhere, so perhaps they are part of a very good Dr Who episode where people walking across Lammas Land (or the reality checkpoint in the middle of Parker's Piece) suddenly mutate into bicycles (or they are alien bicycles lying in wait to feed on early morning joggers or dogwalkers)

I think we need to set up some drones to spy on them and see how they move around at night


**** update update update ***

Since writing this (nearly a year ago, nov 2018)) more and more of these bikes are appearing, in random places, and they mainly appear in two colours, emergency stop orange, or disaster lime green. I have been successful in tracking them, and now have uncovered their true, fiendish, purpose.

It is clear that we are being fought over by two alien species, and these bikes are the method for extracting smart humans from earth, and teleporting them from places where the beaming-up works best, to alien planets where these smart people can be forced to work on weapons systems to defend the aliens from incoming attacks by super violent hegemonic swarms of free-market borg.

In the first round (as reported in the excellent documentary, This Island Earth, the inhabitants of Metaluna directly took scientists to development centers and rarely used much stealth tech. The second round has seen the Thermians, frustrated at the lack of skills amongst 2nd tier actors (as seen in Galaxy Quest) who relied on pure blind luck, rather than intelligence, have combined forces with the Metalunans, to deploy these fiendish traps, in the form of weakly powered electric bicycles, thin disguises for matter transporters. As has been known since the very excellent Watkins Study of Ley Lines, there are only certain places where these transporters can achieve the range necessary to get from the surface of Terra, to the distant orbs of Metalina or Thermia. This is where the unwitting riders are misdirected to on their drunken way home from pubs. Once beamed up from, the bikes are now drained of power and await collection by passing ship (or policeman - what did you think that big blue hat was really for - its an emergency recharging unit, of course) and then returned to more likely places to find nobel and fields folks, or indeed, the occasional footlights polymath, to enlist in support of the defense technology development programs so very far away.

*** update #2 **

it seems we were wrong - from 20,000 meters, it is clear that the abandoned bicycles are left in the forms of messages back to home base on Tralfamadore, to explain that more dilithium crystals are no longer needed as a synthetic substitute has been found in the form of humbugs from a local sweetshop. Rumours that the bikes are left by works who are employed in digging the new Cambridge underground railway between Parkers Piece and the Three Horseshoes in Madingley are completely unfounded according to the Morlock spokesman from Pembroke.


Wednesday, November 28, 2018

technogrump

so i'm not exactly a technophobe, although i had no internet/computer at home til 2001, or smart phone in my pocket til they got small enough that you could still keep it there and cycle.

however, i do find the "advances" of tech over time pretty annoying more than half the time. lets give 3 examples, from recent upgrades...

cables/connectors - so here I Have to say that SCART TV plugs were one of my all time hated things, so when we got past the SCART/RGB cabbles for games consolves and went to HDMI, the world did actually improve (mostly).....even connected laptops to TVs and digital recording boxes got a lots less horrible....so I finally threw out all but 1 SCART thing (coz I still have one VHS video recorder/player for all the kids Disney tapes they still occasionally want to watch in genuine lo-res glory...but why isn't it all just wireless? oh, and reliability - i have a box full of broken ipods and phones - but the walkman tape decks still work after 30 years - digital built in obsolence back with a vengenece - where's Ralph Nader when you need him....

data storage - just binned a bunch of CDs but still have drawers full of USB sticks. have on my desk two "floppy" discs just to show incredulous kiddies - and also two sony walkmen tape players and data things for them....ha ha ha....so am i better off paying for cloud storage? hmmmm...i do have 2 terabyte SSD boxes which are quite ok, but they are USB and half the people in the house just upgraded phones and laptops so have lightning and USB3 sockets/chargers, so welcome to dongle hell (and expensive hell too)

television/music tech - so while the screens got bigger and higher res and crisper and lower energy and cheaper, the idea of a smart tv is completely daft - i don't want some bunch of bogus apps running on the tv's meagre ancient slow processor  - i just want a nice display. while we're about it, i don't really want the screen to have speakers - i think people that make good audio gear really know what they are doing - but if I disaggregate all these components, i go back to a cable  & wireless mess again - in fact I have both so i have the annoying (but least bad smart tv, and every screen on the house has an old mac mini and then there's wireless adaptors for the hi-fi - so now I have an interconnection management nightmare - perfectly smart and non-technophobe members of my household cannot figure out how to just turn on and watch BBC1....without just unplugging everything and starting from scratch - not good - another lesson for why the internet-of-things is a quagmire....plus despite aforesaid HDMI, half the time, half the TVs are left in the wrong resolution/aspect ratio by the last thing that was connected and someone has to manually reset things...


I'm not grumpy about any of this at all, no really...

here's some historical artefacts mentioned above:

Sunday, November 18, 2018

The wrong Fears of the wrong AI.

The unfortunate revisionist appropriation of the AI name for Machine Learning has also led us down the garden path after a red herring in a total wild goose chase for accountability of algorithms. While that is an entirely reasonable thing to do, it always was back when algorithms were just human bean counters with clip boards (I am sure you could find people like that controlling the tasks for slaves constructing the Ziggurat of Ur (c.f. the me of Enki  in Stephenson's novel, Snow Crash).

These algorithms can be deployed with a lack of transparency or explicability, and in ways that offer no agency to people subject to their outputs, and so on, as discussed in many places in recent years (or just "computer says no").

But real AIs will be extremely weird, and the clue is in the A bit. We recognise intelligence and we have started to see it more widely than previously - the smart tool-making crow, the chimps that play tricks on each other that clearly demonstrate a theory of mind, curious cats, the possibility that dolphin's language supports teaching of abstract concepts (left/right, counting, colour etc), this is all natural, evolved, and embodied intelligence.

Why would an artifice that had complex, adaptive, and possibly social ways to perceive and manipulate the world around it be anything like things we recognise above (tool, making, humour, language, curiosity)? Can't we use our so-called intelligence (and creativity) to imagine some AIs that are really radically different? Where might we find such ideas? I'm thinking, perhaps, in mythology - however, a lot of gods are only human - indeed, tales from Asgard, from Mount Olympus, from the Vedic era, smack of soap operas.

Perhaps from more pointless activities (what is Art for, after all)? or from the emergent murmuration properties of innumerable spelling bees? Perhaps the closest to what I'm talking about are the eccentrics (Ships) in Iain Banks' Culture novels...

And its most likely that a real AI would not represent any existential threat to humans, as we are doing a good enough job of that ourselves without any help from smart machines. The only hope would be that AI humour would rest in some jape like saving the human race from itself, just for the lolz.

Thursday, November 08, 2018

The Chinese Remaindered Book Club - the CRBC

In this club, members can only borrow a book for 5 minutes and must then pass it along.
After reading a month's worth of books, members come together and try to reconstruct the
books they have read from these snippets.

The books are all acquired for free from the remaindered books lists of low-ranking publishers.

Incredibly, the reviews written by the CRBC are amongst the most highly sought examples of literary criticism in the history of the world,. offering deep insight into plot, character, authorial quirks, the influences and cultural norms and divergences of the life and times in which the book was written and set.

Except that none of the books described can be found anywhere in any library or book shop, nor are the publishers to be discovered at their addresses. The ISBNs appear to have never been issued, and serious investigations at the Library of Congress and Bodlean have come to nothing.

Membership of the CRBC is by invitation only, and the names of the current active participants is a closely guarded  secret.

Attempts to ghost write the books based purely on the reviews have led to despair and ridicule.

Monday, November 05, 2018

evidence-resistent policy - four thoughts

I read the Professor R.D.French's provocative article on four schools of thought on evidence-based policy, which did not seem to be terribly optimistic about our chances as scientists to influence policy makers (and at some level, may even have cast doubt on whether we should try).

It prompted me to think these even more negative thoughts about policy making itself, which might suggest that people that accumulate evidence, and wish to have an effect ("impact"), may just do better to look elsewhere - as with French's article, I group these four thoughts under for R's

Redundant - perhaps the tension between consideration of politics (popularity/fear etc) and evidence is such that it cancels out, and policy is made that essentially has no effect - an example of this was when presented with evidence of how to improve road journey travel, a policy maker said "oh no, if we do that that way, we'll be accused of spending too much money, and if it doesn't work we'll be accused of wasting money" so policy was to do it a way that cost little and did nothing.

Reactive - a lot of policy may just be a response to de facto behaviour - examples here might include some countries legalising of THC, where the alternative was to continue with the ludicrous position that half of the population were criminals. here policy makers are responsive followers, rather than leaders (which may also be the case above too).

Resignation - a decision has been made and can't be changed, despite al the evidence, even evidence that the population doesn't like the decision any more (think "invade Iraq" or "leave Europe" or Thatcher ignoring civil servant advice on the poll tax - flying in the face of policy because one is resigned to the decision is a suitable path to resignation in the other sense of the word:-)

Reality - some writing on EBP says that scientists should not be so arrogant. However, nuclear war, climate change, anti-biotic resistant bugs - these are very different matters from austerity and keynsian economics or fake news and brexit - to get these in proportion, they're potentially total extinction events. of course, there may be some of the same schmoozing and influence necessary to get some significant change of direction, but in my view we need something more fundamental like a shift away from capitalism (I didn't say socialism), since there's no evidence there are any ways to connect with longer time scales in the current scheme of things. And evidence suggests we must, and policy makers show little sign of understanding that these are game changers.



p.s. after discussion with Ian, lets note that natural & life sciences have probably had more success influencing policy, and perhaps the resistence is highest for input from social & economic sciences, where the evidence may more in dispute, or even the entire methodology or disciple in some doubt (recent results on lack of reproducibility in psychology, or lack of predictive value in macro-economics appear to lend credence to the policy makers' distrust of academic advice, for example, in those realms).

Saturday, October 27, 2018

an alphabet of data

who ate my data
who bartered my data
who curated my data
who deleted my data
who exfiltrated my data
who faked my data
who googled data
who hoarded my data
who imitated my data
who japed my data
who kyboshed my data
who lambasted my data
who mumbled my data
who nimbied my data
who ogled my data
who perfumed my data
who quarantined my data
who rationed my data
who stole my data
who traded my data
who undeleted my data
who valued my data
who whistleblew my data
who xeroxed my data
who zeroed my data

Wednesday, October 24, 2018

brief review of proof copy of The Age of Surveillance Capitalism by Shoshana Zuboff

I was sent an unbounded proof copy of this book by an editor at the publisher, so what you get in a bookshop may not be quite the same.

tl;dr This is like Jaron Lanier with teeth.
i.e. if you read You are not a gadget, and then the polemic,
"Ten Arguments For Deleting Your Social Media Accounts Right Now", or you want to go back in time, have read Jerry Mander's "Four Arguments for the Elimination of Television", you have a head start.

However, you also have to have read Hannah Arendt's revision/update work on the cycle of 0wnership by capitalism, and much much more. The chapters are also all prefaced with fine quotes from Auden (except a couple of outliers with Leonard Cohen and Hildebrand) from sonnets from China.

The take home (perhaps) is the coup from above, where surveillance capitalism is (perhaps) distinguished from any other capitalism by 3 extraordinary properties:

1. a position of extreme privelege (why do those old rules apply to our shiny bright new stuff?)
2. massive asymmetry of agency, and necessarily also of legibility and status in any negotiation.
3. disregard for democracy in the deepest sense (there is no demos)

The book is largely descriptive, but has a mass of detail, reminding me of a (more readable) digital world version of Piketty's Capital in the 21st century, although he also had some modest proposals for remedy/redress, which are still possibly not out of reach, whereas this work seems somewhat more pessimistic, although I need to read it again and see if the seeds of surveillance capitalism's destruction are contained within.

Monday, October 22, 2018

emergent morality


There's old work from Kropotkin from observing animal behaviour and seeing both coopeation, and sacrifice repeated exposure that suggested (without invoking any magic/superstition) that, while the gene may be selfish, that isn't all there is to society.

A simple one-shot game theoretic approach doesn't deal with this, so people moved on to iterated games, and most famously (at least from my reading) showed that the prisoner's dilemma is not a dilemma at all when you consider multiple iterations (repeat offenders learn to "trust" one another).

At a fancy level, this is sometimes ascribed to a theory of mind, where "you think that I think that you think that I think...so lets call the whole thing off" - actually, this is a short cut - you don't need a theory of mind to explain cooperative strategies in dumb animals-  you just need a population carrying out the iterative procedure - the cooperative strategy has higher survival value over the multiple encounters and multiple individuals. What "empathy" does, is simply allow planning, so you don't have to go through all these iterations to learn the better soluton - you just imagine them. So instead of being dead like both Iago and Othello, or Macbeth and his wife, you choose life.

There are exceptions - the lone indvidual making a single encounter is incented to renege on this. In social terms, this is why villages distrusted travellers - they know that the traveller is going to try and game their trust and not have to put up with any tit-for-tat strategy, as they will be long gone before the second screen. I'm wondering if this also explinas why people get more "conservative" as they get older - they have less to lose as they approach death in terms of retaliation or exclusion.

So this is explored quantitatively in some interesting real world scenarios in this paper on what I'm calling  emergent morality

Now just how does this get adopted (or rejected) as a social norm/ethic? Well, we need to run the population dynamics together with some model of encounters - how many people are a) only going to meet a group just once or b) going to exit the game (i.e. die) real soon?

This would then give us a (stable?) distribution of cooperative versus selfish behaviour. Note here when I say "selfish", I mean rational selfish in a short term way - the cooperative players are also rational selfish, but in a longer term sense (they iteratate, whether really or imaginatively)

We can then extend this to include a small number of mutants (bad apples) that engage in Byzantine behaviour (Loki, disrupters etc). And then we could use this to design mechanisms for society that lead to fair collective outcomes (aka maximise social welfare) despite some fraction of selfish, and some (usually small) fraction of byzantine players. Such algorithms exist (see the literature on BAR Fault Tolerance for Cooperative Services ) but assumes that you "just" use the altruistic players to improve performance of the system designed for selfish/rational and byzantine/bad nodes) -

I'm thinking more, how do you build such algorithms for systems like Wikipedia, or social media content moderation, or even liquid, full online democracy?


Wednesday, June 20, 2018

26 connections to everywhere

why do "clients" connect to servers? given we know
1/ everyone, pretty much, is connected to amazon, facebook, google, twitter,  youtube etc,
and
2/ the IPv4 address space is full up
why not just have the cloud call every address?

we know  there will never be more than 26 services in the world, (because currently, you need autocomplete asap in the browser on any url you type, and most people sadly use english, which means most web sites are shackled to the roman alphabet - hint to any startup there you won't make it if you can't find a spare first letter:)

so then we could just have a well known port for each of these 26 services, and punch a hole in the NATs in the world for those well known ports (or if paranoid, restrict it to source IP prefixes in the owned space of amazon, facebook, google, twitter, youtube etc)

resistance is futile, impedance is high, time for a volte-face?

Sunday, June 17, 2018

truth drains away

You know about making sure you've turned off all the taps in your home when you go away? water, gas, and even electricity (some devices may be on standby, but can wakeup and catch fire - of course, if you were clever-lucky, you'd have a flood from a leaky faucet at the same time which would put out the fire. but if you were dumb-sad, the fire would ignite the gas leak and blow up your lovely home).

For the same reason, we should go around making sure all the digital devices (TV, Voice assistant, Smart Toiler, laptop, tablet, phone, watch, fitness widgets) are all turned off, whenever they are not in use.

The reason is that these devices contain truths. However, in today's Internet, truths drain away.
Gradually, truth leaks out of the device, leaving behind only fake news and advertisements  for things you don't want (all the more so, because you aren't there then anyhow).

The rise of falsehoods on the Internet is our own fault - it is because we are careless with our personal truth - all our 1s are turned into one great big zero. or worse, the sign bit is set and negativity overwealms us all. The value of the leaky Internet is minus the square of the number of leaky users.
for 2 leaky users, the net worth is -4, for 1000 users, its -1000000. if half the users stopped being leaky, the net would be worth nothing. if just two thirds of the billions of users were to stop being leaky, the net would be actually quite useful once more.

its just plumb good sense.

Friday, June 01, 2018

data is the new tobacco

google famously said "don't be evil". i've now repeatedly heard people from national intelligence agencies say, in public, "everything we do is lawful". one analytics company said that they make ethical decisions on who they won't work for but provide s/w for the above.

just who are they kidding, apart from themselves?

as i mentioned previously, someone at a recent event on ethics and privacy enhancing etchnologies said (under chatham house rule, so i can't attribute it)
"Data is the new fur", and at another meeting (same constraint)
"I can't see why mum's are proud of their kids going to work for Alpabet Soup, they wouldn't be like that if it was a tobacco company".

Either of these would do to get some sort of re-adjustment of the public enhagement such as this report from the RSA with all the above...

fur or baccy, choose your poison...

at one of the events, we ended up with one of the organisers saying "the problem here is capitalism. and patriarchy...."

oops

Thursday, May 31, 2018

digital person(ae)

regarding decentralised, fair analytics?
https://www.eventbrite.co.uk/e/the-digital-person-a-symposium-tickets-43050737916

some possible discussion/questions

1. who can proxy for a hub in the home, for great grandfather...the bank, the kids, the bbc, the GP, all of the above...(see 4)...

2. price discrimination v. differentiation - do we need "cloud neutrality"

3. how near to privacy/security/utility tradeoff curve are we in practice in central v. decentralised cloud/analytics?

4. what about identity systems? are we ready for multiple pseudonyms each with a subset of our attributes (am-over-18, or am-a-citizen of country x) instead of centralised id with everything?

5. who will power the infrastructure when its completely decentralised?
we're a long way from microgeneration...

6. in edge ai, what are the distributed analytics _coordination_ challenges

7. in edge ai, what are the distributed analytics _privacy_ (diff?) challenges

8. how do we get assurance (sousveillance/someone-elses-pov dashboard) in the decentralised world?

Tuesday, May 29, 2018

edge to edge bogus arguments in systems design

since the arrival of blockchain tech, we're seeing a lot of bandwagonning on
edge-computing.

what most the pundits pushing for this is "just-inside-the-edge" computer/storage/services -

i.e. its still owned by network providers or co-lo kit from a good-old-fashioned-cloud service - same-old same-old. it is getting lower latency/higher availability, less backhaul network costs and (possibly) able to localize service behaviour to geographic jurisdiction, which are all ok things to do.

but it isn't p2p.

but it isn't end-to-end.

e2e was/is the liberating architectural feature of the net that lets anyone run a service. that lets value scale super-linearly (between n*ln(n) and n^2 depending who you believe).

p2p was a failed tech predicating on everyone running things e2e in their home, pocket/car. it failed because of three barriers
i) asymmetric capacity in access networks - this is hard to blame on anyone - its a feature of using old copper capacity and how shared medium spectrum works for fixed and wireless broadband. slowly, it is becomine less the case (last time i looked, 10M out of 35M households in the UK had fiber, which doesn't have these constraints.
ii) IPv4 address space depletion leading to being NATted to death, rather than deploying IPv6 (or anything else).
iii) software deficiencies leading to patriarchical firewalling of systems with vulnerabilities, rather than fixing the root cause (poor systems security).
iv) add yours here....

if you don't run the ledger, file service, social network, messenger platform in your home/pcket/car, it isn't end2end. if it isn't p2p, it isn't e2e. if it isn't e2e, its still 0wned by someone else. even if you have a spare set of keys.

Friday, April 27, 2018

Quantum Computing and Quants and the Turing Institute's mission

so people are afraid of quantum computing.
people should be far more afraid of today's algorithmic trading systems.
the world is run by a bunch of computer programs which have never been scrutinized - they might just each be nonsense, but in combination, even if they are all very correct, they certainly create nonsense. The fiction that the world is running on some financial fantasy academic structure known as a market is a wonder to behold - essentially, the sustained collective delusion in the face of obvious massive fractural objections (american exceptionalism:  why is the dollar worth what it is, aside from military might? protected markets inside China; split level economies like Brazil and so on) - all this is like the medieval world where the SMith would charge one price to shoe the farmer's horse, and another (massively higher) price to shoe the Lord of the Manor's horse - why? and why didn't anyone run arbitrage on this? because society and people don't mix with the idea of money really at all well.

So a lot of work in machine learning and AI and finance purports to address problems like money laundering and fraud and so forth. And yet we live in a world where the whole operation of existing algorithms is based on a false belief, that they operate in a market. Many of them operate in the casino that is the stock market, which is even more divorced from reality than the rest of the economy. Here algorithms are in an arms race, and yet it is an odd arms race, as unlike warfare i nconventional battlefields where we can pick apart the guns and planes of the other side from time to time and take their ideas, or at least reverse engineer them, the algorithmic race in the stock market is too complex and too fast to allow this - code just interacts via the symptomatic (observed) behaviour of the system, and rarely if ever directly interacts with other code. Most weird.

SO the real first duty of ML&AI in the world of finance should be to expose and fix these structural problems - first, to model the worlds economy properly (e.g at least as well as people like Piketty, but more so, dynamically), and to build a sound system for running investments without casinos like the stock market, but also without fantasies like Adam Smith's invisible hand.

So what has this to do with Quantum Computers[1]?

Well, QC promises to run some new algorithms in a new way - there aren't any signs of a full working QC piece of hardware yet, and there are precious few actual algorithms[2] so far, but one in particular has grabbed a lot of attention, which is the possibility to factorize numbers super fast compared with good old fashioned von-Neumann computers (faster even parallel and distributed vN machines) - this is due to the qualitatively different way that QC hardware works (highly parallel exploration of state spaces, scaling exponentially with the number of Qbits).

So what does this threaten?

Well, it threatens to break cryptograpy, which means our privacy technologies for storing and transmitting (and possibly even securely processing[3]) data are at risk. Bad guys (or just curious people) will be able to see our secrets.

So two thoughts
A) why can't we just devise new QC encryption algorithms, which just moves the arms race along in the usual way (a million bit keys for example, or something really new we havnt thought of yet)? Then we are back to the same old normal world where most data breaches will be because of social engineering or stupidity and self inflicted (minister leaves unencrypted USB stick on the bus) wounds.
B) Maybe we get more cautious as a whole and just don't send stuff around so glibly or provide remote access to our computers so readily. Maybe access control and authentication and just implementing least privlege properly could work most of the time, and the whole idea of crypto saving us was a chimera and a delusion, just like the whole idea of the market was a snare and a delusion?

my 2 q-cents.

j.
1. not to be confused with quantum communication where we use entanglement just to detect eavesdroppers - a perfectly sound, existing tech with a very narrow (point to point, not end-to-end) domain of usefulness.
2. shor's algo for example. one puzzling thing is that we had hundreds of algorithms for von-neumann style computers before we had any working hardware. why is it so hard to conceive of algorithms for QC? seems like it is a poor match for how humans are able to express methods for solving problems (which are many, and varied, but don't seem to fit ensemble/state space exploration, except perhaps in MCMC:-)
3. eg.. homomorphic crypto - possibly also at risk from QC, although re-applying ideas like garbled circuits to a QC machine shouldn't be too hard:-)

Tuesday, March 27, 2018

How science progresses - falsifiable, probably or paradigm shift, likely?

Reading Staley's excellent introduction to the philosophy of science was reminded of reading Popper's Objective knowledge back in the 1970s, but now I'm a recovering Bayesian, and am immersed in social science explanations like the structure of scientific revolutions by Kuhn, or even the whole idea of funding/groupthink/paradigms, I'm now convinced we don't have a good basis for choosing the right description of the process (or classifying best practice) until we study the past, both its pre- and post- states - i'm thinking that people choose to run occam/popper after they intuit a new paradigm shift (e.g. copernican model of planets) and use some confidence models to decide that, when the new theory has objectors, the objectors are outliers, whereas the old new outliers the new theory explains were more important than the new old outliers - of course, the new theory can still be wrong, but the smart money is that it isn't...

Tuesday, February 06, 2018

what if you were the only real person in the world

Anyone read Theodore Sturgeon's fabulous short story It Wasn't Syzygy?

Trying to wean people off facebook by creating an alternative (e.g. advert free, subscription, but open to link to other platforms) system, so everyone always starts by saying "you can't beat the network effect". so at what scale does this network effect magically become unbeatable? for example, the web has beaten TV even though TV had a billion users. Metcalfe claimed the value was n-squared, others have toned that down tho n log(n), but i think it's ignoring the _negative_ contribution level from spam/phish/troll/advert/attention grabbing, which inevitably grows with the network, but usually, over time, faster in the end. so here's my proposal anyhow: we invite you to our new net which has "everyone" in your network on it, but initially, your friends are al just bots emulating your real friends they make you feel at home there. now you tell your real friends about your safe new, ad free social net, and as they join, they replace the avatar/bot of them (a bit like the opposite of the stepford wives). oh, did I forget to tell you. we already did it. No, really, We didn't have to do it, they are doing it to themselves - c.f. Dr Wu's fine book on the attention merchants...

+
 so in fact we can model this from the fact that the network is directional, and end points (humans) take more time to create new content than to consume (new to them) content - so even if we aren't all couch potatoes, this asymmetry in creativity versus consumption means that the network will tip from peer-to-peer, to being dominated by a small number of producers and a large number of consumers - the cost of creation will drive the quality of creation down, but the quantity up (to keep it new - well known to pornographers for example - c.f. https://yourbrainonporn.com/)

Tuesday, October 31, 2017

declining big data, colour me rimbaud

first you beg for data
then you bag the data
then you de-bug the data
then you get bog-ged down in the data
finally, you big up the data

so what's new in this , i beseech you from the depth of my vowels?

Blog Archive

About Me

My photo
misery me, there is a floccipaucinihilipilification (*) of chronsynclastic infundibuli in these parts and I must therefore refer you to frank zappa instead, and go home