• Progress with PayGap.ie

    I’ve made some great connections in the last few months with my PayGap.ie portal, and wanted to share some of what I’ve been doing.

    I visited the Geary Institute of Public Policy to speak at one of their lunchtime seminars about the portal, and following this, they invited me to write a short paper for PublicPolicy.ie. That paper outlined the state of reporting so far, what I’ve learned, and suggested some policy improvements that could be made.

    I visited Phoenix FM to talk to them about the portal, and also gave them a brief statement on the announcement of a government portal (at long last!)

    There’s more coming, and I’m very grateful for the opportunities I’ve had to share my work with others so far!

  • New Gender Pay Gap Portal

    New Gender Pay Gap Portal

    If you’ve been on my site before, chances are you’ve seen the link in the menu to the 2022 gender pay gap database. When I set up this page on my website last year, I had to work around a number of tricky limitations and ended up having to insert custom code into WordPress to display the data the way I wanted to. There wasn’t a plugin (or three) that I could use to load the data from a database and allow for it to be searched, updated, etc. easily.

    I found some ways to make it work for the first year, but then the government announced that they would not actually have a portal for 2023 either, and that they didn’t have a clear date for when there would be a portal available. I started to try and expand my existing setup, and ran into many of the same issues as last year, compounded by the fact that I was trying to manage and display two different sets of data. I kept running into the fact that it would just be better to build something more custom. Over the last week or so I’ve spent time doing just that. It’s all the same data, displayed as it was before in searchable sortable tables, but with a number of improvements.

    The new site has the full 2022 database, and I’m building the 2023 database at the moment. There’s also a form where you can submit reports with their info directly, to help me build 2023’s dataset more quickly (and any future datasets too).

    This new sub-site will let me keep expanding the dataset, and hopefully expand to include some visualisations of data, comparisons, etc.

    If you appreciate what I’m doing here and want to help, one of the best ways you can do so is to use the form on the new site and submit links to any company’s gender pay gap report, and ideally pull out the headline figures for me.

    Watch this space for updates as the 2023 database grows!

  • What we owe to each other

    I attended a talk recently about migraine, and included in the talk was a demo and a quick blurb about a new-ish medical device to potentially treat migraine and other headache conditions (an external vagus nerve stimulation device, for the curious). It seems an interesting development, since previous incarnations of the same required surgery and an implanted device, but when I did a little more investigating I was disappointed to discover that the device operates on a subscription model. Every 93 days, you have to buy a new card to “activate” the device and make it work for another block of time. It’s not to do with the monitoring of a patient, or wanting a clinical touchpoint every 3 months or so (because you can also opt for the very expensive “36 months in one go” option), it is simply a business model – sell a device that becomes an expensive paperweight if a subscription is not maintained.

    Over the last few days, it has prompted me to think about the landscape we are building for ourselves – one populated with smart devices, subscription devices, as well as an increasing cohort of medical devices – what it will look like in the future, what we owe to customers of these devices if we are involved in making them, and ultimately, what we owe to each other.

    Subscription Business Models

    Subscription-based business models are nothing new – chances are you’re paying for at least one subscription service yourself. For many businesses they are an attractive choice as they mean a continuous revenue stream, a constant inflow of cash you can plan around, rather than one big bang purchase and then a drought. And lots of people are fine with paying for subscription models, even if they don’t love them, but what if we’re talking about more than just streaming music or paying monthly for Photoshop? What if instead of software or an intangible thing, we’re talking about physical devices?

    Physical devices with a subscription model aren’t exactly new, and they’ve had their problems – Peloton came under fire in 2021 after it seemed to release an update that forced a subscription onto its users and rendered its treadmills useless without one. BMW were recently the subject of much negative press for their subscription model heated seats – shipping cars with the physical equipment needed to heat the seats, but locking it behind a subscription paywall. And HP Instant Ink subscribers found that once they cancelled the Instant Ink service, the ink left in their cartridges stopped working, even though it was still sitting their in their printers.

    This is all very annoying, but mostly you could argue the above are luxuries – your seats aren’t heated, your day still goes on. But these are not the only kinds of devices that, increasingly, are coming with subscriptions.

    What happens when your bionic eyes stop working?

    The merging of technology and medicine is, to a certain extent, inevitable. People have unofficially relied on technology to supplement and assist with medical issues for a long time now (such as those with diabetes hacking pumps to essentially make artificial pancreas, a process known as looping, or people with vision impairments using apps to see through the camera, receiving audio descriptions), and as time goes on, manufacturers are joining the market with “official” solutions. There is huge potential to make lives better with assistive technologies, by automating processes that were manual or artificially replacing senses to name just two examples. Often these developments have been lauded as the “way of the future” and a huge step forward for humanity, but what happens when the initial shine passes?

    A CNN article from 2009 speaks about Barbara Campbell, a woman who was diagnosed with retinitis pigmentosa – a condition which gradually robbed her of her sight. In 2009, she was participating in an FDA approved study of an artificial retina – a technological solution to her impaired vision, a microchip to help her see again by stimulating the retina electrically in the way that light should be. Combined with a pair of sunglasses and a camera to capture the world around her, the devices allowed her to see again, with her interpretation of the new signals improving all the time. By all accounts, it’s a dream scenario – technology that is really doing good and changing someone’s life for the better.

    Now, in 2022, things have changed. In 2020, the company that manufactured these implants had financial difficulty. Their CEO left the company, employees were laid off, and when asked about their ongoing support, Second Sight told IEEE Spectrum that the layoffs meant it “was unable to continue the previous level of support and communication for Argus II centers and users.” Around 350 patients worldwide have some form of Second Sight’s implants, and as the company wound down operations, it told none of them. A limited supply of VPU (video processing units) and glasses are available for repairs or replacements, and when those are gone, patients are figuratively and literally in the dark.

    Barbara Campbell was in a NYC subway station changing trains when her implant beeped three times, and then stopped working for good.

    Now patients are left with incredibly difficult decisions. Do they continue to rely on a technology which changed their lives but which has been deemed obsolete by the company, that may cause problems with procedures such as MRIs, with no support or repair going forward? Or do they undergo a potentially painful surgery to remove the devices, accruing more medical costs and removing what sight they have gained? Do they wait until the implant fails to remove it, or do they remove it now, gambling on whether it might continue working for many years? Do they walk around for the rest of their lives with obsolete, non-functional technology implanted in them, waiting for the day it fails and replacement parts can no longer be found?

    Meanwhile, Second Sight has moved on, promising to invest in continuing medical trials for Orion, their new brain implant (also to restore vision) for which it received NIH funding. Second Sight are also proposing a merger with an biopharmaceutical company called Nano Precision Medical (NPM). None of Second Sight’s executives will be on the leadership team of the new company. Will those who participated in the Orion trials to date continue to receive support in the future, or even after this merger?

    IEEE Spectrum have written a comprehensive and damning article examining the paths taken by Second Sight, piecing together the story though talking to patients, former doctors and employees, etc. and although it’s clear that Strickland and Harris know more about this than anyone, even they can’t get a good answer from the companies about what happens now to those who relied on the technology. Second Sight themselves don’t have a good answer.

    Subscription Paperweights

    Second Sight’s bionic eyes didn’t come with a subscription, but they should have come with a duty of care that meant their patients never had to worry about their sight permanently disappearing due to a bug that no one would ever fix or a wire failing. And while bionic eyes are an extreme example of medical tech, they’re an excellent example of the pitfalls that this new cohort of subscription-locked medical devices may leave patients in.

    Lets return for a moment to the device that started me down this line of thought – the external vagus nerve stimulator. I have had a difficult personal journey with migraine treatment. I have tried many different drug-based therapies, acute and daily preventatives, and have yet to find one which has been particularly effective or that didn’t come with intolerable side effects. I am now at a point where the options available to me are specialist meds available only with neurologist consults and special forms and exceptions, or the new and exciting world of migraine medtech devices. And the idea of a pocketable device that I use maybe twice a day, perhaps again if I am experiencing an attack, is appealing. With fewer side effects, no horrible slow taper off a non-working med, and let’s be honest, a cool cyberpunk vibe, I’m more than willing to give this kind of thing a try. Or I would be, if it wasn’t tied to a subscription model.

    Because I can’t help but think of Barbara, and the day her eyes stopped. And I can’t help but think about how many companies fail, suddenly, overnight, with no grand plan for a future or a graceful wind down. And I can’t help but worry that I might finally find something to sooth this terrible pain in my head, to give me my life back, only to have it all disappear a year down the line because the company fails and I have no one to renew a subscription with, and my wonder-device becomes an expensive and useless piece of e-waste.

    The decision to add a subscription model to a device such as this is financial, not medical. Internal implantable vagus nerve stimulators already exist for migraine and other conditions, and you don’t renew those by tapping a subscription card. This is a decision motivated by revenue stream.

    To whom do you have a duty of care?

    The migraine vagus nerve device is not alone. When I shared the news about this device with a friend, she told me that her consultant had been surprised to find that her smartphone-linked peak flow meter did not have a subscription attached. Subscription medtech devices have become a norm without many noticing, because many people do not (and may never) rely on devices like this to exist.

    The easy argument here is that companies deserve to recoup their expenses – they invested in the devices, in their testing and development, in their production. If the devices are certified by medical testing boards, if they underwent clinical trials, there is significant costs associated with that, and given the potential market share of some of these devices, if they simply sell them for a one-time somewhat affordable cost, they will never see a return on their investment. This, in turn, will discourage others from developing similar devices. And look, it’s hard to refute that because it is true – it is expensive to do all of these things, and a company will find it very hard to continue existing if they simply sell a handful of devices each year and have no returning revenue stream. If this were purely about the numbers, this would be the last you’d hear from me on the topic. But it’s not.

    If you develop a smart plug with a subscription model, and your company fails, this is bad news for smart plug owners, but a replacement exists. A world of replacements, in fact. And the option to simply unplug the smart device and continue without it is easy, and without major consequence. The ethical consequences are low. But developing a medtech device is simply not the same. It is about so much more than the numbers. This is not about whether someone can stream the latest Adele album, this is about ongoing health and in some cases lives, and this is an area of tech that should come with a higher burden than a smart doorbell or a plug.

    When you make any sort of business plan, you’ll consider your investors, your shareholders, perhaps your staff, and certainly your own financial health, but when it comes to medtech, these aren’t the only groups of people to whom you should owe your care and consideration. Your circle of interested parties extends to people who may rely on your device for their health, for their lives, beyond just a simple interest or desire to have a product that works. Simply put, is your duty of care to your shareholders, or to your patients?

    What do we owe to each other

    Do you owe the same duty of care to a smart doorbell owner as to a smart heart monitor owner? Who will face a tougher consequence if your company fails without warning?

    Second Sight took grants and funding to continue developing and trialling their brain implant while they quietly shelved the products they had already implanted in real patients – is this ethical? Is it right? How can anyone trust their new implant, knowing how patients using their previous implant were treated and the position they were left in? And is it right to continue to grant funding to this research?

    Companies right now are developing devices which lock people into a subscription model that will fail if and when the company fails, at a time when we are all concerned about the impact of technology on the environment, conscious of e-waste, and trying to reduce our carbon footprint. They are developing devices that work for 12 months and then must be replaced with new ones. Is it right to develop throwaway medical devices that stop working simply so that you can lock people into a renewing subscription/purchase model?

    It is undeniable that technology can help where current medical options have failed. We have already seen this with devices that are on the market, and with new devices that arrive. We should want to pursue these avenues, to make lives better and easier for those who need help. We should fund these technologies, spur on innovation and development in these areas, and help everyone to reach their fullest potential.

    But we owe it to each other to push for better while we do. To push back on devices that will fail because someone’s payment bounces. To push back on devices that only have subscription models and no option to purchase outright. To push for higher standards of care, better long term support and repair plans which can exist even if the company fails. To push for companies to be held to these standards and more, even if it makes things more difficult for them. And to push companies to keep developing, even with these standards in place, to keep developing even though it is hard.

    We deserve a duty of care that extends not just to lifetime of device, but the lifetime of a patient.

    This isn’t just about home security, or smart lights – this is people’s health, their lives. The duty of care should be higher, the ethical burden stronger. We owe it to each other to not allow this world to become one where vision and pain relief and continued survival depends on whether or not you can pay a subscription.

  • Facial recognition is terrible at recognising faces.

    If you’ve ever used a Snapchat, Instagram, or TikTok filter, you’ve probably used facial recognition technology. It’s the magic that makes it possible for the filters to put the virtual decorations in the right place, it’s why your beauty filter eyeshadow (usually) doesn’t end up on your cheeks.

    It’s fun and cute, but chances are that if you’ve never had any issues with these filters, it’s because your face looks a lot like mine. Unfortunately, the same cannot be said for many other people.

    Why don’t zoom backgrounds work for me?

    This image has an empty alt attribute; its file name is ZoomBackgroundFaceDetect-1024x277.jpeg

    During the pandemic I, along with many other people, used Zoom virtual backgrounds extensively – calm pictures to hide my office background, funny pictures to communicate my current stress levels, you name it. I found they worked fairly well, perhaps struggling a little to handle the edges of my curly hair, but my experience wasn’t universal. When a faculty member asked Colin Madland why the fancy Zoom backgrounds didn’t work for him, it didn’t take too long to debug. Zoom’s facial detection model simply failed to detect his face.

    

    Why can’t I complain about it on twitter?

    This image has an empty alt attribute; its file name is twittercropping-1024x635.png
    On twitter, the longer image I shared was cropped….to include only Colin’s face.

    When Colin tweeted about this experience, he noticed something interesting with Twitter’s auto-cropping for mobile. Twitter crops photos when you view them on a phone, because a large image won’t fit on screen. They developed a smart cropping algorithm which attempted to find the “salient” part of an image, so that it would crop to an interesting part that would encourage users to click and expand, instead of cropping to a random corner which may or may not contain the subject.

    Guess which part of the image Twitter’s algorithm cropped to. Why did Twitter decide that Colin’s face was more “salient” than his colleague’s?

    How does this happen?

    Facial Recognition Technology (FRT) is an example of Narrow Artificial Intelligence (Narrow AI). Narrow AI is an AI which is programmed to perform a single task. This is not Data from Star Trek, not the replicants from Blade Runner, this is more like the drinking bird from The Simpsons which works…until it doesn’t.

    Facial recognition algorithms are trained to do one thing, and one thing only – recognise faces. And the way you train an algorithm to recognise faces (the way you train any of these narrow algorithms) is by showing it a training dataset – a set of images that you know are faces – and telling it that all the images it sees are faces, so it should learn to recognise features in these images as parts of a face. But there isn’t an “intelligence” deciding what a face looks like, and while it is possible to try and help the algorithm by providing a descriptive dataset, it’s not possible to direct this “learning” specifically. The algorithm is a closed box. Once the algorithm has “learned” what a face is, you can test it by showing it images that do, or do not, contain faces, and see if it correctly tells you that there is a face in the image but, crucially, it is still very very difficult to tell what that algorithm is using to decide if a face is present. Is it looking for a nose? Two eyes? Hair in a certain location?

    Is it even looking at the faces at all?

    Of course it’s looking at the faces to determine if faces are present, right? How else would it do it?

    Well, if you’ve ever done any reading about AI recognition, you’ll undoubtedly be amused by the above statement, because (as I mentioned above) you can’t really specifically direct the “learning” that happens when an algorithm is figuring things out, and in perhaps their most human-like trait, AI algorithms take shortcuts. There’s a well known case of an algorithm trained to determine whether the animal in a photo was a wolf or a husky dog, and by all accounts the algorithm worked very well. Until it didn’t.

    The researchers intentionally did not direct the learning away from what the AI picked up the first time it tried to understand the images, and so what the AI actually learned was that wolves appear on light backgrounds and huskies don’t. It wasn’t even looking at the animals themselves – just the photo background. And thus, once it was tried on a larger dataset, it began incorrectly identifying wolves and dogs based on the background alone.

    This situation might have been constructed just for research, but the hundreds of AI tools developed to help with Covid detection during the pandemic were not, and as you may have already guessed, those algorithms by and large were not the solution to the problem, and instead had the same problem as above.

    A review of thousands of studies of these tools identified just 62 models that were of a quality to be studied, and of those, none were found to be suitable for clinical use due to methodological flaws and/or underlying bias. Because of the datasets used for training these models, they didn’t learn to identify Covid. One dataset commonly used contained paediatric patients, all between age 1-5. The resulting model learned not to identify Covid, but to identify children.

    Other models learned to “predict” Covid based on a person’s position while being x-rayed, because patients scanned lying down were more likely to be seriously ill and the models simply learned to tell which x-rays were taken standing up and which were taken lying down. In some other cases the models learned what fonts each hospital used and fonts from hospitals with more serious caseloads became predictors of Covid.

    And there are so many other incidents of AI failing to recognise something correctly or failing to have some human oversight or context applied that the AIAAAC has an entire repository to catalogue them.

    What has this got to do with Facial Recognition Technology in Ireland?

    Quite a lot. I’ve already shown you that AI models are often not very good at recognising the things they are supposed to be trained to recognise. Much of this is down to the training dataset that you use – the pictures you show your algorithm to teach it how to recognise things.

    You may have already spotted a problem here because, of course, not all faces are alike – some people have facial differences that may mean their face does not have a typical structure, some may have been injured in a way that has changed their face. And, of course, the world contains a wide variety of skin tones. And if you are asking if the kinds of datasets available to train these models contain a diverse set of faces, incorporating all of the above and more, then the answer is a resounding no. Efforts are being made by some groups to address this, but progress is slow.

    And what this boils down to is this: if your training dataset only contains typical faces, then your facial recognition algorithm will only learn to recognise and identify typical faces. If you train your facial detection models using data that isn’t diverse, you will release software that doesn’t work for lots of faces. This is a well known, pre-existing, and long running problem in tech.

    Moreover, if you are not careful with your dataset, you will reinforce existing stereotypes and bias, a problem which is more severe for certain groups, such as women and POC. Models don’t just learn what you want them to learn, they also learn stereotypes like “men don’t like shopping” when they are trained on datasets in which most shoppers are female.

    Can’t we put in some checks and balances, some safeguards?

    Well, yes and no. While people are working to help people improve datasets, there are many datasets out there which contain demographic information which will likely be used to help train and build national or large-scale models, and since this demographic information represents how things currently are, it will also represent how people have been impacted by existing or past biases (e.g. historically poorer parts of a country, or groups of people who have been oppressed and, as a result, have lower rates of certain demographic indicators such as college education or home-owning). It’s hard to escape these biases when they are built into the data because they are still built into our societies.

    Additionally, in order for there to be checks and balances, we would need those in charge to understand the implications of all of the above (and more) and to care enough to enforce them by writing legislation with nuance and skill. This is a complex area that has caused issues in many countries that have tried to adopt some form of FRT.

    We have examples in our own country’s recent history about how our government has legislated for (and cared about) personal and biometric data, and their record is not good. An investigation by the Data Protection Commission found significant and concerning issues with the way the Public Services Card had been implemented, and the scope expanded without oversight or additional safeguards, sharing data between organisations that were never intended to be in the scope of the card. The Commission said, in its findings, that “As new uses of the card have been identified and rolled-up from time to time, it is striking that little or no attempt has been made to revisit the card’s rationale or the legal framework on which it sits, or to consider whether adjustments may be required to safeguards built into the scheme to accommodate new data uses.” The government’s first response to this report was not to adjust course or review the card internally itself, but to appeal this ruling and continue to push the card without any revisiting. (They dropped the appeal quietly in December 2021).

    On this basis, I do not believe that the government in its current form has the capacity or the will to legislate safely for the use of FRT. A first foray into gathering public data en masse resulted in illegal scope creep, extending the card’s reach far beyond what was permitted without any announcement or oversight, and no review or change to the safeguards. This is something that simply cannot be permitted to happen when it comes to the use of facial recognition technology, which has the potential to be infinitely more invasive, undermining rights to privacy and data protection, and (with flawed datasets) potentially leading to profiling of individuals, groups, and areas.

    Facial recognition technology is not fit for purpose. Existing models are not good at recognising a diversity of faces, and are unable to account for the biases built into the datasets that are necessary to train and build them. It cannot be a one-stop solution for enforcing laws.

  • On Gender Quotas

    The Citizens’ Assembly has today voted for a program of reforms on gender equality in Ireland, including some recommendations around extending gender quotas, and ahead of the predictable backlash for gender quotas, I want to share some thoughts on the inevitable “best person for the job” rhetoric.

    A frequent refrain when people mention gender quotas is that it should just be “the best person for the job”, and that gender shouldn’t matter, but the people who make this argument rarely pause to consider or explore the sexist ideal they prop up with this statement. Let’s dig into that now.

    Studies have shown that when people are blinded to gender, the choices they make represent the actual spectrum of gender much more accurately. We see it in jobs, we see it in award nominations, we see it in all aspects of life. Which means that something different is happening when panels aren’t blinded. We’ve seen that panels are affected by unconscious bias, and end up hiring those who look like them, sound like them, etc. And we’ve seen that people who don’t fit the already established “mold” get left out of this process – even before we step into the interview, we have seen that biased algorithms filter out CVs of women and people of colour, and job descriptions discourage applications from underrepresented groups. We face an uphill battle to improve gender equality in hiring.

    Why not just more unconscious bias training?

    So why quotas? Why not just more training? Can’t we just trust that people will address their unconscious biases, or wait until we reach a more balanced representation organically?

    No. We can’t.

    Unconscious bias training remains a controversial topic. When people propose unconscious bias training, it is often met with resistance and mockery, and people questioning whether the training leads to real change. There have been some studies to examine the effect of unconscious bias training and right now, the evidence suggests that while the training does raise awareness of these biases, the overall effect is not translating into significant behavioural changes. It is worthwhile, but it is not enough.

    And so here we are, with mandated gender quotas. Why? Frankly, because for years, you were asked nicely and you ignored it. Many of the studies which show gender bias in hiring are decades old, this is not a new problem, and people have been raising it for a very very long time. Maybe you were given training about why diverse hiring matters, about unconscious bias, and you ignored it or didn’t internalise it enough to action it. Maybe you’ve never examined your job descriptions to see why all of your candidates look the same. So now your hand has to be forced with quotas, because you won’t do it voluntarily, and people should not have to wait ten more lifetimes for you to decide it suits you to make a change.

    But don’t you think it should be the best person?

    If we loop back to our original thesis, that it should always be just “the best person for the job” regardless of gender, I actually agree. It should be the “best” person. But the unspoken part of this is that you are saying that this is currently how things are actually done, that this idea of “best person regardless” is the current status quo. And there, I must firmly disagree.

    When you say “the best person” and imply that that’s what is happening right now, you’re propping up a myth, a status quo that isn’t. The status quo isn’t always hiring the best person, it’s hiring the one you like, and very often, the one you like is the one you match. And when the hiring panel is predominantly old white men, guess who matches them?

    Your status quo is a myth

    When you say “best person for the job” this is the unasked question which shows the problem with your statement: If we currently hire “the best person regardless of gender” then why are all of those best people white men? For decades? Really, not a single other person was better? Honestly? If, at the moment, the best person already always got the job, then why is there still such a lack of diversity in hiring? What is the reason?

    If we currently hire “the best person regardless of gender” then why are all of those best people white men?

    Please, honestly, examine this thing that you are implicitly saying. If you think that right now, we always hire the best person regardless of gender, then you are also saying that the current gender representation everywhere is an accurate reflection of skill and qualification. You are saying the the only bias which exists is one which would cause someone to hire an incompetent woman over a man because “diversity” when a literal embarrassment of riches of evidence shows the very opposite. And if you don’t understand why such a statement might cause me to raise my eyebrows, well you’ve got rather a lot of catching up to do.

    Can we completely eliminate bias from hiring? Maybe not. And maybe not soon. But gender quotas can force us to shine a light on how we currently hire, and make people think outside their current status quo.

    The research

  • Have you ever been told that your name is incorrect?

    Your name is one of the first things you say to people you meet, it is how you present yourself to the world. It is personal and special. But what if you are told that your name is incorrect due to lazy or thoughtless programming every time you try to book an airline ticket, access banking, healthcare, or any number of services online? This is all too often the case for people around the world, due to lazy or inconsiderate programming.

    What’s a fada?

    Diacritical marks are the marks which guide pronunciation, and they appear in numerous languages – if you’re a native english speaker, you might not use them frequently, but they can change not just the sound, but the meaning of a word. A fada changes the word “sean” (meaning old, pronounced shan) to the name Seán (pronounced shawn), changing the a sound to an aw sound. And if it is your given name, then to include the fada is correct. It is as crucial a part of the spelling of your name as any of the letters. Yet, all over the internet, people who try to include the fadas, accents, umlauts (or other diacritical marks) in their names are told their name is incorrect, invalid, or wrong.

    When it comes to including these diacritical marks on online forms, we too often hear the refrain that it’s a “technical issue”, but that doesn’t quite get to the heart of it, and also implies that it is very difficult to fix or perhaps not even possible. That’s not really true though.

    Back when people were first defining how computers would speak to each other, a character set was agreed upon, so that communication would be consistent. This character set was ASCII (American Standard Code for Information Interchange) and due to memory limitations of the time, ASCII could only fit 128 characters. This is enough for all the letters, numbers, and punctuation marks used in english, but not nearly enough to include all of the “special” characters used by other languages (such as a letter with an accent, á). But these characters aren’t special, they are a part of the language, as much as any other character.

    Competing standards and character sets exist and have done for decades now. Character sets such as Unicode support all the characters in different languages. So why aren’t we using them? Well, it’s probably two reasons:

    • Many older systems continue to use ASCII (such as legacy internal systems at banks and airlines) because they were designed when other character sets weren’t available, and many companies are running much older software than you would imagine at the core of their operations
    • Many things, such as databases and development platforms, default to non-inclusive character sets when you install them, and people don’t bother to change them before moving code into production because it doesn’t occur to them, and then it becomes a larger issue to fix because the system is already in use

    I don’t think either of these reasons are a good enough excuse. Legacy systems should be updated, and when you are developing a new system, there is absolutely no good reason to not begin your architecture with support for other languages.

    Irish has the status as the national and first official language of Ireland, and whether or not you speak it frequently, it is a common feature of our road signs, official documents, and yes, our names. And yet Irish people have had to battle national/state bodies for refusing to accept fadas in their names, and our own Data Protection Commission has decided decided against them. Gearóidín McEvoy points out that fadas aren’t exactly a new invention, so why should we have to fight for their inclusion?

    Your name is too long, too short

    How long is too long? And how can a name even be too long? Well, if you’re going to take part in a census in Ireland, you might be surprised to find how short the space is for a name on the form. The sample form for the 2016 census is available here and you can see that there is space for just 22 characters, including any spaces you might need. If you have a long name, you’re out of luck. And this is far from just an Irish problem. In Hawaii, Janice “Lokelani” Keihanaikukauakahihulihe’ekahaunaele had to fight the government to have her full name displayed on her official ID cards, and she spoke of her dismay at her name being treated like “mumbo-jumbo” and the disrespect she felt when a policeman told her to change her name back to her maiden name to have it fit on her license.

    Patrick McKenzie lives in Japan, and forms there accommodate typically Japanese names, but with 8 characters in his surname, and most Japanese surnames rarely exceeding 4 characters, Patrick routinely can’t fill in his name properly. Inspired by this, Patrick has also written a blog which notes falsehoods that programmers believe about names, which I highly recommend you read.

    I have also known friends with shorter surnames (e.g. two character surnames) to have significant difficulties with online forms, with their legal surname declared “too short”.

    The reality is that, particularly when you think globally, there is no “too short” or “too long” surname, and arbitrary character limits on form fields cause unnecessary difficulties for people who have to butcher their name to make it fit, and then face questioning from others when the name on the ticket doesn’t match the name they put into the form.

    First name and last name please

    If you have ever filled out an online form, chances are you have been asked to split your name, filling out your first name in one box, and your surname/second name/last name in a second box. But what if that is not how your name is structured? Around the world, names are structured in a number of ways that far exceed the constraint of “first” and “last” name. Many countries have names that contain multiple family names, part of a mother or father’s name, different endings depending on the sex of the child being named, etc. Moreover, the idea of a “first” name simply doesn’t translate to a number of cultures, who order parts of the name differently as a matter of course, or depending on the situation. For example, in the Chinese name Mao Ze Dong, the “first” piece of this name (reading left to right) appears to be Mao, but this is in fact the family name. Dong is the given or “first” name.

    The W3 has an excellent blog which discusses the issues with forms and personal names, which includes a number of clear examples of the ways in which the idea of a “first” name breaks down, and it should be mandatory reading for anyone who is designing a form. They note a key question that form designers should be asking themselves before writing a single line of code – do you actually need to have separate fields for given name and family name? If so, why? Could you not simply ask for a user’s full name as they would typically write it, and then if necessary, ask what they prefer to be called in communications so that you can still auto-populate your “Hey <name>” email?

    Inclusive form fundamentals

    A multitude of online forms fail to support diacritical marks, or declare that names are too short or too long based on simple biases and the incorrect assumption that everyone has a first and last name that looks like our own (or considers their name in terms of first and last).

    Beyond the frustration that this causes people, it is also dehumanising, insulting, and demeaning. Instead of telling the person “sorry, our system doesn’t handle this and that is our fault” the error messages tell people that their name is wrong, that they are wrong. It tells them they don’t know how to spell their name, or that their name is invalid. It makes people feel like their name or their culture is disrespected. It underlines the idea that this system is not built with everyone in mind, just with people who look like those who built it.

    It presents a barrier for someone every time they use your system, every time they are told they are wrong. It is an unfriendly user experience that turns users away.

    It might require extra work in development, to retrofit existing systems to support extra characters, or to ensure that inputs are validated so that special characters are processed and stored correctly in the underlying databases, but the alternative is unacceptable. The time to begin this work is long overdue.

    Your name is not invalid, our form is.

    Key points

    • Inclusive form design makes your product better
    • Inclusive error messages should focus on the system, not the user – if your system can’t handle a character, the character is not invalid, your system needs to be improved.
    • Not everyone considers their name in terms as simple as “first” and “last”
    • And you should ask yourself if you even need a name split this way, or are you just defaulting to the forms you recognise from elsewhere?
    • Special characters should be supported from the very beginning. They aren’t an edge case, they are critical.

  • Actually inclusive engineering

    I want to talk about ethics, diversity, and inclusion in engineering, how we often miss the mark, the impact that has, and the changes we can make to truly bring change from the inside out. My goal is to explain why this is important, and show you some examples where a simple decision resulted in a barrier for someone.

    Why does this matter? Why is it important to be thinking about ethics when we’re developing software? Because software (apps, websites, etc) is becoming the fabric of society – increasingly it is involved in everything we do, from shopping for groceries to online banking to socialising. There is very little in our lives now that is not touched, in some way, by software.

    As we integrate software into more and more areas of our lives, we are also increasingly turning to automated and predictive solutions to perform tasks that were once manual. We are asking computers to do “human” things, more open-ended “thinking” tasks, but computers aren’t human. Most people, when they think of AI, think of something like Data from Star Trek. The reality of AI however is that we have “narrow” AI – models which are trained to do a specific thing and that thing only. These models are unable to add context to their decisions, to take additional factors into account if they are not in the data model, or even to question their own biases. It takes in data, and returns data.

    Lastly, we often spend a lot of time discussing how we will implement something, but perhaps not as much time discussing whether we should implement something. We know that it is possible to build software which will have in-app purchases, and that it’s possible to incentivise those in-app purchases so that they are very attractive to app users. We have seen that it is possible for people to target this marketing towards children – answering the “can”, but not addressing the “should we?”

    When I say we should consider the “should” rather than the “can”, what do I really mean? I’m going to show some real world examples where decisions made during product design ripple out into the world with negative effects. In each of these examples, there probably wasn’t malicious intent, but the result is the same – a barrier for an individual. Most of these examples are not due to programming errors, but by (poor) design.

    Have you ever accidentally subscribed to Amazon Prime?

    Do you know what a dark UX pattern is? You’ve probably encountered one, even if you’ve never heard the term. Have you ever accidentally opted-in to something you meant to deselect, found an extra charge on a bill that you didn’t even realise you had signed up for? Have you ever tried to cancel a service, only to discover that the button to “cancel” is hidden below confusing text, or the button that looks like a cancel button actually signs you up for even longer? How about accidentally signing up to Amazon Prime when you just wanted to order a book? These are dark UX patterns – design changes that are designed to trick the user. They can be beneficial for the person who implements them, but usually to the detriment of the user. In the image above, we see two buttons to add your tickets to the basked. An optional donation can be added with the tickets, but the option to add without donation is much harder to read. It also points backwards, implying visually that this would bring you back a step. Is the value of this donation worth the confusion? Is this ethical? Should a donation be opt-out or opt-in?

    Have you ever been told that your name is incorrect?

    Your name is one of the first things you say to people you meet, it is how you present yourself to the world. It is personal and special. But what if you are told that your name is incorrect due to lazy or thoughtless programming every time you try to book an airline ticket, access banking, healthcare, or any number of services online? A multitude of online forms fail to support diacritical marks, or declare that names are too short or too long based on simple biases and the incorrect assumption that everyone has a first and last name that looks like our own. Instead, we should be asking – do we need to separate people’s names? Why do you need a “first” and “last” name? Could we simply have a field which would accommodate a user’s name, whatever form that takes, and then another which asks what they prefer to be called?

    Let’s talk about everyday barriers

    We’ve never been more aware of handwashing, and a lot of places are using automatic soap or hand sanitiser dispensers to ensure that people stay safe without having to touch surfaces. But what if they don’t work for you? Many soap dispensers use near-infrared technology, which sends out invisible light from an infrared LED bulb for hands to reflect the light back to a sensor. The reason the soap doesn’t just spill out all day is because the hand acts to bounce back the light and close the circuit, activating the soap dispenser. If your hand has darker skin, and actually absorbs that light instead, then the sensor will never trigger. Who tests these dispensers? Did a diverse team develop these or consider their installation?

    Why don’t zoom backgrounds work for me?

    If you’re like me, you’ve been using meeting backgrounds either to have some fun or to hide untidy mixed working spaces while adapting to working from home during this past year. When a faculty member asked Colin Madland why the fancy Zoom backgrounds didn’t work for him, it didn’t take too long to debug. Zoom’s facial detection model simply failed to detect his face. If you train your facial detection models using data that isn’t diverse, you will release software that doesn’t work for lots of faces. This is a long running problem in tech, and companies are just not addressing it.

    Why can’t I complain about it on twitter?

    On twitter, the longer image I shared was cropped….to include only Colin’s face.

    When Colin tweeted about this experience, he noticed something interesting with Twitter’s auto-cropping for mobile. Twitter crops photos when you view them on a phone, because a large image won’t fit on screen. They developed a smart cropping algorithm which attempted to find the “salient” part of an image, so that it would crop to an interesting part that would encourage users to click and expand, instead of cropping to a random corner which may or may not contain the subject. Why did twitter decide that Colin’s face was more “salient” than his colleague’s? It could be down to the training data for their model, once again – they used a dataset of eye tracking information,  training their model to look for the kinds of things in an image that people look at when they look at an image. Were the photos tested diverse? Were the participants diverse? Do people just track to “bright” things on a screen. It certainly seems there was a gap and the end result is insulting. Users tested the algorithm too, placing white and black faces on opposite ends of an image to see how twitter would crop them. The results speak for themselves. Twitter said they tested for bias before shipping the model….but how?

    This impacts more than social media. It could impact your health

    Pulse oximeters measure oxygen saturation. If you’ve ever stayed in a hospital chances are you’ve had one clamped to your finger. They use light penetration to measure oxygen saturation, and they often do not work as well on darker skin. This has come to particular prominence during the pandemic, because hospitals overwhelmed with patients started spotting differences in oxygen levels reported by bloodwork and by the pulse ox. This could impact clinical treatment decisions, as they report higher oxygen saturations than are actually present. This could lead to a delay in necessary clinical treatment when a patient’s o2 level drops below critical thresholds.

    This could change the path your life takes

    COMPAS is an algorithm widely used in the US to guide sentencing by predicting the likelihood of a criminal reoffending. In perhaps the most notorious case of AI prejudice, in May 2016 the US news organisation ProPublica reported that COMPAS is racially biased. While COMPAS did predict reoffending with reasonable accuracy, black people were twice as likely to be rated a higher risk but not actually reoffend. The graphs show that risk scores are very far from normal distribution – they are skewed heavily towards low risk for white defendants. In multiple real life examples from the ProPublica analysis, the black defendant was rated as a higher risk, despite fewer previous offences, and in both cases that individual did not reoffend, although the “lower risk” defendant did. 

    And these are, sadly, just selected examples. There are many, many, many more.

    Clang, clang, clang went the trolley

    As we come to the end of the real world examples, I want to leave you with a hypothetical that is becoming reality just a little bit too fast. Something that many people are excited about is the advent of self-driving cars. These cars will avoid crashes and keep drivers safe and allow us to do other things with our commute. But….

    Have you ever heard of the trolley problem? It’s a well known simple question that is often used to explore different ethical beliefs. In case you aren’t familiar with this yet, the picture above is a fair summary. Imagine you are walking along and you see a trolley, out of control, speeding down the tracks. Ahead, on the tracks, there are five people tied up and unable to move. The trolley is headed straight for them. You are standing some distance off in the train yard, next to a lever. If you pull this lever, the trolley will switch to a different set of tracks. However, you notice that there is one person on the side track. You have two options:

    • Do nothing and allow the trolley to kill the five people on the main track.
    • Pull the lever, diverting the trolley onto the side track where it will kill one person.

    What is the right thing to do?

    The tricky thing is that there are numerous ways to try and decide what is right, and there isn’t really a right answer. As humans we can perhaps use more about the context to aid us in a decision, and we can take in all of the information about the situation even if we have not encountered the situation before, but even then we still can’t always come to a right answer. So how do we expect a smart car to decide?

    While we might not see a real life trolley problem in our lifetimes, the push towards self driving cars will almost certainly see a car presented with variations on this problem – in avoiding an accident, does the car swerve to hit one pedestrian to save five? Does it not swerve at all, to preserve the life of the driver? Given what we know about recognition software as it currently stands, will it accurately recognise every pedestrian?

    How will the car decide? And who is responsible for the decision that it makes? The company? The programmer who implemented the algorithm?

    I don’t have an answer for this one, and I’m not sure that anyone does. But there is a lot that we can do to action inclusive and diverse programming in our jobs, every single day, so that we remove the real barriers that I’ve already shown.

    What can we do?

    First and foremost, diversity starts from the very bottom up. We need to be really inclusive in our design – think about everyone who will use what you make and how they will use it, and really think beyond your own experience.

    Make decisions thoughtfully – many of the examples I’ve shown weren’t created with malicious intent, but they still hurt, dehumanised, or impaired people. Sometimes there isn’t going to be a simple answer, sometimes you will need to have a form with “first name” and “last name”, but we can make these decisions thoughtfully. We can choose to not “go with the default” and consider the impact of our decisions beyond our own office.

    Garbage in, garbage out – if you are using a dataset, consider where it came from. Is it a good representative set? Is your data building bias into the system, or is it representative of all of our customers?

    Inclusive hiring – when many diverse voices can speak, we spot more of these problems, and some of them won’t make it out the door. Diverse teams bring diverse life experiences to the table, and show us the different ways our “defaults” may be leaving people out in the cold.

    Learn more – In the coming days and weeks, I’ll be sharing more links and some deep dives into the topics I’ve raised above, because there is so much more to say on each of them. I’m going to try and share as many resources and expert voices as I can on these topics, so that we can all try to make what we make better.