ibtimes.co.uk

Sarmyth , (edited ) to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

Facial recognition still struggles with really bad mistakes that are always bad optics for the business that uses it. I'm amazed anyone is still willing to use it in its current form.

It's been the norm that these systems can't tell the difference between people of dark pigmentation if it even acknowledges it's seeing a person at all.

Running a system with a decade long history or racist looking mistakes is bonkers in the current climate.

DarkThoughts ,

You make it sound like that's a bad thing.

Buddahriffic ,

I'm curious how well the systems can differentiate doppelgangers and identical twins. Or if some makeup is enough to fool it.

PM_Your_Nudes_Please ,

Facial recognition uses a few key elements of the face to hone in on matches, and traditional makeup doesn’t obscure any of those areas. In order to fool facial recognition, the goal is often to avoid face detection in the first place; Asymmetry, large contrasting colors, obscuring one (or both) eyes, hiding the oval head shape and jawline, and rhinestones (which sparkle and reflect light nearly randomly, making videos more confusing) seem to work well. But as neural nets improve, they also get harder to fool, so what works for one system may not work for every system.

CV Dazzle (originally inspired by dazzle camouflage used on some warships) is a makeup style that tries to fool the most popular facial recognition systems.

https://lemmy.world/pictrs/image/52d8f120-e959-481d-ac06-49b718824128.jpeg

https://lemmy.world/pictrs/image/c93a006b-8d63-4e49-9733-a5509cbf5f8f.jpeg

Note that those tend to obscure the bridge of the nose, the brow line, the jawline, etc… Because those are key identification areas for facial recognition.

Buddahriffic ,

Yeah, if we can still recognize those as faces, it's possible for a neural net to do so as well.

But I'm talking more about differentiating faces than hiding entirely from such systems. Like makeup can be used to give the illusion that the shape of the face is different with false contour shading. You can't really change the jawline (I think... I'm not skilled in makeup myself but have an awareness of what it can do) but you can change where the cheekbones appear to be, as well as the overall shape of the cheeks, and you can adjust the nose, too (unless it's a profile angle).

I think the danger in trying to hide that you have a face entirely is that if it gets detected, there's a good chance that it will be flagged for attempting to fool the system because those examples you gave are pretty obvious, once you know what's going on.

It would be like going in to a bank with a ski mask to avoid being recognized vs going in as Mrs Doubtfire. Even if they are just trying to do banking that one time, the ski mask will attract unwanted attention while using a different face would accomplish the goal of avoiding attention.

nifty ,
@nifty@lemmy.world avatar

Functional and subversive, just how I like my makeup

UnderpantsWeevil , (edited )
@UnderpantsWeevil@lemmy.world avatar

The catch is that its only really a problem for the people getting flagged. Then you're guilty until proven innocent, and the only person to blame is a soulless machine with a big button that reads "For customer support, go fuck yourself".

As security theater, its cheap and easy to implement. As a passive income stream for tech bros, its a blank check. As a buzzword for politicians who can pretend they're forward-thinking by rolling out some vaporware solution to a non-existent problem, its on the tip of everyone's tongue.

I'm old enough to remember when sitting Senator Ted Kennedy got flagged by the Bush Admin's No Fly List and people were convinced this is the sort of shit that would reform the intrusive, incompetent, ill-conceived TSA. 20 years later... nope, it didn't. Enjoy techno-hell, Brits.

rottingleaf , (edited ) to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

So cute.

For old stuff things like minority rights and all other principles about making people comfortable apply, and reliability standards with a lot of nines have to be met.

For new stuff - "if it fails 1/100 times, then it's fine, so screw you".

See, everybody (or at least people whose voices are heard, not us dumb fucks, authentic Zuck quote btw) wants all this tech bro surveillance centralized obscure blackbox ambiguous crap so fucking badly that other things don't matter.

Boeing planes dropping outta sky? Wait till "AI" reaches nuclear energy. Or until autonomous armed police drones roam your area, as something easier to imagine. (I've just remembered that in Star Wars police drones on Coruscant are unarmed, both under Republic and under Empire. EU writers couldn't imagine our times' degree of stupidity EDIT: so I'm imagining it now.)

TrickDacy ,

Fear mongering

ABCDE ,

How so? India doesn't recognise people in some unusual but not impossible circumstances, such as those without fingerprints.

DarkThoughts ,

Look at China.

TrickDacy ,

They are living in the same timeline as us. The difference is level of authoritarianism, not that they have warped into the future.

DarkThoughts ,

And what do you think those mass surveillance tools are supposed to be used for?

CeeBee ,

For new stuff - "if it fails 1/100 times, then it's fine, so screw you".

The FMR (false match rate) for these systems is 1e-6. Which translates to about 0.000001% error rate or 99.999999% accuracy. And that number was from about 3 or 4 years ago. They are much better today.

How do I know? I worked in that industry building that kind of tech and read through the 500+ page NIST report on testing results for all the various Face Recognition vendors.

NeoNachtwaechter ,

How do I know? I worked in that industry building that kind of tech

Thank you for pleading guilty :)

CeeBee ,

LMAO. You have no idea what I built the system for, and I have no skin in the game anymore as I moved on to a completely different industry that doesn't even use AI at all.

The implications of your argument is the same with flat earthers where they demand photographic proof of a spherical Earth, but when they are shown photos from space they simply say it's fake and NASA is in the lie.

Sometimes you just can't get past people's preconceived biases regardless of the facts.

notabot ,

It doesn't really matter whether the FMR is one in a hundred or one in a million, for the uses it's being put to it's still too high. If it was only being used as one factor for authenticating someone (I.e. the 'thing for are') but still required the other factor(s) (the 'thing you know' and the 'thing you have') then it'd be adaquate.

As it stands, when it's being used either to pick someone out without further scrutiny, or to make payments with no further checks, it's simply unacceptable. There's good arguments to say it's not just the error rate is unacceptable, but that the technology simply shouldn't be used in those scenarios as it's entirely inappropriate, but that's a separate discussion.

CeeBee ,

The truth is the numbers I cited are the 1:N stats. The 1:1 numbers are far higher, because you can immediately control for distance, lighting, angle, and gaze direction.

With my system I worked with had a 1:1 match rate that was statistically perfect (but nothing is ever 100%).

the technology simply shouldn't be used in those scenarios as it's entirely inappropriate, but that's a separate discussion.

Agreed. Its use as a payment system is just ridiculous.

GluWu , (edited )

[Thread, post or comment was deleted by the author]

  • Loading...
  • CeeBee ,

    That was a really garbage system then. Like disgracefully bad Fisher Price quality.

    The reality is that there are more crap systems than really good ones out there. And there are as many algorithms and different ways of doing it as there are companies.

    The system I developed was so good, even when we tried all kinds of shenanigans to trip it up, we just couldn't do it.

    rottingleaf ,

    You know what overfitting is, right?

    Other than that, if this system of yours makes 1 error in a million scans, that's still not very good, if that's treated as "virtually no errors" as in no talking to manager, no showing ID as a fallback, so on. Say, if it were employed in Moscow subway, that'd mean a few unpleasant errors preventing people from getting where they need every day.

    CeeBee ,

    You know what overfitting is, right?

    As you reply to someone who spent a decade in the AI industry.

    This has nothing to do with overfitting. Particularly because our matching algorithm isn't trained on data.

    The face detection portion is, but that's simply finding the face in an image.

    The system I worked with used a threshold value that equates to an FMR of 1e-07. And it wasn't used in places like subways or city streets. The point I'm making is that in the few years of real world operation (before I left for another job) we didn't see a single false detection. In fact, one of the facility owners asked us to lower the threshold temporarily to verify the system was actually working properly.

    rottingleaf ,

    This has nothing to do with overfitting. Particularly because our matching algorithm isn’t trained on data.

    Good to know.

    The face detection portion is, but that’s simply finding the face in an image.

    So you are saying yourself that your argument has nothing to do with what's in the article?..

    CeeBee ,

    So you are saying yourself that your argument has nothing to do with what's in the article?..

    OP said "reliability standards with a lot of nines have to be met". All I'm saying is that we're already there.

    rottingleaf ,

    Well, the place you worked at is already there. Those stores - possibly not.

    Also I said that about new and shiny stuff like what they call "AI".

    CeeBee ,

    Fair enough

    ricdeh ,
    @ricdeh@lemmy.world avatar

    A probability of 10^-6^ corresponds to 10^-4^ %, not 10^-6^ %.

    CeeBee , (edited )

    Ok, sure

    Edit: the truth is that saying x% accuracy isn't entirely correct, because the Numbers just don't work that way. It's just a way we convey the data to the average person. I can't count the amount of times I've had asked "ok, but what doesn't mean in terms of accuracy? What's the accuracy percentage?"

    And I understand what you're saying now. Yes I did have the number written down incorrectly as a percentage. I'm on mobile this whole time doing a hundred other things. I added two extra digits.

    Jackthelad , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    Well, this blows the "if you've not done anything wrong, you have nothing to worry about" argument out of the water.

    raspberriesareyummy ,

    That argument was only ever made by dumb fucks or evil fucks. The article reports about an actual occurrence of one of the problems of such technology that we (people who care about privacy) have warned about from the beginning.

    Maggoty ,

    The US killed that argument a long time ago. We shot it in the back and claimed it had a gun.

    dependencyinjection , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    Could attaching infrared leds to glasses or caps be away to prevent these systems from capturing people’s faces.

    Think I saw something on black mirror once or might have been another show.

    Possible business opportunity selling to crackheads lol.

    fiercekitten , (edited )

    I think IR lights on glasses can blind cameras, and there are also infrared-blocking glasses that also reflect IR light from cameras back at them. So yes, adding lights or reflectors can be effective.

    But we shouldn't even have to consider these workarounds to maintain our privacy and security. And even if we start wearing these glasses or fooling the systems, governments will outlaw the use of such circumventing tech.

    Our reaction to this needs to be "we will not allow this tech to be used against us, period." Ban it for law enforcement, ban it for commerce, ban it in airports and schools.

    sugar_in_your_tea ,

    governments will outlaw the use of such circumventing tech

    They can't stop all of us. If they outlaw using it, I'd support (or even organize) a protest group that organizes widespread use of them.

    fiercekitten ,

    Well in that case, one brand of such glasses is Reflectacles 😎

    sugar_in_your_tea ,

    Reflectacles

    Cool! Looks like they also have a discount for Monero purchases. I might have to pick one up.

    I wish they would make a line available for cheap so a charity could hand them out for free, ~$150 is a bit much.

    squidspinachfootball ,

    Very neat, I wonder how effective they are at confusing facial recognition and 3D facial scanning systems? Not that we often encounter the 3D scanning, but an interesting aspect to consider.

    sugar_in_your_tea ,

    I just wonder if I'd get in trouble for wearing them at the airport. That's the most likely place for 3D facial scanning.

    But AFAIK, most facial recognition systems use IR, whether they're 2D or 3D, so I would guess they'd be pretty effective.

    dependencyinjection , (edited )

    Thanks for the informative and thought out reply.

    Also, thanks for killing my joke business idea by mentioning a brand that makes these.

    I would like to see them in the rim of a cap too. As I always have a cap in my head.

    Edit: Hot damn they’re expensive. Business idea is back.

    Edit 2: If they have blockers to block IR. Would UV light have the same issue or other wavelengths that are not visible to humans.

    Edit 3: UV is dumb. We don’t need cancer.

    Final solution is we all convert to Islam and pretend to be women.

    You999 ,

    Ultraviolet would be a terrible idea as UV-A and UV-B can cause skin cancer over prolonged exposure and UV-C will straight up burn your eyes and skin.

    dependencyinjection ,

    Thanks for saving my dumbass. Got enough going on to deal with skin cancer haha.

    setsneedtofeed ,
    @setsneedtofeed@lemmy.world avatar

    I presume at that point the store would just have security walk out the person wearing the hat.

    dependencyinjection ,

    How would they know it had lights in it?

    It could be a ravers led hat that isn’t turned on.

    setsneedtofeed ,
    @setsneedtofeed@lemmy.world avatar

    If the AI is flagging faces and immediately alerting employees, it is likely also going to throw up a flag for abnormal interference like that. Or if it doesn't do it now, it is a feature that could be added if such hats become a common enough.

    uis , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech
    @uis@lemm.ee avatar

    Is it even legal? What happened to consumer protection laws?

    PenisWenisGenius ,

    I'm going to think twice before enerting any big box store because of this.

    afraid_of_zombies ,

    I am amazed that anyone is still using them. It is the 21st century, just shop online.

    intensely_human ,

    Last time I ordered big boxes online, they just shipped me empty boxes. I don’t know how they screwed that up, but then I’ve always gone to big box stores so I can actually see the big boxes I’m buying.

    afraid_of_zombies ,

    Fair

    deafboy ,
    @deafboy@lemmy.world avatar

    No need to avoid such stores, just wear a protection to avoid your face from being falsely flagged.

    https://lemmy.world/pictrs/image/78e5db4f-50dc-466c-9e82-02eb25f7a5d2.webp

    ChillPenguin ,

    Welcome to 2024. Where we move closer and closer to black mirror being real due to our technology improvements.

    Buddahriffic ,

    Plus our allowing sociopaths to hold so much power.

    intensely_human ,

    Allowing bad character to reign generally speaking. Sociopaths are like seed crystals; it’s the rest of us implanting their way of life that do most of the evil, even if they’re the ones providing the purest examples.

    We need to be good people, not just nice people, in order to improve things.

    maynarkh ,

    Brexit. The EU has laws forbidding stuff like this.

    uis ,
    @uis@lemm.ee avatar

    I don't think this is legal in Russia, Ukraine and Belarus either.

    ChaoticEntropy , (edited ) to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech
    @ChaoticEntropy@feddit.uk avatar

    Oh good... janky oversold systems that do a lot of automation on a very shaky basis are also having high impacts when screwing up.

    Also "Facewatch" is such an awful sounding company.

    intensely_human ,

    Cara Filter wouldn’t be bad, just for the ring if it

    muntedcrocodile , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    We have so many dystopian futures and we decided to invent a new one.

    Lith ,
    @Lith@lemmy.sdf.org avatar

    Actually this one feels pretty similar to watch_dogs. Wasn't this the plot to watch_dogs 2?

    sfxrlz , (edited )

    Now I’m interested in the plot of watch dogs 2…

    Edit: it’s indeed the plot of watch dogs 2

    https://en.m.wikipedia.org/wiki/Watch_Dogs_2

    In 2016, three years after the events of Chicago, San Francisco becomes the first city to install the next generation of ctOS – a computing network connecting every device together into a single system, developed by technology company Blume. Hacker Marcus Holloway, punished for a crime he did not commit through ctOS 2.0 …

    rmuk ,

    Also, a kickass soundtrack by Hudson Mohawke.

    Suavevillain , (edited ) to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech
    @Suavevillain@lemmy.world avatar

    People who blindly support this type of tech and AI being slapped into everything always learn the hard way when a case like this happens.

    squid_slime , (edited )
    @squid_slime@lemm.ee avatar

    Sadly there won't be any learning, the security company will improve the tech and continue as usual.

    This shit is here to stay :/

    EngineerGaming , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech
    @EngineerGaming@feddit.nl avatar

    Are you prohibited from covering your face in the stores like this?

    sugar_in_your_tea ,

    Idk, but as someone who has a fair skin tone and thus likely wouldn't trigger a false positive, I am prohibited from entering the store based purely on the fact that they use face recognition software. Screw that.

    EngineerGaming ,
    @EngineerGaming@feddit.nl avatar

    I mean wouldn't the prohibited person just have to wear some mask to avoid triggering? And if this is prohibited - wouldn't that be problematic when it comes to hijabs?

    sugar_in_your_tea ,

    I'm guessing they use eye tracking, so they'd probably need something to block IR facial recognition. Another user mentioned Reflecticles, which should work but are a bit pricey.

    NeoNachtwaechter , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    This raises questions about how 'good' this technology is.

    But it also raises the question of how well your police can deal with false suspicions and false accusations.

    CeeBee ,

    This raises questions about how 'good' this technology is.

    No it doesn't. For every 10 million good detections you only hear about the 1 or 2 false detections. The issue here are the policies around detections and how to verify them. Some places are still taking a blind faith approach to the detections.

    NeoNachtwaechter ,

    For every 10 million good detections you only hear about the 1 or 2 false detections.

    Considering the impact of these faults, it is obviously not good enough.

    CrayonMaster ,

    But that really says more about the user then the tech. This issue here isn't that the tech has too many errors, it's that stores use it and it alone to ban people despite it having a low but well known error rate.

    NeoNachtwaechter ,

    says more about the user then the tech.

    "You need to have a suitable face for our face recognition?" ;-)

    stores use it and it alone to ban people

    No. Read again. The stores did not use technology, they used the services of that tech company.

    nyan ,

    stores use it and it alone to ban people despite it having a low but well known error rate.

    And it is absolutely predictable that some stores would do that, because humans. At very least, companies deploying this technology need to make certain that all the store staff are properly trained on what it does and doesn't mean, including new hires who arrive after the system is put in. Forcing that is going to require that a law be passed.

    CeeBee ,

    Considering the impact of these faults, it is obviously not good enough.

    I was throwing a rough number out there, but the true error rate is lower than what I said. But when with those numbers this tech is statistically safer than driving a car.

    The other half of the equation is policy management. Every single one of these systems should operate with human-in-the-loop. Meaning after a detection is made, it goes over to a person to make a "real" determination for accuracy.

    explodicle ,

    That doesn't sound as cost-effective as just losing the millionth customers.

    CeeBee ,

    Because using FR for a payment system is dumb.

    Everyone here seems to be hyperfixated on the payment system aspect.

    I'm talking purely in the context of FR tech being vastly better than what people think, since everyone has this Idea that FR tech doesn't work.

    DarkThoughts ,

    Good as in ethical, not in capability. Facial recognition (and similar mass surveillance tech) is simply a tool for authoritarianism and should be banned for general usage. There's literally no good reason why this should be widely used at all.

    Vipsu , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech
    @Vipsu@lemmy.world avatar

    Can't wait for something like this get hacked.
    There'll be a lot explaining to do.

    dogsnest ,
    @dogsnest@lemmy.world avatar

    I read this in a Ricky Ricardo voice.

    Jayjader , (edited )

    Still, I think the only way that would result in change is if the hack specifically went after someone powerful like the mayor or one of the richest business owners in town.

    Rooskie91 , to Technology in Delivery Goes Wrong: New Cybertruck Slices Owner's Wrist During Inspection

    Finding new and exciting ways to make cars more dangerous. What an inovator.

    Emmie , (edited ) to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    There we go guys. It’s funny when nutty conspiracy theorists are against masks when they should be wearing frikin balaclavas

    Clbull , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    Please tell me a lawyer is taking this on pro bono and is about to sue the shit out of Facewatch.

    Alph4d0g , to Technology in UK Woman Mistaken As Shoplifter By Facewatch, Now She's Banned From All Stores With Facial Recognition Tech

    She got more than she Home Bargained for

  • All
  • Subscribed
  • Moderated
  • Favorites
  • kbinchat
  • All magazines