Not liking that all of the time travelers on 4chan seem to be pointing towards a 2027 surprise like this.
forkDestroyer
This is why I donate blood. Gotta offload some microplastics.
This is how I found out that lobsters don't have a single centralized brain like humans do.
"if you do it correctly" holds a lot of weight to this argument. I'd be worried of anyone who wants to start from scratch, instead of building on the current foundation.
You make good points, though. I'm just more skeptical than you.
I regret to inform you that salaries in tech are not as glorious as I thought they'd be. I'd be surprised to have enough to own a farm any time soon.
Would be nice to be able to afford a house, though.
instead voting for independent anarchist parties which try to get rid of as many laws and government institutions and also nationalize anything the people will be better served by, under collective ownership.
Once the laws are gone/things are deregulated, the corpos will likely take over. Nationalizing sounds good, but likely won't end well without regulations, imo.
I'm being a bit extra but...
Your statement:
The article headline is wildly misleading, bordering on being just a straight up lie.
The article headline:
A Developer Accidentally Found CSAM in AI Data. Google Banned Him For It
The general story in reference to the headline:
- He found csam in a known AI dataset, a dataset which he stored in his account.
- Google banned him for having this data in his account.
- The article mentions that he tripped the automated monitoring tools.
The article headline is accurate if you interpret it as
"A Developer Accidentally Found CSAM in AI Data. Google Banned Him For It" ("it" being "csam").
The article headline is inaccurate if you interpret it as
"A Developer Accidentally Found CSAM in AI Data. Google Banned Him For It" ("it" being "reporting csam").
I read it as the former, because the action of reporting isn't listed in the headline at all.
^___^
Google isn't the only service checking for csam. Microsoft (and other file hosting services, likely) also have methods to do this. This doesn't mean they also host csam to detect it. I believe their checks use hash values to determine if a picture is already clocked as being in that category.
This has existed since 2009 and provides good insight on the topic, used for detecting all sorts of bad category images:
Lived a block away from a Catholic church for several years. Bells did not ring for half an hour, but I do remember them ringing a little longer for some holidays.
I guess other Catholic churches handle it differently?
Idk, bells sounding the hour are basically like any clock to me (I like clocks so I'm biased). Spoken prayer over a loud speaker is much more invasive imo.
There's no point in taking moderation seriously. It's a volunteer position and the tools to handle slop aren't there.
I hear they're bringing back digg.
I also hear it's a former Reddit C-suite that's bringing it back, so it'll probably be more of the same.
Convinced a majority of the use cases are from corporations mandating employees to use it.