I am a bit worried about the number of people here who, like me, are fairly recent arrivals, and who are using the so-called 'extreme' and 'unreasonable' reactions to them wanting to 'hack' the fediverse to write off the whole conversation around consent as somehow not relevant.
The NSA monitors anything you do on the internet anyway, so why are you complaining about tech bros wanting to harvest out in the open, yada yada.
It seems we need to define what consent actually is.
AGAIN.
I think blurring the lines between public and private spaces is the
opposite of informing consent. Cultivating unrealistic expectations of
“privacy” and control in what are ultimately public spaces is actually
bad.
I tried to single out the world wide web, as opposed to the internet
at large, because the two are not synonymous. It’s rather absurd to
publicly serve webpages to any querying IP address and maintain that
the receiving computer is not to save said pages to disk.
All this to say: I find it difficult to argue that web publications
should or could be exempt from aggregation and archival (or scraping,
to put it another way). I understand that the ease with which bots do
this can be disconcerting, however.
If we stay with the cafe bulletin board, getting a detailed overview of all the
postings on the board is akin to scraping the whole thing. If we extend
our analogy instead to a somewhat more significant example, library
catalogs do the same with books, magazines, and movies.
This is the cost of publishing, be that in print or online. It must be
expected that some person has a copy of every- and anything one has
ever written or posted publicly, and perhaps even catalogued it. A way around
this might be to move away from the web to another part of the internet,
like Matrix, as alma suggested.
I assume the non-consensual collection of various (meta-)data is what
you refer to when talking about intrusion and money making.
Lemmy, like many projects, seeks to offer an alternative to corporate,
data-gobbling social media sites, but doesn’t eliminate the ability
to search through its webpages.
I think Besse makes a great point here:
I tried to single out the world wide web, as opposed to the internet at large, because the two are not synonymous. It’s rather absurd to publicly serve webpages to any querying IP address and maintain that the receiving computer is not to save said pages to disk.
All this to say: I find it difficult to argue that web publications should or could be exempt from aggregation and archival (or scraping, to put it another way). I understand that the ease with which bots do this can be disconcerting, however.
If we stay with the cafe bulletin board, getting a detailed overview of all the postings on the board is akin to scraping the whole thing. If we extend our analogy instead to a somewhat more significant example, library catalogs do the same with books, magazines, and movies.
This is the cost of publishing, be that in print or online. It must be expected that some person has a copy of every- and anything one has ever written or posted publicly, and perhaps even catalogued it. A way around this might be to move away from the web to another part of the internet, like Matrix, as alma suggested.
I assume the non-consensual collection of various (meta-)data is what you refer to when talking about intrusion and money making. Lemmy, like many projects, seeks to offer an alternative to corporate, data-gobbling social media sites, but doesn’t eliminate the ability to search through its webpages.