• 1 Post
  • 28 Comments
Joined 1 year ago
cake
Cake day: June 20th, 2023

help-circle


  • this brings up the question: what is a book? what is art? if an “AI” can now churn out the next harry potter sequel and people literally can’t tell that it’s not written by JK Rowling, then what does that mean for what people value in stories? what is a story? is this a sign that we humans should figure something new out, instead of reacting according to an outdated protocol?

    yes, authors made money in the past before AI. now that we have AI and most people can get satisfied by a book written by AI, what will differentiate human authors from AI? will it become a niche thing, where some people can tell the difference and they prefer human authors? or will there be some small number of exceptional authors who can produce something that is obviously different from AI?

    i see this as an opportunity for artists to compete with AI, rather than say “hey! no fair! he can think and write faster than me!”







  • this is what i’m frustrated with. why do all these engineers let themselves be told what to do even if it makes a worse-functioning tool? that’s not real engineering.

    “because they’ll get fired”

    not if enough of them do the thing that should’ve been what got them interested in engineering in the first place.

    maybe we shouldn’t call them engineers, but something else relating to being the one who does the dirty work for institutions that aim to steal people’s attention and decrease their quality of life.

    and if they do get fired, then they should join together and make the reasonable company that makes good tools for human use.


  • ok, you make good points, but i feel like the algorithm could work to not have the system grind to a halt. i’d have to look at other examples where this has been done. but maybe i am overly-optimistic and it’s not possible.

    who would pay for those nodes you are querying

    the people who are already running nodes, like lemmy.world, lemmy.ml, me, etc. i run some services on my home server that i let anyone use, because i have the hardware and the bandwidth to be able to afford it. there are enough people who have the necessary hardware and bandwidth to contribute to it at minimal detriment to them. it’s already an open-source project where people volunteer their time to code it.

    i’ll read up on oxen network.

    in an anonymous way

    wait who said anything about anonymous? what are talking about being anonymous? there would still be user accounts.

    if I don’t want to aggregate all the posts in the world by myself (as you are suggesting), then I’ll have to fine someone to do it for me

    this is already what is done, except that the data is not stored in a replicated and distributed manor. you get all the posts in the world of a community of an instance. it is one server, with all the data stored on its harddrive, like a traditional website. in what i’m proposing, this is also what would happen in many cases, because the thing wouldn’t requery the entire network every time you request posts, there would be a time threshold, like how posts are cached on your local mobile device for most social media apps. posts would be cached on the server.

    now, yes, this architecture would in fact result in more network traffic occurring between each and every node, as they receive updates about events on other nodes. so that would be extra burden upon the hosts. but i believe it is something we can work through.


    1. you connect to some lemmy instance on your web browser
    2. the client application (lemmy web app) authenticates your login credentials by first checking its own user database, if it doesn’t find you (which it should because by default you’d be connecting to an instance that you’ve already used, and if done through a mobile app for example it would automatically find the best instance to use by lowest latency), it send out a message to the nodes(instances) that it knows about, searching for your user, recursively, when found, sent back and stored in each node that was part of the searching. (there’d be some threshold of tree depth so the unsuccessful branches don’t keep going forever, and some other algorithmic details to prevent redundant network activity)
    3. you navigate to your subscribed communities feed, lemmy shows you the posts that are already on the node that you are directly connected to, then asynchronously sends out a request to the surrounding nodes to pull more posts from those communities, recursively reaching out to adjacent nodes, again avoiding repeatedly hitting the same node via algorithmic details which we can discuss further if you wish, sending back the info up the tree to your primary node. now a bunch of servers have duplicated community data, like a distributed storage system, but you, the user, don’t know about all that stuff that just happened behind the scenes. your GUI is updated accordingly
    4. now you can interact with these posts, make new posts, and each interaction will be sent out to all the relevant nodes in a reverse process.
    5. another user on the other end can visit some community that you just posted to, and a request will again be propagated through the network, but starting from his node, and eventually reaching some node that has your new post.

    the advantages of this:

    • if a node goes down, not all of the community and user data is lost, because its neighbor nodes have replicated the data
    • if i am hosting a node, and have limited bandwidth and storage, i can specify limits so that my network is not unintentionally DoSed. so this implies that when the prior-described processes are occurring, some instances will not store the data they are pushing through, which is fine, and one of the intended features of this distributed architecture
    • similar to previous point, each instance can have a whitelist or blacklist of communities (for either storage and/or data passing), defined by the admin, if he/she wishes to tailor the content for example to keep it related to content they are interested in rather than being forced to serve everyone on the network. it’s like if someone wants to help a little bit but they don’t have all the bandwidth and storage in the world, they can, instead of having to handle traffic for a bunch of irrelevant-to-them communities.







  • to expound:

    the tankie instance or the nutballs on the fascist instance

    here you reveal a conceptual misunderstanding, or rather, a part of the lemmy architecture which i disagree with. there shouldn’t be a concept of a “interest X instance” etc. it should be similar to a distributed storage model. so the concept of a community is not per-instance, it’s just an abstract thing that exists in conceptual space.



  • everyone has all the duplicated data.

    everyone does not have all the duplicated data. they only have the data that they need – the data requested by a user who happens to be using some instance.

    handling defederating is a good point. there could be malicious nodes that would be damaging to the network. i suppose there could be a community-mainted ledger of known malicious nodes (similar to minecraft usernames of known hackers), and the admins of the servers would maintain a blacklist. (obviously you configure that your instance’s blacklist would be automatically synced with this ledger)

    the mega community idea could be good. where is this being discussed?