25+ yr Java/JS dev
Linux novice - running Ubuntu (no windows/mac)

  • 0 Posts
  • 627 Comments
Joined 1 year ago
cake
Cake day: June 14th, 2023

help-circle



  • I don’t think I would agree that just because something is public that it’s a public forum. I feel like the public has to own it as well. I looked it up and maybe it’s because I predate social media by rather a lot, but I think of it in the classical sense:

    Public forums are typically categorized into three types:

    1. Traditional Public Forums: Long-established spaces like parks or sidewalks, where people have historically exercised their rights to free speech and assembly.
    2. Designated Public Forums: Areas that the government intentionally opens up for public expression, such as town halls or school meeting rooms.
    3. Limited Public Forums: Spaces opened for specific types of discussions or activities but with certain restrictions on the subject matter or participants.

    The important factor being public ownership of the forum. I will concede that it has colloquially come to include public social media, but I think it’s important to distinguish that it’s not really the same thing at all as has been discussed through most of our history.

    Food for thought. I just think calling them public forums attaches too much importance to a profit seeking endeavor.






  • You made a lot of points here. Many I agree with, some I don’t, but I specifically want to address this because it seems to be such a common misconception.

    It does and it doesn’t discard the original. It isn’t impossible to recreate the original (since all the data it gobbled up gets stored somewhere in some shape or form and can be truthfully recreated, at least judging by a few comments bellow and news reports). So AI can and does recreate (duplicate or distribute, perhaps) copyrighted works.

    AI stores original works like a dictionary does. All the words are there, but the order and meaning is completely gone. An original work is possible to recreate by randomly selecting words from the dictionary, but it’s unlikely.

    The thing that makes AI useful is that it understands the patterns words are typically used in. It orders words in the right way far more often than random chance. It knows “It was the best of” has a lot of likely options for the next word, but if it selects “times” as the next word, it’s far more likely to continue with, “it was the worst of times.” Because that sequence of words is so ubiquitous due to references to the classic story. But over the course of following these word patterns, it will quickly glom onto a different pattern and create a wholly new work from the original “prompt.”

    There are only two cases in which an original work should be duplicated: either the training data is far too small and the model is overtrained on that particular work, or the work is the most derivative text imaginable lacking any flair or originality.

    Adding more training data makes it less likely to recreate any original works.

    I am aware of examples where it was claimed an LLM reproduced entirely code functions including original comments. That is either a case of overtraining, or far too many people were already copying that code verbatim into their own, thus making that work very over represented in the training data (same thing, but it was infringing developers who poisoned the data, not researchers using bad training data).

    Bottom line: when created with enough data, no original works are stored in any way that allows faithful reproduction other than by chance so random that it’s similar to rolling dice over a dictionary.

    None of this means AI can do no wrong, I just don’t find the copyright claim compelling.






  • I wouldn’t call it ignorance. My bliss is based on considerable reading and contemplation. Way more than ought to be necessary, all things considered. I wish some of the understanding I’ve developed over the years would’ve been explained when I was young. There are things which, once explained, drive the logical mind to an inevitable conclusion.


  • I’ve been there. Posted my story, but I didn’t talk about the lifelong anxiety that comes with a lengthy layoff. Continually pursued higher pay at shittier jobs to try to get ahead of things for when the rug gets pulled out from under me again. It’s corrosive. Losing income and insurance when everyone is counting on you to provide makes you feel like your self-worth is completely tied to your job and ability to provide.


  • Which time?

    First time happened after I’d been with my first real job for ten years because the business was changing and there wasn’t a role for me. I was out of work for 7 shitty months trying to have my own business starting from the few customers we had left when they let me go. It was right after I bought a house and had a baby. It was fucking awful.

    Second time was after COVID. First we all took a 10% pay cut to avoid layoffs. Then two months later when federal assistance expired, they cut 1/3 of the company across the board. I’m a little fucking bitter about that to be honest, but I had a new remote job lined up within a couple of weeks that paid quite a bit better.

    Last time was 5 months ago. Just got hired this week. Start next month. It sucked. Wiped out my whole retirement savings, so I get to start over at 51. But we made it through and potentially I won’t have to switch companies again.


  • I mostly get what you’re saying, though I don’t have the requisite understanding to follow formal proofs, but if there is one thing I do know for certain, it’s that “understanding” is anthropomorphizing and shorthand for something that is very much not understanding in a human context at all.

    I get that it can be hard to find the right words to explain a some of these emergent phenomena, but I think it’s misleading to use words that make AI appear to have a thought process akin to anything we could understand as such—at least in settings where folks might not understand the shorthand as such.

    And maybe everyone here is aware of that, but it makes me uneasy, hence this comment to hopefully make that point.