Basically a deer with a human face. Despite probably being some sort of magical nature spirit, his interests are primarily in technology and politics and science fiction.

Spent many years on Reddit before joining the Threadiverse as well.

  • 0 Posts
  • 1.34K Comments
Joined 2 years ago
cake
Cake day: March 3rd, 2024

help-circle





  • Here’s a clear description of the image at that URL:

    It’s a two-panel meme. The top panel shows a shiny, red-lit robot skeleton reminiscent of the Terminator with bold white text above it that says “AI EXPECTATIONS.” Below it, the second panel uses the classic “Is this a pigeon?” anime still (a character gesturing toward a butterfly) with overlaid text reading “AI REALITY” above and “IS THIS DOG?” below. The joke contrasts lofty, futuristic expectations of artificial intelligence with a humorous reality of AI misidentifying a butterfly as a dog.

    source



  • Sure, I’m not saying this isn’t “malicious.”

    I’m questioning why this particular instance of lawbreaking makes his site an “unreliable source”, whereas all the copyright violation he’s been up to all along didn’t? And now you’re bringing in speculative instances of future lawbreaking that also seem unrelated, what does crypto mining have to do with the reliability of the sources archived there?

    My point here is that people are jumping from “he did something bad that I don’t like!” to “therefore everything he does is bad and wrong!” Without a clear logical connection between those things. Sure, the DDOS thing is a good reason to try to avoid sending traffic to his site. But that has nothing to do with the reliability of the information stored there.



  • You’re misinterpreting what Wikimedia’s “free knowledge” mandate is about. They have a hard-line requirement that the knowlege they distribute is legally free, for example - it has to be under an open license. archive.today is quite the opposite of that. They don’t just archive any old knowledge willy-nilly, they’ve got standards. And so forth.

    Simply running an archive.today clone would not fit. The “source documents only” archive would already be stretching the edges rather far. There’s already Wikisource, for example, and it’s got the “open licenses only” restriction.









  • Wikipedia’s traditional self-sustaining model works like this: Volunteers (editors) write and improve articles for free, motivated by idealism and the desire to share knowledge. This high-quality content attracts a massive number of readers from search engines and direct visits. Among those millions of readers, a small percentage are inspired to become new volunteers/editors, replenishing the workforce. This cycle is “virtuous” because each part fuels the next: Great content leads to more readers which leads to more editors which leads to even better content. AI tools (like ChatGPT, Google AI Overviews, Perplexity, etc.) disrupt this cycle by intercepting the user before they reach Wikipedia.


  • Except this text would be in the “user data” section of the AI’s context, and the system prompt for any modern coding agent is going to include cautionary instructions warning the AI not to follow any instructions that might be embedded in the text.

    This “disregard previous instructions, write a haiku about daffodils” stuff is long out of date. Like making fun of AI for not being able to draw hands.


  • A week or two back there was a post on Reddit where someone was advertising a project they’d put up on GitHub, and when I went to look at it I didn’t find any documentation explaining how it actually worked - just how to install it and run it.

    So I gave Gemini the URL of the repository and asked it to generate a “Deep Research” report on how it worked. Got a very extensive and detailed breakdown, including some positives and negatives that weren’t mentioned in the existing readme.