• 0 Posts
  • 47 Comments
Joined 2 years ago
cake
Cake day: June 12th, 2023

help-circle


  • Really? I mean, it’s melodramatic, but if you went throughout time and asked writers and intellectuals if a machine could write poetry, solve mathmatical equations, and radicalize people effectively t enough to cause a minor mental health crisis, I think they’d be pretty surprised.

    LLMs do expose something about intelligence, which is that much of what we recognize as intelligence and reason can be distilled from sufficiently large quantities of natural language. Not perfectly, but isn’t it just the slightest bit revealing?


  • A child may hallucinate, lie, misunderstand, etc, but we wouldn’t say the foundations of a complete adult are not there, and we wouldn’t assess the child as not conscious. I’m not saying that LLMs are conscious because they say so (they can be made to say anything), but rather that it’s difficult to be confident that humans possess some special spice of consciousness that LLMs do not, because we can also be convinced to say anything.

    LLMs can reason (somewhat unreliably) with a fraction of a human brains compute power while running on hardware that was made for graphics processing. Maybe they are conscious, but only in some pathetically small way, which will only become evident when they scale up, like a child.



  • Why can’t complex algorithms be conscious? In fact, ai can be directed to reason about themselves, context can be made to be persistent, and we can measure activation parameters showing that they are doing so.

    I’m sort of playing devil’s advocate here, but, “Consciousness requires contemplation of self. Which requires the ability to contemplate.” Is subjective, and nearly any ai model, even rudimentary ones, are capable of insisting that they contemplate themselves.






  • Yes, sorry, where I live it’s pretty normal for cars to be diesel powered. What I meant by my comparison was that a train, when measured uncritically, uses more energy to run than a car due to it’s size and behavior, but that when compared fairly, the train has obvious gains and tradeoffs.

    Deepseek as a 600b model is more efficient than the 400b llama model (a more fair size comparison), because it’s a mixed experts model with less active parameters, and when run in the R1 reasoning configuration, it is probably still more efficient than a dense model of comparable intelligence.



  • This article is comparing apples to oranges here. The deepseek R1 model is a mixture of experts, reasoning model with 600 billion parameters, and the meta model is a dense 70 billion parameter model without reasoning which preforms much worse.

    They should be comparing deepseek to reasoning models such as openai’s O1. They are comparable with results, but O1 cost significantly more to run. It’s impossible to know how much energy it uses because it’s a closed source model and openai doesn’t publish that information, but they charge a lot for it on their API.

    Tldr: It’s a bad faith comparison. Like comparing a train to a car and complaining about how much more diesel the train used on a 3 mile trip between stations.




  • I mean no offense, but it sounds like you have poorly developed social skills. I used to as well.

    You could try reframing it in your mind:

    It’s not faking, it’s practice.

    If you pick up an instrument for the first time to practice, you will sound terrible, and possibly be discouraged, but if you practice for hundreds of hours you’ll be able to play it for real.

    Babies and children aren’t born knowing how to express interest or sympathize. You certainly weren’t. Children have to learn how to do this. It is possible that you need to practice if you want to build intimate relationships. There is no shortcut to this.


  • peanuts4life@lemmy.blahaj.zonetoAsk Lemmy@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    9
    arrow-down
    2
    ·
    6 months ago

    (35, he/him) This is how I met my first girlfriend, in reverse. I was lonely and had mentally committed to making a friend in a manic episode. I saw her on a bench reading and asked if I could sit next to her. I had a book with me too, and began to read. Then, I asked what she was reading. We became friends, and later dated for 2.5 years.

    I’ve spent a lot of time wandering around hoping people would talk to me. I used to feel like it was worthless, because 95% of the time no one will talk to you, but those odds aren’t so bad in hindsight. Go into public 100 times a year and you’ll have 5 decent shots at making a friend. Make one friend a year, and you’ll probably have more social opportunities than you want to deal with.

    I’ve met people randomly in public like this perhaps 6 times.

    There are other factors other than randomness:

    1. I’m very friendly to people. I like to ask questions once a conversation gets going, and I get animated on just about any topic. I talk to myself a lot, so even when I’m not exposed to people I’m practiced, in a way.

    2. There is usually an activity involved. Reading a book together, drawing on an airplane, posting art on a blog, taking classes together, being at the same work event, hiding in the same hard to find corner of the library. These are all situations from my life, and they typically involve a shared activity, or a creative outlet. This is probably why people recommend joining clubs / going to bars, advice I’ve never taken, but I see the reasoning.

    I don’t mean to project that my social life is great! I’ve been terribly lonely during much of it, and these experiences I’m describing took place over several years. However, if I could boil down my successes, I’d say they cultivating a curiousty in others and publically engaging in my hobbies has been the best way to make friends (and occasionally lovers).


  • I’m not a medical expert, but typically your skin is pretty good at keeping pathogens out, at least when you clean yourself regularly.

    Keeping your hands clean (and thus your eyes and mouth) should be a priority. A relatively cheap and easy thing to focus on would be cheap, disposable gloves. Buy them in bulk, and carry a wad in a pocket. You can turn old ones inside out and use them as little trash bags. Change them out whenever you feel like it.

    If you don’t have a clothes washer at home, consider buying a cheap portable unit that can drain into the shower or sink. I have one, and it rocks.

    You might feel a little crazy, but if you have the spare cash, buy some bleach spray and paper towels and wipe down the elevator. It should only take 5 minutes. It could be the case that it only needs to be cleaned every few weeks.

    Remember that by keeping your space and person clean, you are doing a lot to stay healthy. People work draining septic tanks for a living and are exposed to sewage, but stay healthy because of good hygiene habits in the long term. I don’t mean to minimize your situation, because I’d feel crazy too, but just keep in mind you’re already doing a good job.



  • https://openai.com/index/how-openai-is-approaching-2024-worldwide-elections/

    Here is a direct quote from openai:

    “In addition to our efforts to direct people to reliable sources of information, we also worked to ensure ChatGPT did not express political preferences or recommend candidates even when asked explicitly.”

    It’s not a conspiracy. It was explicitly thier policy not to have the ai discuss these subjects in meaningful detail leading up to the election, even when the facts were not up for debate. Everyone using gpt during that period of time was unlikely to receive meaningful information on anything Trump related, such as the legitimacy of Biden’s election. I know because I tried.

    This is ostentatiously there to protect voters from fake news. I’m sure it does in some cases, but I’m sure China would say the same thing.

    I’m not pro China, I’m suggesting that every country engages in these shenanigans.

    Edit it is obvious that a 100 billion dollar company like openai with it’s multude of partnerships with news companies could have made gpt communicate accurate and genuinely critical news regarding Trump, but that would be bad for business.