• 0 Posts
  • 33 Comments
Joined 2 years ago
cake
Cake day: June 29th, 2023

help-circle

  • I don’t think the term AI has been used in a vague way, it’s that there’s a huge disconnect between how the technical fields use it vs general populace and marketing groups heavily abuse that disconnect.

    Artificial has two meanings/use cases. One is to indicate something is fake (video game NPC, chess bots, vegan cheese). The end product looks close enough to the real thing that for its intended use case it works well enough. Looks like a duck, quacks like a duck, treat it like a duck even though we all know it’s a bunny with a costume on. LLMs on a technical level fit this definition.

    The other definition is man made. Artificial diamonds are a great example of this, they’re still diamonds at the end of the day, they have all the same chemical makeups, same chemical and physical properties. The only difference is they came from a laboratory made by adult workers vs child slave labor.

    My pet theory is science fiction got the general populace to think of artificial intelligence to be using the “man-made” definition instead of the “fake” definition that these companies are using. In the past the subtle nuance never caused a problem so we all just kinda ignored it


  • I think a better definition would be “achieve something in an unintended or uncommon way”. Fits the bill on what generally passes in the tech community as a “hack” while also covering some normal life stuff.

    Getting a cheaper flight booked by using a IP address assigned to a different geographical location? Sure I’d call that a life hack. Getting a cheaper flight by booking a late night, early morning flight? No, those are deliberately cheaper

    Also re: your other comment about not making a reply at all, sometimes for people like us it’s just better to not get into internet fights over semantics (no matter how much fun they can be)


  • That’s kind of the point though, isn’t it?

    If I were to post with “Extend the plank!” there’s a near zero chance that even fans of the movie, or even the franchise, I’m thinking of will get the movie right. If I instead say “Who am I to argue with the Captain of the Enterprise” a normie might guess Star Trek, a true nerd and fan of the franchise will peg that instantly as from Star Trek Generations

    Edit: That said, there are several lines in this thread that aren’t necessarily only recognizable to fans or people familiar with the movie, but instead just pop culture references.



  • You don’t. In C everything gets referenced by a symbol during the link stage of compilation. Libraries ultimately get treated like your source code during compilation and all items land in a symbol table. Two items with the same name result in a link failure and compilation aborts. So a library and a program with main is no bueno.

    When Linux loads an executable they basically look at the program’s symbol table and search for “main” then start executing at that point

    Windows behaves mostly the same way, as does MacOS. Most RTOS’s have their own special way of doing things, bare metal you’re at the mercy of your CPU vendor. The C standard specifies that “main” is the special symbol we all just happen to use



  • I hadn’t thought of that before, and I can think of several characters who’ve said things I doubt the writers would want attributed to them. I just want to see quotes from fiction being clearly labeled as such, and not using the grandiose of a character’s title to add weight to the quote.

    For example when I see people quote Admiral William Adama on how when the military becomes the police, the people become the enemy of the state. That was Ron Moore writing a character for a show set in a post apocalyptic universe where the only survivors are hanging out on military ships, not a real world seasoned officer’s opinion. Is it an interesting point worth discussing? Sure, but I’m not putting it in the same category of 5-Star General Dwight Eisenhower’s warnings about the military industrial complex










  • Realistically no. The support needed to manage the devices we all use is just insane, and I think a lot of people take for granted how the x86 platform has evolved over the last few decades. The ARM landscape does not have the standards set that x86 does and that will always hold it back. Qualcomm learned long ago that it’s within their best interest to be constantly changing the SoCs and never really documenting/supporting them very well because it forces all of the downstream vendors to do constant refreshes. Toss in the development hellscape my fellow programmers created ourselves and we get the vicious cycle we’re in today where Google saying they’ll support a device for longer than a few years was the headline sales pitch

    -typed on a Pixel 8 which was purchased due to that sales pitch




  • Heh, I guess this shows my corporate software dev experience. Whenever I’ve taught git workflows it was always paired with a work ticketing system where any changes you were making were ideally all one single set of changes. If you need a feature or bug fix someone else was doing that was being done on another branch which you could pull into your code early and for tracking purposes we always made sure the other person merged into main first. The only time I’ve seen per line manipulation with git was when someone made a ton of changes in a file and wanted to revert a handful of lines.

    Everything else you mentioned I’ve had a web git host like gitlab or bitbucket for, but I kinda put that more into peer review workflow than git itself