• 0 Posts
  • 78 Comments
Joined 2 years ago
cake
Cake day: March 3rd, 2024

help-circle
  • i personally have pushed back on every “infinite scrolling” feature request from product designers. first, you think you need it; you don’t. second, you think it’s just so nifty! it isn’t. oh is your content is dynamically generated? what was wrong with Reddit’s pager that launched that site into popularity?

    it’s unnecessary complexity that hides information from the user, makes API calls (which are, spoilers, paginated) more complicated, can cause the obvious memory/resource consumption issues, and just generally disempowers the user. which i guess on a social media app is the point. but totally counter to the goals of a fleet management system lol





  • i think it’s easy to make comments like this from the peanut gallery, with the benefit of hindsight and a self-selected group of users who will agree. but Apple should be legally obligated to address this. the solution can’t be “this idiot didn’t spend his nights and weekends doing 3-tier backups and high availability infrastructure diversity!”; that’s not scalable. if we just accept that companies can do this, they will continue to. but this has been on the front page of HackerNews. it’ll probably make it to Tim Apple’s desk eventually, so we’ll see what shakes out.






  • my point is that it’s hard to program someone’s subjective, if written in whatever form of legalese, point of view into a detection system, especially when those same detection systems can be used to great effect to train systems to bypass them. any such detection system would likely be an “AI” in the same way the ones they ban are and would be similarly prone to mistakes and to reflecting the values of the company (read: Jack Dorsey) rather than enforcing any objective ethical boundary.








  • but LLMs do represent a significant technological leap forward. i also share the skepticism that we haven’t “cracked AGI” and that a lot of these products are dumb. i think another comment made a better analogy to the dotcom bubble.

    ETA: i’ve been working in ML engineering since 2019, so i can sometimes forget most people didn’t even hear about this hype train until ChatGPT, but i assure you inference hardware and dumb products were picking up steam even then (Tesla FSD being a classic example).


  • i know it’s popular to be very dismissive, but a lot of “AI” has already been integrated into normal workflows. AI autocomplete in development text editors, software keyboards, and question asking bots isn’t going away. speech-to-text, “smart eraser”, subject classification, signal processing kernels like DLSS and frame generation, and so many more will be with us and improving for a long time. Transformers, machine learning optimized chips, and other ML fields are going to be with us for a long time. the comparison to NFTs is either angst or misunderstanding.