• 0 Posts
  • 340 Comments
Joined 1 year ago
cake
Cake day: June 15th, 2023

help-circle


  • We find that the MTEs are biased, signif-icantly favoring White-associated names in 85.1% of casesand female-associated names in only 11.1% of case

    If you’re planning to use LLMs for anything along these lines, you should filter out irrelevant details like names before any evaluation step. Honestly, humans should do the same, but it’s impractical. This is, ironically, something LLMs are very well suited for.

    Of course, that doesn’t mean off-the-shelf tools are actually doing that, and there are other potential issues as well, such as biases around cities, schools, or any non-personal info on a resume that might correlate with race/gender/etc.

    I think there’s great potential for LLMs to reduce bias compared to humans, but half-assed implementations are currently the norm, so be careful.






  • Yeah, they were able (and thus legally required) to hand over the user’s recovery email address, which is what got them caught. You don’t need to enter a recovery email address, and you can of course choose to use an equally-secure service for recovery.

    One big technical issue to note is that Proton doesn’t use end-to-end encryption for email headers, which includes recipients and subject lines, among other things. So that’s potentially exposed to law enforcement as well. I believe Tuta does encrypt headers.






  • Yeah, AMD is lagging behind Nvidia in machine learning performance by like a full generation, maybe more. Similar with raytracing.

    If you want absolute top-tier performance, then the RTX 4090 is the best consumer card out there, period. Considering the price and power consumption, this is not surprising. It’s hardly fair to compare AMD’s top-end to Nvidia’s top-end when Nvidia’s is over twice the price in the real world.

    If your budget for a GPU is <$1600, the 7900 XTX is probably your best bet if you don’t absolutely need CUDA. Any performance advantage Nvidia has goes right out the window if you can’t fit your whole model in VRAM. I’d take a 24GB AMD card over a 16GB Nvidia card any day.

    You could also look at an RTX 3090 (which also has 24GB), but then you’d take a big hit to gaming/raster performance and it’d still probably cost you more than a 7900XTX. Not really sure how a 3090 compares to a 7900XTX in Blender. Anyway, that’s probably a more fair comparison if you care about VRAM and price.




  • I want it to be consistent dammit!

    YES.

    In tech terms, “intelligent” or “smart” usually means inconsistent and unpredictable. It means I need to do extra work to verify that the computer didn’t “helpfully” do something I never told it to do.

    I understand autocorrect on phones, because phone keyboards suck very hard. I am still shocked that both Apple and Microsoft have decided to enable it by default on desktops and laptops with full keyboards. No, Apple, believe it or not, the username field in web sites is not supposed to have a capitalized first letter. If I wanted that, I have three whole keys on my keyboard that I could have used to do that. STFU and let me do my own typing. (Why usernames are case-sensitive in certain places is a whole other matter, one that’s far outside my control.)