• 0 Posts
  • 22 Comments
Joined 2 years ago
cake
Cake day: September 6th, 2024

help-circle

  • I’m going to enjoy torturing my 14-year-old self. My 14-year-old self was a shithead. But I was raised in a conservative Catholic house, and at that age I firmly embraced the version of reality common among the Fox News set. I was that annoying conservative high schooler. Sure I was repping hard, but I was still an idiot.

    Now I’m a late-30s trans woman, about to celebrate 8 years of marriage to my wonderful husband.

    The things I can say. I’m going to haunt this kid’s dreams.



  • You asked a question no one can answer.

    Instead of asking impossible questions, I suggest just using a bit of logic. Officially, YouTube removed the like/dislike because they felt people were prejudging videos before viewing them themselves. Unofficially, people speculate they did it to have greater control of what people watch. But in either case, such a change would only make sense if plenty of people were checking the ratio prior to viewing. If no one ever paid attention to it, then there wouldn’t be anything to be gained by tampering with it.



  • The real issue is that since any fingerprint that can be mandated for AI content must be algorithmically implemented, then that fingerprint can be algorithmically removed.

    For example, let’s say companies voluntarily choose or are forced to integrate text fingerprinting into LLM output. Automated AI writing detection tools already exist, but they’re not reliable. But in principle we could make the output of LLMs easy to identify. Maybe we force them to adopt subtle but highly unique patterns of word choice, punctuation, sentence structure, etc. Then if any student attempted to upload an LLM-generated essay to their course website, the system could with high accuracy flag it as AI generated.

    But…if those patterns are so clear and unambiguous, it also means they can be easily detected by third party tools. If one person can code ChatGPT to add special fingerprinting to the text ChatGPT creates, another person can create a program that you can paste ChatGPT text into that will remove that fingerprinting.




  • I mean, what exactly is wrong with it? Age gap aside, I really don’t see anything wrong with say a young faculty member getting with an undergrad. Imagibe a prof in their late twenties and an undergrad in their early twenties. As long as the student isn’t one of their current or likely future students, I see nothing morally wrong with it. Now if it’s a 50 year old prof with a 19 year old student, that’s a different matter. But the problem there is the age gap, not the prof/student status.





  • This just doesn’t pass the smell test to me. Why would they need to go to the trouble? It’s not like they can’t figure out who is trans through name and gender change court records, health records, etc. Even if these aren’t gathered in one list anywhere, it wouldn’t be hard to have an AI run through name change records and flag anyone that changed their name from a typically masculine to a typically feminine name, and vice versa.

    Sure, there are some very paranoid/privacy-minded trans folks that never put their transness in government or medical records. They never change their name legally, they DIY HRT, etc. And those folks would be difficult to find. However, I imagine most such folks would be far too cautious to join a trans-specific dating app.