• 0 Posts
  • 317 Comments
Joined 1 year ago
cake
Cake day: June 22nd, 2023

help-circle


  • The difference is that Uber’s model of using an app to show you the route, give driver feedback, be able to report problems and monitor and track the driver, etc. is actually a huge improvement to both rider safety and experience compared to calling a cab company and then waiting who knows how long for someone to show up and hopefully bring you where you want to go.

    Not saying that their model of gig workers, or dodging up front training is good, but they legitimately offered up a fundamentally better taxi experience than anything that came before, which I think encouraged regulators to really drag their feet on looking into them.






  • They claim he made a threat. The article failed to print his side of the story for some curious reason. It isn’t printing any testimony from the bystanders, either.

    Fair enough, supposedly they were wearing body cams so hopefully some of what actually happened can be answered objectively, I’m just pointing out what the article said. If he didn’t make a threat or have a knife, then tasering him is a wild escalation, it’s just that if he did, then the police can’t really just let him get on a train.

    Cops will often lie about the danger of a suspect in order to justify elevating their use-of-force. That said, they weren’t that concerned by his unreasonableness when they deployed tasers into the crowd first. They didn’t switch to guns until they realized the tasers weren’t going to work.

    Again, assuming what the article says is true, which is a big assumption, it’s not that crazy to taser a guy who just got onto a train with a knife and threatened to you. At that point you’re looking at a potential mass stabbing incident if you do nothing.

    Again, who knows, maybe the cops are blowing his behaviour wildly out of proportion, I’m just saying that, based on the article, it sounds like he wasn’t just gunned down for jumping a turnstile.





  • “This isn’t a meeting about the budget per se”

    “This isn’t exactly a meeting about the budget”

    If you finish those sentences, it becomes clear why per se is used:

    “This isn’t a meeting about the budget per se, it’s a meeting about how much of the budget is spent on bits of string”

    “This isn’t exactly a meeting about the budget, it’s a meeting about how much of the budget is spent on bits of string”

    In this situation, using per se provides a more natural sentence flow because it links the first part of the sentence with the second. It’s also shorter and fewer syllables.

    “Steve’s quite erudite.”

    “Steve’s quite intellectual.”

    I think intellectual might be a closer synonym, but intellectual often has more know-it-all connotations than erudite which seems to often refer to a more pure and cerebral quality.

    “Tom and Jerry is a fun cartoon because of the juxtaposition of the relationship between cat and mouse.”

    “Tom and Jerry is a fun cartoon because of the side by side oppositeness of the relationship between cat and mouse that is displayed

    For those to say precisely the same thing it would have to be more like the above which doesn’t really roll off the tongue.

    “I don’t understand, can you elucidate that?”

    “I don’t understand, can you explain?”

    Elucidate just means to make something clear in general, explaining something usually inherently implies a linguistic, verbal, explanation, unless otherwise stated.

    Honestly, these all seem like very reasonable words to me for the most part. I can understand not using them in some contexts, but for the most part, words exist for a reason, to describe something slightly differently, and it takes forever to talk and communicate if we only limit ourselves to the most basic unnuanced terms.


  • When people use industry specific jargon and acronyms with someone not in their industry.

    It is a very simple rule of writing and communication. You never just use an acronym out of nowhere, you write it out in full the first time and explain the acronym, and then after that you can use it.

    Artificial diamonds can be made with a High Temperature, High Pressure (HTHP) process, or a …

    Doctors, military folk, lawyers, and technical people of all variety are often awful at just throwing out an acronym or technical term that you literally have no way of knowing.

    Usually though, I don’t think it’s a conscious effort to sound smart. Sometimes, it’s just people who are used to talking only with their coworkers / inner circle and just aren’t thinking about the fact that you don’t have the same context, sometimes it’s people who are feeling nervous / insecure and are subconsciously using fancy terms to sound like they fit in, and sometimes it’s people using specific terminology to hide the fact that they don’t actually understand the concepts well enough to break them down further.



  • The work is reproduced in full when it’s downloaded to the server used to train the AI model, and the entirety of the reproduced work is used for training. Thus, they are using the entirety of the work.

    That’s objectively false. It’s downloaded to the server, but it should never be redistributed to anyone else in full. As a developer for instance, it’s illegal for me to copy code I find in a medium article and use it in our software. I’m perfectly allowed to read that Medium article, learn from it, and then right my own similar code.

    And that makes it better somehow? Aereo got sued out of existence because their model threatened the retransmission fees that broadcast TV stations were being paid by cable TV subscribers. There wasn’t any devaluation of broadcasters’ previous performances, the entire harm they presented was in terms of lost revenue in the future. But hey, thanks for agreeing with me?

    And Aero should not have lost that suit. That’s an example of the US court system abjectly failing.

    And again, LLM training so egregiously fails two out of the four factors for judging a fair use claim that it would fail the test entirely. The only difference is that OpenAI is failing it worse than other LLMs.

    That’s what we’re debating, not a given.

    It’s even more absurd to claim something that is transformative automatically qualifies for fair use.

    Fair point, but it is objectively transformative.



  • You said open source. Open source is a type of licensure.

    The entire point of licensure is legal pedantry.

    No. Open source is a concept. That concept also has pedantic legal definitions, but the concept itself is not inherently pedantic.

    And as far as your metaphor is concerned, pre-trained models are closer to pre-compiled binaries, which are expressly not considered Open Source according to the OSD.

    No, they’re not. Which is why I didn’t use that metaphor.

    A binary is explicitly a black box. There is nothing to learn from a binary, unless you explicitly decompile it back into source code.

    In this case, literally all the source code is available. Any researcher can read through their model, learn from it, copy it, twist it, and build their own version of it wholesale. Not providing the training data, is more similar to saying that Yuzu or an emulator isn’t open source because it doesn’t provide copyrighted games. It is providing literally all of the parts of it that it can open source, and then letting the user feed it whatever training data they are allowed access to.


  • LLMs use the entirety of a copyrighted work for their training, which fails the “amount and substantiality” factor.

    That factor is relative to what is reproduced, not to what is ingested. A company is allowed to scrape the web all they want as long as they don’t republish it.

    By their very nature, LLMs would significantly devalue the work of every artist, author, journalist, and publishing organization, on an industry-wide scale, which fails the “Effect upon work’s value” factor.

    I would argue that LLMs devalue the author’s potential for future work, not the original work they were trained on.

    Those two alone would be enough for any sane judge to rule that training LLMs would not qualify as fair use, but then you also have OpenAI and other commercial AI companies offering the use of these models for commercial, for-profit purposes, which also fails the “Purpose and character of the use” factor.

    Again, that’s the practice of OpenAI, but not inherent to LLMs.

    You could maybe argue that training LLMs is transformative,

    It’s honestly absurd to try and argue that they’re not transformative.