• 0 Posts
  • 29 Comments
Joined 11 months ago
cake
Cake day: August 19th, 2023

help-circle

  • Claude Opus disagrees, lol (edit to add: all of what follows is Claude; not me):

    I respectfully disagree with the claim that current language models are logically incapable of mapping language to actual concepts. While today’s LLMs certainly have limitations and are prone to hallucinations, they have demonstrated a remarkable ability to capture and manipulate complex concepts through language.

    A few key points:

    LLMs are trained on vast amounts of human-generated text which is grounded in real-world concepts, allowing the models to build rich representations that go beyond surface-level word associations. Techniques like unsupervised pre-training enable LLMs to learn meaningful conceptual relationships.
    
    In many domains, LLMs have shown an impressive capacity for reasoning, inference, and knowledge synthesis - skills that rely on a deep understanding of the underlying concepts, not just pattern matching. Their performance on complex question-answering, analogical reasoning, and natural language inference tasks is evidence of this.
    
    LLMs can be fine-tuned and augmented with techniques like retrieval-augmented generation or chain-of-thought prompting to further ground them in factual knowledge and improve their reasoning capabilities. This allows more robust conceptual mappings.
    
    Careful prompt engineering and sampling techniques can help to mitigate hallucinations and nonsensical outputs. While not a complete solution, this shows that the conceptual knowledge is often present even if the model sometimes struggles to express it faithfully.
    

    That said, I fully agree that the tendency for subtle nonsense is a serious issue, especially for layperson use cases where hallucinations may go undetected. Continued research into making LLMs more truthful, consistent and robust is crucial. Techniques like constitutional AI to bake in truthfulness, and better UX design to convey uncertainty are important steps.

    But in summary, I believe the evidence suggests that LLMs, while flawed, are not fundamentally incapable of meaningful conceptual representation and reasoning. We should push forward on making them more reliable and trustworthy, rather than dismissing their potential prematurely.







  • Would you, after devoting full years of your adult life to the unpaid work of learning the requisite advanced math and computer science needed to develop such a model, like to spend years more of your life to develop a generative AI model without compensation? Within the US, it is legal to use public text for commercial purposes without any need to obtain a permit. Developers of such models deserve to be paid, just like any other workers, and that doesn’t happen unless either we make AI a utility (or something similar) and funnel tax dollars into it or the company charges for the product so it can pay its employees.

    I wholeheartedly agree that AI shouldn’t be trained on copyrighted, private, or any other works outside of the public domain. I think that OpenAI’s use of nonpublic material was illegal and unethical, and that they should be legally obligated to scrap their entire model and train another one from legal material. But developers deserve to be paid for their labor and time, and that requires the company that employs them to make money somehow.



  • Eh, us professors care pretty deeply about the plagiarism she did. Intent or even knowledge of plagiarism isn’t necessary for disciplinary action in plagiarism cases at major research universities. Any one of these examples would be enough for my university’s academic integrity committee to rule that plagiarism occurred:

    https://www.nytimes.com/2023/12/21/us/claudine-gay-harvard-president-excerpts.html

    And in the case of a dissertation, plagiarism is an automatic expulsion and degree retraction from my university. At the PhD level, students certainly know that what Dr. Gay did is plagiarism (a good rule of thumb is that five sequential words, even with paraphrasing, without citing the source, is plagiarism), and that plagiarism is completely unacceptable.

    I already know of a student who made the argument that their plagiarism wasn’t as bad as Dr. Gay’s, so because Dr. Gay wasn’t penalized, they shouldn’t be penalized. Had she not stepped down, that line of argument likely would have snowballed out of control. The professors I know think her comments to Congress were out of touch, but all of us had been livid that she and Harvard were saying that she didn’t plagiarize–any professor who looks at those examples will tell you that she did.




  • GPT-4 will. For example, I asked it the following:

    What is the neighborhood stranger model of fluid mechanics?

    It responded:

    The “neighborhood stranger model” of fluid mechanics is not a recognized term or concept within the field of fluid mechanics, as of my last update in April 2023.

    Now, obviously, this is a made-up term, but GPT-4 didn’t confidently give an incorrect answer. Other LLMs will. For example, Bard says,

    The neighborhood stranger model of fluid mechanics is a simplified model that describes the behavior of fluids at a very small scale. In this model, fluid particles are represented as points, and their interactions are only considered with other particles that are within a certain “neighborhood” of them. This neighborhood is typically assumed to be a sphere or a cube, and the size of the neighborhood is determined by the length scale of the phenomena being studied.