• Aria@lemmygrad.ml
    link
    fedilink
    arrow-up
    14
    ·
    1 month ago

    There’s no link. Did they? Was it trained on ChatGPT answers? Because a lot of small team LLMs are trained on ChatGPT answers.

      • Aria@lemmygrad.ml
        link
        fedilink
        arrow-up
        17
        ·
        1 month ago

        Yeah okay so they’re literally just talking about using ChatGPT answers in the training set. Everyone does this, it’s completely fair game. Hell, the R1 Qwen that everyone is using now to run ‘R1’ on their own device is Qwen distilled with R1.

        (Not that I believe in copyright to begin with).

        • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
          link
          fedilink
          arrow-up
          10
          ·
          1 month ago

          Exactly and it’s just stunning hypocrisy on the part of OpenAI who have been sucking up all data with complete disregard for ownership. Also agree that copyright wouldn’t even be necessary in a socialist society, while it only ends up being abused under capitalism.

          • WhatWouldKarlDo@lemmygrad.ml
            link
            fedilink
            English
            arrow-up
            5
            ·
            1 month ago

            I would argue that a new socialist country should retain copyright, but drastically shorten its duration (back to the original 7 years or less). The reason being that copyright is the mechanism that protects the copyleft. Without copyright, the GPL would no longer be enforceable. I’m no fan of copyright and would love to see it eventually abolished, but it has its uses today.

        • cfgaussian@lemmygrad.ml
          link
          fedilink
          arrow-up
          9
          ·
          edit-2
          1 month ago

          Yeah the real problem with this is that ChatGPT is going to give a lot of garbage answers when it comes to political questions, and if DeepSeek takes in too much of that, it will also start spitting out very bad takes. It’s like copying off of your classmate in a test in school, but that classmate is only known for getting good grades in some classes but not so good grades in others. You need to be careful to only copy on the subjects you know they’re actually competent at.

      • ViVerVo@lemmygrad.ml
        link
        fedilink
        English
        arrow-up
        11
        ·
        1 month ago

        Lol, what’s ClosedAI going to do? Hide ChatGPT’s answers behind a summary (like with the CoT) 🤣

          • ViVerVo@lemmygrad.ml
            link
            fedilink
            English
            arrow-up
            0
            ·
            1 month ago

            From Deepseek:

            Chain of Thought (CoT) in LLMs refers to a prompting technique that guides large language models to articulate intermediate reasoning steps when solving a problem, mimicking human-like logical progression. Here’s a concise breakdown:

            1. Purpose: Enhances performance on complex tasks (e.g., math, logic, commonsense reasoning) by breaking problems into sequential steps, reducing errors from direct, unstructured answers.

            2. Mechanism:

              • Prompting Strategy: Users provide examples (few-shot) or explicit instructions (zero-shot, e.g., “Let’s think step by step”) to encourage step-by-step explanations.
              • Output Structure: The model generates tokens sequentially, simulating a reasoned pathway to the answer, even though internal processing remains parallel (via transformer architecture).
            3. Benefits:

              • Accuracy: Improves results on multi-step tasks by isolating and addressing each component.
              • Transparency: Makes the model’s “thinking” visible, aiding debugging and trust.
            4. Variants:

              • Few-Shot CoT: Examples with detailed reasoning are included in the prompt.
              • Zero-Shot CoT: Direct instructions trigger step-by-step output without examples.
              • Self-Consistency: Aggregates answers from multiple CoT paths to select the most consistent one.
            5. Effectiveness: Particularly impactful for tasks requiring structured reasoning, while less critical for simple queries. Research shows marked accuracy gains in benchmarks like math word problems.

            In essence, CoT leverages the model’s generative capability to externalize reasoning, bridging the gap between opaque model decisions and interpretable human problem-solving.

            Example: