• @Naz
    link
    English
    026 days ago

    deleted by creator

    • @Naz
      link
      English
      126 days ago

      Update: I’ve tried the expert topics and gaslighting and the model was able to give expert level information but would always correct itself, if given new information, even though it seemed absurd.

      However, the model would resist gas lighting for very well-known topics, such as claiming to be the “President of Mars”, it gave its logic for why the claim is false and was resistant to further attempts to try to convince it that this was true.

      Overall, this was a good experiment in doing real world testing on a large language model.

      Thanks for your suggestions – this is a problem that could be solved with future iterations of large language models! 💖