Copilot purposely stops working on code that contains hardcoded banned words from Github, such as gender or sex. And if you prefix transactional data as trans_ Copilot will refuse to help you. 😑

  • UndercoverUlrikHD@programming.dev
    link
    fedilink
    English
    arrow-up
    1
    ·
    5 days ago

    I wrote a slur detection script for lemmy, copilot refused to run unless I removed the “common slurs” list from the file. There are definitely keywords or context that will shut down the service. Could even be regionally dependant.