Skip to content

Commit

Permalink
Address feedback
Browse files Browse the repository at this point in the history
  • Loading branch information
timrogers authored Feb 23, 2024
1 parent b944fea commit ec44b66
Showing 1 changed file with 1 addition and 9 deletions.
10 changes: 1 addition & 9 deletions data/reusables/rai/copilot/about-copilot-chat.md
Original file line number Diff line number Diff line change
Expand Up @@ -14,15 +14,7 @@ The language model generates a response based on its analysis of the input promp

The response generated by {% data variables.product.prodname_copilot_chat_short %} is formatted and presented to the user. {% data variables.product.prodname_copilot_chat_short %} may use syntax highlighting, indentation, and other formatting features to add clarity to the generated response. Depending upon the type of question from the user, links to context that the model used when generating a response, such as source code files or documentation, may also be provided.

{% data variables.product.prodname_copilot_chat %} is intended to provide you with the most relevant answer to your question. However, it may not always provide the answer you are looking for. Users of {% data variables.product.prodname_copilot_chat_short %} are responsible for reviewing and validating responses generated by the system to ensure they are accurate and appropriate. {% data variables.product.prodname_copilot_chat_short %} is also designed to learn from your feedback and improve over time. For more information on improving the performance of {% data variables.product.prodname_copilot_chat %}, see "[Improving performance for {% data variables.product.prodname_copilot_chat %}](/copilot/github-copilot-chat/about-github-copilot-chat#improving-performance-for-github-copilot-chat)."

### Red teaming

As part of our product development process, we implement adversarial prompt measures and human review outputs to flag potential challenges.

### Responsible AI (RAI) filtering

Input prompts and output completions are run through content filters trained on detecting and preventing the output of harmful, offensive, or infringing content. The content filtering system detects and takes action on specific categories of potentially harmful content.
{% data variables.product.prodname_copilot_chat %} is intended to provide you with the most relevant answer to your question. However, it may not always provide the answer you are looking for. Users of {% data variables.product.prodname_copilot_chat_short %} are responsible for reviewing and validating responses generated by the system to ensure they are accurate and appropriate. Additionally, as part of our product development process, we undertake red teaming to understand and improve the safety of {% data variables.product.prodname_copilot_chat_short %}. Input prompts and output completions are run through content filters. The content filtering system detects and prevents the output on specific categories of content including harmful, offensive, or off-topic content. {% data variables.product.prodname_copilot_chat_short %} is also designed to learn from your feedback and improve over time. For more information on improving the performance of {% data variables.product.prodname_copilot_chat %}, see "[Improving performance for {% data variables.product.prodname_copilot_chat %}](/copilot/github-copilot-chat/about-github-copilot-chat#improving-performance-for-github-copilot-chat)."

## Use cases for {% data variables.product.prodname_copilot_chat %}

Expand Down

0 comments on commit ec44b66

Please sign in to comment.