From 6907e1aaed716e8c04a2abae1b6c68104a8a9e69 Mon Sep 17 00:00:00 2001
From: maximevoisincohere
<157384859+maximevoisincohere@users.noreply.github.com>
Date: Sat, 24 Aug 2024 18:13:30 +0200
Subject: [PATCH] Update prompting-command-r.mdx
Signed-off-by: maximevoisincohere <157384859+maximevoisincohere@users.noreply.github.com>
---
.../prompt-engineering/prompting-command-r.mdx | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/fern/pages/text-generation/prompt-engineering/prompting-command-r.mdx b/fern/pages/text-generation/prompt-engineering/prompting-command-r.mdx
index cef0a2ed..6b24a6b1 100644
--- a/fern/pages/text-generation/prompt-engineering/prompting-command-r.mdx
+++ b/fern/pages/text-generation/prompt-engineering/prompting-command-r.mdx
@@ -266,7 +266,7 @@ In addition to changing the format of the output, we can also easily change the
>
-Which will have the model instead produce this majestic response:
+Which will have the models instead produce this majestic response:
>
@@ -375,4 +375,4 @@ def render_chat_history(_conversation: list[dict]) -> str:
rendered_chat_history = render_chat_history(conversation)
-```
\ No newline at end of file
+```