Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[pull] main from openai:main #7

Merged
merged 5 commits into from
May 6, 2023
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
polishes wording in a few places
  • Loading branch information
ted-at-openai committed May 5, 2023
commit b59f105ed16be7922baf5b003f7e815d92c0c95c
10 changes: 5 additions & 5 deletions examples/Question_answering_using_embeddings.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@
"\n",
"Specifically, this notebook demonstrates the following procedure:\n",
"\n",
"1. Prepare search data (once)\n",
"1. Prepare search data (once per document)\n",
" 1. Collect: We'll download a few hundred Wikipedia articles about the 2022 Olympics\n",
" 2. Chunk: Documents are split into short, mostly self-contained sections to be embedded\n",
" 3. Embed: Each section is embedded with the OpenAI API\n",
Expand All @@ -97,7 +97,7 @@
"\n",
"### Costs\n",
"\n",
"Because GPT is more expensive than embeddings search, a system with a high volume of queries will have its costs dominated by step 3.\n",
"Because GPT is more expensive than embeddings search, a system with a decent volume of queries will have its costs dominated by step 3.\n",
"\n",
"- For `gpt-3.5-turbo` using ~1,000 tokens per query, it costs ~$0.002 per query, or ~500 queries per dollar (as of Apr 2023)\n",
"- For `gpt-4`, again assuming ~1,000 tokens per query, it costs ~$0.03 per query, or ~30 queries per dollar (as of Apr 2023)\n",
Expand Down Expand Up @@ -574,7 +574,7 @@
"source": [
"Thanks to the Wikipedia article included in the input message, GPT answers correctly.\n",
"\n",
"In this particular case, GPT was intelligent enough to realize that the original question was underspecified, as there were three curling gold medals, not just one.\n",
"In this particular case, GPT was intelligent enough to realize that the original question was underspecified, as there were three curling gold medal events, not just one.\n",
"\n",
"Of course, this example partly relied on human intelligence. We knew the question was about curling, so we inserted a Wikipedia article on curling.\n",
"\n",
Expand All @@ -591,7 +591,7 @@
"\n",
"To save you the time & expense, we've prepared a pre-embedded dataset of a few hundred Wikipedia articles about the 2022 Winter Olympics.\n",
"\n",
"To see how we constructed this dataset, or to modify it, see [Embedding Wikipedia articles for search](Embedding_Wikipedia_articles_for_search.ipynb)."
"To see how we constructed this dataset, or to modify it yourself, see [Embedding Wikipedia articles for search](Embedding_Wikipedia_articles_for_search.ipynb)."
]
},
{
Expand Down Expand Up @@ -1011,7 +1011,7 @@
"source": [
"Despite `gpt-3.5-turbo` having no knowledge of the 2022 Winter Olympics, our search system was able to retrieve reference text for the model to read, allowing it to correctly list the gold medal winners in the Men's and Women's tournaments.\n",
"\n",
"However, it still wasn't quite perfect - the model failed to list the gold medal winners from the Mixed doubles event."
"However, it still wasn't quite perfectthe model failed to list the gold medal winners from the Mixed doubles event."
]
},
{
Expand Down