depricated model GPT3Ada changed to GPT3Babbage002 (#843)
* depricated model GPT3Ada changed to GPT3Babbage002 * Delete test.mp3
This commit is contained in:
@@ -141,7 +141,7 @@ func main() {
|
|||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
req := openai.CompletionRequest{
|
req := openai.CompletionRequest{
|
||||||
Model: openai.GPT3Ada,
|
Model: openai.GPT3Babbage002,
|
||||||
MaxTokens: 5,
|
MaxTokens: 5,
|
||||||
Prompt: "Lorem ipsum",
|
Prompt: "Lorem ipsum",
|
||||||
}
|
}
|
||||||
@@ -174,7 +174,7 @@ func main() {
|
|||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
req := openai.CompletionRequest{
|
req := openai.CompletionRequest{
|
||||||
Model: openai.GPT3Ada,
|
Model: openai.GPT3Babbage002,
|
||||||
MaxTokens: 5,
|
MaxTokens: 5,
|
||||||
Prompt: "Lorem ipsum",
|
Prompt: "Lorem ipsum",
|
||||||
Stream: true,
|
Stream: true,
|
||||||
|
|||||||
@@ -82,7 +82,7 @@ func ExampleClient_CreateCompletion() {
|
|||||||
resp, err := client.CreateCompletion(
|
resp, err := client.CreateCompletion(
|
||||||
context.Background(),
|
context.Background(),
|
||||||
openai.CompletionRequest{
|
openai.CompletionRequest{
|
||||||
Model: openai.GPT3Ada,
|
Model: openai.GPT3Babbage002,
|
||||||
MaxTokens: 5,
|
MaxTokens: 5,
|
||||||
Prompt: "Lorem ipsum",
|
Prompt: "Lorem ipsum",
|
||||||
},
|
},
|
||||||
@@ -99,7 +99,7 @@ func ExampleClient_CreateCompletionStream() {
|
|||||||
stream, err := client.CreateCompletionStream(
|
stream, err := client.CreateCompletionStream(
|
||||||
context.Background(),
|
context.Background(),
|
||||||
openai.CompletionRequest{
|
openai.CompletionRequest{
|
||||||
Model: openai.GPT3Ada,
|
Model: openai.GPT3Babbage002,
|
||||||
MaxTokens: 5,
|
MaxTokens: 5,
|
||||||
Prompt: "Lorem ipsum",
|
Prompt: "Lorem ipsum",
|
||||||
Stream: true,
|
Stream: true,
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ func main() {
|
|||||||
resp, err := client.CreateCompletion(
|
resp, err := client.CreateCompletion(
|
||||||
context.Background(),
|
context.Background(),
|
||||||
openai.CompletionRequest{
|
openai.CompletionRequest{
|
||||||
Model: openai.GPT3Ada,
|
Model: openai.GPT3Babbage002,
|
||||||
MaxTokens: 5,
|
MaxTokens: 5,
|
||||||
Prompt: "Lorem ipsum",
|
Prompt: "Lorem ipsum",
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -169,7 +169,7 @@ func TestCreateCompletionStreamRateLimitError(t *testing.T) {
|
|||||||
var apiErr *openai.APIError
|
var apiErr *openai.APIError
|
||||||
_, err := client.CreateCompletionStream(context.Background(), openai.CompletionRequest{
|
_, err := client.CreateCompletionStream(context.Background(), openai.CompletionRequest{
|
||||||
MaxTokens: 5,
|
MaxTokens: 5,
|
||||||
Model: openai.GPT3Ada,
|
Model: openai.GPT3Babbage002,
|
||||||
Prompt: "Hello!",
|
Prompt: "Hello!",
|
||||||
Stream: true,
|
Stream: true,
|
||||||
})
|
})
|
||||||
|
|||||||
Reference in New Issue
Block a user