diff --git a/components/retriever/multiquery/main.go b/components/retriever/multiquery/main.go index ecec1c3..a010581 100644 --- a/components/retriever/multiquery/main.go +++ b/components/retriever/multiquery/main.go @@ -103,11 +103,9 @@ func main() { func newChatModel(ctx context.Context, baseURL, apiKey, modelName string) (model.ChatModel, error) { return openai.NewChatModel(ctx, &openai.ChatModelConfig{ - BaseURL: baseURL, - APIKey: apiKey, - Model: modelName, - ByAzure: true, - APIVersion: "2024-06-01", + BaseURL: baseURL, + APIKey: apiKey, + Model: modelName, }) } diff --git a/compose/chain/main.go b/compose/chain/main.go index 476688f..7023fa7 100644 --- a/compose/chain/main.go +++ b/compose/chain/main.go @@ -35,7 +35,7 @@ import ( func main() { openAPIBaseURL := os.Getenv("OPENAI_BASE_URL") openAPIAK := os.Getenv("OPENAI_API_KEY") - modelName := os.Getenv("MODEL_NAME") + modelName := os.Getenv("OPENAI_MODEL_NAME") ctx := context.Background() // build branch func @@ -84,17 +84,13 @@ func main() { return "你的叫声是怎样的?", nil })) - modelConf := &openai.ChatModelConfig{ + // create chat model node + cm, err := openai.NewChatModel(context.Background(), &openai.ChatModelConfig{ BaseURL: openAPIBaseURL, APIKey: openAPIAK, Model: modelName, Temperature: gptr.Of(float32(0.7)), - ByAzure: false, - APIVersion: "2024-06-01", - } - - // create chat model node - cm, err := openai.NewChatModel(context.Background(), modelConf) + }) if err != nil { log.Panic(err) return diff --git a/compose/graph/tool_call_agent/tool_call_agent.go b/compose/graph/tool_call_agent/tool_call_agent.go index 529c730..32f03cc 100644 --- a/compose/graph/tool_call_agent/tool_call_agent.go +++ b/compose/graph/tool_call_agent/tool_call_agent.go @@ -34,7 +34,7 @@ import ( func main() { - //openAIBaseURL := os.Getenv("OPENAI_BASE_URL") + openAIBaseURL := os.Getenv("OPENAI_BASE_URL") openAIAPIKey := os.Getenv("OPENAI_API_KEY") modelName := os.Getenv("OPENAI_MODEL_NAME") @@ -50,17 +50,13 @@ func main() { schema.UserMessage("{user_query}"), ) - modelConf := &openai.ChatModelConfig{ - //BaseURL: openAIBaseURL, - APIKey: openAIAPIKey, - //ByAzure: true, + // 2. create an instance of ChatModel as 2nd Graph Node + chatModel, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ + BaseURL: openAIBaseURL, + APIKey: openAIAPIKey, Model: modelName, Temperature: gptr.Of(float32(0.7)), - APIVersion: "2024-06-01", - } - - // 2. create an instance of ChatModel as 2nd Graph Node - chatModel, err := openai.NewChatModel(ctx, modelConf) + }) if err != nil { logs.Errorf("NewChatModel failed, err=%v", err) return diff --git a/compose/graph/tool_call_once/tool_call_once.go b/compose/graph/tool_call_once/tool_call_once.go index c60d0e1..4a4ed2f 100644 --- a/compose/graph/tool_call_once/tool_call_once.go +++ b/compose/graph/tool_call_once/tool_call_once.go @@ -33,7 +33,7 @@ import ( ) func main() { - //openAIBaseURL := os.Getenv("OPENAI_BASE_URL") + openAIBaseURL := os.Getenv("OPENAI_BASE_URL") openAIAPIKey := os.Getenv("OPENAI_API_KEY") modelName := os.Getenv("OPENAI_MODEL_NAME") @@ -46,16 +46,12 @@ func main() { schema.UserMessage("{query}"), ) - modelConf := &openai.ChatModelConfig{ - //BaseURL: openAIBaseURL, - APIKey: openAIAPIKey, - //ByAzure: true, + chatModel, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ + BaseURL: openAIBaseURL, + APIKey: openAIAPIKey, Model: modelName, Temperature: gptr.Of(float32(0.7)), - APIVersion: "2024-06-01", - } - - chatModel, err := openai.NewChatModel(ctx, modelConf) + }) if err != nil { logs.Fatalf("NewChatModel failed, err=%v", err) } diff --git a/compose/graph/two_model_chat/two_model_chat.go b/compose/graph/two_model_chat/two_model_chat.go index 2a9efe0..97451f3 100644 --- a/compose/graph/two_model_chat/two_model_chat.go +++ b/compose/graph/two_model_chat/two_model_chat.go @@ -35,27 +35,23 @@ import ( ) func main() { - //openAIBaseURL := os.Getenv("OPENAI_BASE_URL") + openAIBaseURL := os.Getenv("OPENAI_BASE_URL") openAIAPIKey := os.Getenv("OPENAI_API_KEY") modelName := os.Getenv("OPENAI_MODEL_NAME") ctx := context.Background() - modelConf := &openai.ChatModelConfig{ - //BaseURL: openAIBaseURL, - APIKey: openAIAPIKey, - //ByAzure: true, - Model: modelName, - Temperature: gptr.Of(float32(0.7)), - APIVersion: "2024-06-01", - } - type state struct { currentRound int msgs []*schema.Message } - llm, err := openai.NewChatModel(ctx, modelConf) + llm, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ + BaseURL: openAIBaseURL, + APIKey: openAIAPIKey, + Model: modelName, + Temperature: gptr.Of(float32(0.7)), + }) if err != nil { logs.Fatalf("new chat model failed: %v", err) } diff --git a/flow/agent/multiagent/host/journal/host.go b/flow/agent/multiagent/host/journal/host.go index e124a5d..b3b31d9 100644 --- a/flow/agent/multiagent/host/journal/host.go +++ b/flow/agent/multiagent/host/journal/host.go @@ -27,8 +27,7 @@ func newHost(ctx context.Context, baseURL, apiKey, modelName string) (*host.Host chatModel, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ BaseURL: baseURL, Model: modelName, - //ByAzure: true, - APIKey: apiKey, + APIKey: apiKey, }) if err != nil { return nil, err diff --git a/flow/agent/multiagent/plan_execute/main.go b/flow/agent/multiagent/plan_execute/main.go index 48584cb..c06cae1 100644 --- a/flow/agent/multiagent/plan_execute/main.go +++ b/flow/agent/multiagent/plan_execute/main.go @@ -44,7 +44,7 @@ func main() { ctx := context.Background() deepSeekModel, err := deepseek.NewChatModel(ctx, &deepseek.ChatModelConfig{ - Model: os.Getenv("DEEPSEEK_MODEL_ID"), + Model: os.Getenv("DEEPSEEK_MODEL_NAME"), APIKey: os.Getenv("DEEPSEEK_API_KEY"), BaseURL: os.Getenv("DEEPSEEK_BASE_URL"), }) @@ -54,7 +54,7 @@ func main() { arkModel, err := ark.NewChatModel(ctx, &ark.ChatModelConfig{ APIKey: os.Getenv("ARK_API_KEY"), - Model: os.Getenv("ARK_MODEL_ID"), + Model: os.Getenv("ARK_MODEL_NAME"), }) if err != nil { log.Fatalf("new Ark model failed: %v", err) diff --git a/flow/agent/react/react.go b/flow/agent/react/react.go index f24de10..b73b5fc 100644 --- a/flow/agent/react/react.go +++ b/flow/agent/react/react.go @@ -38,18 +38,16 @@ import ( func main() { openAIAPIKey := os.Getenv("OPENAI_API_KEY") - // openAIBaseURL := os.Getenv("OPENAI_BASE_URL") + openAIBaseURL := os.Getenv("OPENAI_BASE_URL") openAIModelName := os.Getenv("OPENAI_MODEL_NAME") ctx := context.Background() - // prepare chat model + // prepare ChatModel chatModel, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ - // BaseURL: openAIBaseURL, // if using Azure or other models conforming to OpenAI protocol, set this to your model's base URL - APIKey: openAIAPIKey, - Model: openAIModelName, - // ByAzure: true, // if using Azure's OpenAI API, set this to true - APIVersion: "2024-06-01", + BaseURL: openAIBaseURL, + APIKey: openAIAPIKey, + Model: openAIModelName, }) if err != nil { logs.Errorf("failed to create chat model: %v", err) diff --git a/quickstart/chat/openai.go b/quickstart/chat/openai.go index acb044e..03220a9 100644 --- a/quickstart/chat/openai.go +++ b/quickstart/chat/openai.go @@ -27,9 +27,12 @@ import ( func createOpenAIChatModel(ctx context.Context) model.ChatModel { key := os.Getenv("OPENAI_API_KEY") + modelName := os.Getenv("OPENAI_MODEL_NAME") + baseURL := os.Getenv("OPENAI_BASE_URL") chatModel, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ - Model: "gpt-4o", // 使用的模型版本 - APIKey: key, // OpenAI API 密钥 + BaseURL: baseURL, + Model: modelName, + APIKey: key, }) if err != nil { log.Fatalf("create openai chat model failed, err=%v", err) diff --git a/quickstart/todoagent/main.go b/quickstart/todoagent/main.go index 47018f4..7297f18 100644 --- a/quickstart/todoagent/main.go +++ b/quickstart/todoagent/main.go @@ -33,6 +33,8 @@ import ( func main() { openAIAPIKey := os.Getenv("OPENAI_API_KEY") + openAIModelName := os.Getenv("OPENAI_MODEL_NAME") + openAIBaseURL := os.Getenv("OPENAI_BASE_URL") ctx := context.Background() @@ -59,7 +61,8 @@ func main() { // 创建并配置 ChatModel chatModel, err := openai.NewChatModel(ctx, &openai.ChatModelConfig{ - Model: "gpt-4o", + BaseURL: openAIBaseURL, + Model: openAIModelName, APIKey: openAIAPIKey, Temperature: gptr.Of(float32(0.7)), })