From 28c59c39103bc7b4574b263a869a371bdd47eb3e Mon Sep 17 00:00:00 2001
From: 0007 <0007@qq.com>
Date: Wed, 27 Aug 2025 19:57:35 +0800
Subject: [PATCH] Add File
---
.../java/com/agentsflex/llm/qwen/QwenLlm.java | 132 ++++++++++++++++++
1 file changed, 132 insertions(+)
create mode 100644 agents-flex-llm/agents-flex-llm-qwen/src/main/java/com/agentsflex/llm/qwen/QwenLlm.java
diff --git a/agents-flex-llm/agents-flex-llm-qwen/src/main/java/com/agentsflex/llm/qwen/QwenLlm.java b/agents-flex-llm/agents-flex-llm-qwen/src/main/java/com/agentsflex/llm/qwen/QwenLlm.java
new file mode 100644
index 0000000..0c8207e
--- /dev/null
+++ b/agents-flex-llm/agents-flex-llm-qwen/src/main/java/com/agentsflex/llm/qwen/QwenLlm.java
@@ -0,0 +1,132 @@
+/*
+ * Copyright (c) 2023-2025, Agents-Flex (fuhai999@gmail.com).
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package com.agentsflex.llm.qwen;
+
+import com.agentsflex.core.document.Document;
+import com.agentsflex.core.llm.BaseLlm;
+import com.agentsflex.core.llm.ChatOptions;
+import com.agentsflex.core.llm.StreamResponseListener;
+import com.agentsflex.core.llm.client.BaseLlmClientListener;
+import com.agentsflex.core.llm.client.HttpClient;
+import com.agentsflex.core.llm.client.LlmClient;
+import com.agentsflex.core.llm.client.LlmClientListener;
+import com.agentsflex.core.llm.client.impl.SseClient;
+import com.agentsflex.core.llm.embedding.EmbeddingOptions;
+import com.agentsflex.core.llm.response.AiMessageResponse;
+import com.agentsflex.core.parser.AiMessageParser;
+import com.agentsflex.core.prompt.Prompt;
+import com.agentsflex.core.store.VectorData;
+import com.agentsflex.core.util.LogUtil;
+import com.agentsflex.core.util.StringUtil;
+import com.alibaba.fastjson.JSON;
+import com.alibaba.fastjson.JSONObject;
+import com.alibaba.fastjson.JSONPath;
+
+import java.util.HashMap;
+import java.util.Map;
+
+public class QwenLlm extends BaseLlm {
+
+ private final HttpClient httpClient = new HttpClient();
+
+ public AiMessageParser aiMessageParser = QwenLlmUtil.getAiMessageParser(false);
+ public AiMessageParser streamMessageParser = QwenLlmUtil.getAiMessageParser(true);
+
+
+ public QwenLlm(QwenLlmConfig config) {
+ super(config);
+ }
+
+
+ @Override
+ public AiMessageResponse chat(Prompt prompt, ChatOptions options) {
+ Map headers = new HashMap<>();
+ headers.put("Content-Type", "application/json");
+ headers.put("Authorization", "Bearer " + getConfig().getApiKey());
+
+ //非流式返回,需要设置为 false
+ options.setEnableThinking(false);
+
+ String payload = QwenLlmUtil.promptToPayload(prompt, config, options, false);
+ String endpoint = config.getEndpoint();
+ String response = httpClient.post(endpoint + "/compatible-mode/v1/chat/completions", headers, payload);
+
+ if (config.isDebug()) {
+ LogUtil.println(">>>>receive payload:" + response);
+ }
+
+ if (StringUtil.noText(response)) {
+ return AiMessageResponse.error(prompt, response, "no content for response.");
+ }
+
+ JSONObject jsonObject = JSON.parseObject(response);
+ JSONObject error = jsonObject.getJSONObject("error");
+
+ AiMessageResponse messageResponse = new AiMessageResponse(prompt, response, aiMessageParser.parse(jsonObject));
+
+ if (error != null && !error.isEmpty()) {
+ messageResponse.setError(true);
+ messageResponse.setErrorMessage(error.getString("message"));
+ messageResponse.setErrorType(error.getString("type"));
+ messageResponse.setErrorCode(error.getString("code"));
+ }
+
+ return messageResponse;
+ }
+
+
+ @Override
+ public void chatStream(Prompt prompt, StreamResponseListener listener, ChatOptions options) {
+ LlmClient llmClient = new SseClient();
+ Map headers = new HashMap<>();
+ headers.put("Content-Type", "application/json");
+ headers.put("Authorization", "Bearer " + getConfig().getApiKey());
+ headers.put("X-DashScope-SSE", "enable"); //stream
+
+ String payload = QwenLlmUtil.promptToPayload(prompt, config, options, true);
+ LlmClientListener clientListener = new BaseLlmClientListener(this, llmClient, listener, prompt, streamMessageParser);
+
+ String endpoint = config.getEndpoint();
+ llmClient.start(endpoint + "/compatible-mode/v1/chat/completions", headers, payload, clientListener, config);
+ }
+
+
+ @Override
+ public VectorData embed(Document document, EmbeddingOptions options) {
+ String payload = QwenLlmUtil.promptToEnabledPayload(document, options, config);
+ Map headers = new HashMap<>();
+ headers.put("Content-Type", "application/json");
+ headers.put("Authorization", "Bearer " + getConfig().getApiKey());
+
+ String url = config.getEndpoint() + "/compatible-mode/v1/embeddings";
+ String response = httpClient.post(url, headers, payload);
+
+ if (config.isDebug()) {
+ LogUtil.println(">>>>receive payload:" + response);
+ }
+
+ if (StringUtil.noText(response)) {
+ return null;
+ }
+
+ VectorData vectorData = new VectorData();
+ double[] embedding = JSONPath.read(response, "$.data[0].embedding", double[].class);
+ vectorData.setVector(embedding);
+
+ return vectorData;
+ }
+
+}