From f121f0f34be7610fa3463e9f9a670fbbaf5268be Mon Sep 17 00:00:00 2001 From: fatwang2 Date: Wed, 21 Feb 2024 22:36:44 +0800 Subject: [PATCH] =?UTF-8?q?=E6=96=B0=E5=A2=9E=E6=94=AF=E6=8C=81gemini?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .DS_Store | Bin 6148 -> 8196 bytes README.md | 17 +- gemini.js | 287 +++++++++++++++++++++++++++++++++ cloudflare-worker => openai.js | 0 4 files changed, 299 insertions(+), 5 deletions(-) create mode 100644 gemini.js rename cloudflare-worker => openai.js (100%) diff --git a/.DS_Store b/.DS_Store index c7e1fff5f731ea9351e136cc84230ae332c73485..58d08315eda215309431e189ee5806cc7d214409 100644 GIT binary patch delta 271 zcmZoMXmOBWU|?W$DortDU;r^WfEYvza8E20o2aMAD6}zPH}hr%jz7$c**Q2SHn1=X zZRTM)$~f7AMR2kTn=BiHBhUb?$?j~%lf&2rggF`V844Iu8S)qs88R947_u0OC(mV- zpWMY}#M%usc+KR4K$Q*bER1U=|76orVFU^4{Raaei-CcMA&9{hXs8Q=FOb#+V!z3b z?2P>ECHx*rjq?4PWiXjPTdE(^xY>Mnqna%P%@0qzIxPkgy fLEhRd$nl+dGQWuDWPctG4n~NN88*lB%wYxq?4U&O delta 250 zcmZp1XfcprU|?W$DortDU=RQ@Ie-{Mvv5r;6q~50$SANeU^g?Pz-AtSql_+0SAkM` z|G@yrVqg$tNM^`k$Y&^JNMT3=vJx4J7*ZK@8OniTMGV!pwu8v{sMZ*rip?qmxgfysTs za*XpPuN5|Bp9gWp#6q{l>>M0|%s^#8AixbITtUv?SoocJGQW%`$TJK~5RZZU%CI?} HXAUy}DhN3` diff --git a/README.md b/README.md index 3962988..13bac08 100644 --- a/README.md +++ b/README.md @@ -8,6 +8,7 @@ Buy Me A Coffee # 版本更新 +- V0.1.6,20240221,支持Gemini模型,暂时可通过cloudflare worker的方式配置 - V0.1.5,20240205,支持新闻搜索,快速浏览新闻更方便 - V0.1.4,20240120,支持Zeabur一键部署,非常方便,强烈推荐! - V0.1.3,20240120,支持本地部署,可自行部署在自己的服务器上 @@ -17,7 +18,7 @@ 更多历史更新,请参见 [版本记录](https://github.com/fatwang2/search2ai/releases) # 产品介绍 -- search2ai,让你的 OpenAI API 支持联网,搜索、新闻、网页总结,大模型会根据你的输入判断是否联网,不是每次都联网搜索,不需要安装任何插件,也不需要更换key,直接在你常用的 OpenAI 三方客户端替换自定义地址即可,也支持自行部署,不会影响使用 OpenAI 的其他功能,如画图、语音等 +- search2ai,让你的 LLM API 支持联网,搜索、新闻、网页总结,已支持OpenAI、Gemini,大模型会根据你的输入判断是否联网,不是每次都联网搜索,不需要安装任何插件,也不需要更换key,直接在你常用的 OpenAI/Gemini 三方客户端替换自定义地址即可,也支持自行部署,不会影响使用 OpenAI 的其他功能,如画图、语音等,Gemini的画图功能适配中 @@ -32,9 +33,15 @@ # 使用方法 **直接使用:替换客户端自定义域名为下面地址即可** +OpenAI ``` https://api.search2ai.online ``` +Gemini +``` +https://geminiapi.search2ai.online +``` + 如图所示
@@ -43,7 +50,7 @@ https://api.search2ai.online
-demo站体验地址:[demo站](https://search2ai.online/demo),根据提示使用你的key,更换自定义域名即可 +demo站体验地址:[OpenAI](https://search2ai.online/demo),根据提示使用你的key,更换自定义域名即可 **Zeabur一键部署(强烈推荐)** @@ -91,9 +98,9 @@ http://localhost:3014/v1/chat/completions ``` **cloudflare worker部署** -1. 复制[search2ai](https://search2ai.online/cloudflare)的代码,不需要任何修改!在cloudflare的worker里部署,上线后的worker的地址可作为你接口调用时的自定义域名地址,注意拼接,worker地址仅代表v1前的部分 XXX/v1/chat/completions +1. 复制[openai.js](https://search2ai.online/cloudflare)或者[gemini.js](gemini.js)或者的代码,不需要任何修改!在cloudflare的worker里部署,上线后的worker的地址可作为你接口调用时的自定义域名地址,注意拼接,worker地址仅代表v1前的部分 -2. 在worker中配置变量 +2. 在worker中配置变量(部署openai时需要配置,gemini暂不支持) ![效果示例](pictures/worker.png) - SEARCH_SERVICE:暂时支持google、bing、serpapi、serper、duckduckgo(推荐),必填 - APIBASE:如果你在用 OpenAI 三方代理,可在这里填入,注意不需要加v1,非必填 @@ -118,7 +125,7 @@ http://localhost:3014/v1/chat/completions 为保证更新,也可以先fork本项目后自己在vercel上部署 # 后续迭代 -- 支持Gemini +- Gemini支持流式输出、兼容非聊天场景 - 支持Azure OpenAI - 修复Vercel项目流式输出问题 - 提升流式输出的速度 diff --git a/gemini.js b/gemini.js new file mode 100644 index 0000000..cb521d5 --- /dev/null +++ b/gemini.js @@ -0,0 +1,287 @@ +addEventListener('fetch', event => { + event.respondWith(handleRequest(event.request)) +}) +const corsHeaders = { + 'Access-Control-Allow-Origin': '*', + 'Access-Control-Allow-Methods': 'GET, POST, OPTIONS', // 允许的HTTP方法 + 'Access-Control-Allow-Headers': 'DNT,User-Agent,X-Requested-With,If-Modified-Since,Cache-Control,Content-Type,Range,Authorization,x-goog-api-client,x-goog-api-key', + 'Access-Control-Max-Age': '86400', // 预检请求结果的缓存时间 +}; +async function search(query) { + console.log('search function started with query:', query); + + try { + const url = "https://search.search2ai.one"; + const headers = {"Content-Type": "application/json"}; + const body = { + "query": query, + "search_service": "duckduckgo" + }; + + const response = await fetch(url, { + method: 'POST', + headers: headers, + body: JSON.stringify(body) + }); + + if (response.ok) { + console.log('search function successfully completed'); + return response.json(); + } else { + throw new Error("Unable to make request"); + } + } catch (error) { + return {"status": "ERROR: " + error.message}; + } + } + +async function parse_function_response(message) { + if (!message[0] || !message[0]["functionCall"]) { + console.log('Invalid message:', message); + return { function_name: 'ERROR', function_response: 'Invalid message' }; + } + const function_call = message[0]["functionCall"]; + const function_name = function_call["name"]; + + console.log("Gemini: Called function " + function_name ); + + let function_response; + try { + const arguments = function_call["args"]; + + if (function_name === 'search') { + // 检查 args 参数是否包含 query 属性 + if (!arguments.hasOwnProperty('query')) { + function_response = "ERROR: Missing query parameter"; + console.log('Missing query parameter'); + return { function_name, function_response }; + } + + // 获取 query 参数的值 + const query = arguments.query; + + // 调用 search 函数并获取结果 + function_response = await search(query); + return { function_name, function_response }; // 直接返回 + } else { + function_response = "ERROR: Called unknown function"; + console.log('Called unknown function:', function_name); + } + } catch (error) { + function_response = "ERROR: Invalid arguments"; + console.log('Invalid arguments:', error.message); + } + console.log('Function response:', function_response); + + return {function_name, function_response}; +} + + +async function run_conversation(api_key, message) { + const date = new Date(); + const timeZone = 'Asia/Shanghai'; + const formatter = new Intl.DateTimeFormat('en-US', { dateStyle: 'full', timeZone }); + const currentDate = formatter.format(date); + if (!message) { + console.log('Invalid message:', message); + return { error: 'Invalid message' }; + } + const customMessage = [ + { + "role":"user", + "parts":[ + { + "text": `Today is ${currentDate}.You are a friendly intelligent assistant with the ability to search online, hopefully you will go online when the user asks for something that requires internet access, otherwise just answer, try to be as simple and clear as possible when answering the user's question, and you can use emoji to make your conversations more interesting!` + } + ] + }, + { + "role": "model", + "parts":[ + { + "text": "okay" + } + ] + }, + ]; + message = [...customMessage, ...message]; + console.log('Running conversation with message:', message); + const originalMessage = [...message]; + let functionResponseJson; + + const definitions = [ + { + "name": "search", + "description": "search on the Interent when the users want something new to know", + "parameters": { + "type": "object", + "properties": { + "query": { + "type": "string", + "description": "The query to search" + } + } + } + } + ]; + + const data = { + "contents": message, + "tools": [{ + "functionDeclarations": definitions + }] + }; + + const response = await fetch("https://generativelanguage.googleapis.com/v1beta/models/gemini-pro:generateContent?key="+api_key, { + method: 'POST', + body: JSON.stringify(data) + }); + + if (!response.ok) { + console.log('Received error response from run_conversation'); + // 修改: 返回表示出错的 Response 对象 + return new Response(JSON.stringify({ error: "Error fetching from Google Language API" }), { + headers: { ...corsHeaders,'content-type': 'application/json' }, + status: 500 // 代表出现 Internal Server Error 的错误码 + }); + } + console.log('Received successful response from run_conversation'); + + let responseJson = await response.json(); + + if (!responseJson["candidates"][0]["content"]) { + console.log("ERROR: No content in response"); + console.log(responseJson); + return; + } + + message = responseJson["candidates"][0]["content"]["parts"]; + + if (message[0]["functionCall"]) { + const {function_name, function_response} = await parse_function_response(message); + + const functionResponseData = { + "contents": [ + + ...originalMessage + , + { + "role": "model", + "parts": [ + ...message,] + },{ + "role": "function", + "parts": [{ + "functionResponse": { + "name": function_name, + "response": { + "name": function_name, + "content": function_response + } + } + }] + }], + "tools": [{ + "functionDeclarations": definitions + }] + }; + console.log('functionResponseData:', functionResponseData); + const functionResponse = await fetch("https://generativelanguage.googleapis.com/v1beta/models/gemini-pro:generateContent?key="+api_key, { + method: 'POST', + body: JSON.stringify(functionResponseData) + }); + + if (!functionResponse.ok) { + console.log('Received error response from run_conversation'); + return; + } + + functionResponseJson = await functionResponse.json(); + + if (!functionResponseJson["candidates"][0]["content"]) { + console.log("ERROR: No content in response"); + console.log(functionResponseJson); + return new Response(JSON.stringify({ error: "No content received from Google Language API"}), { + headers: { ...corsHeaders,'content-type': 'application/json' }, + status: 400 // 代表 Bad Request 的错误码 + }); + } + } else { + functionResponseJson = responseJson; + } + // 将响应封装成一个 Response 对象,然后返回 + return new Response(JSON.stringify(functionResponseJson), { + headers: { ...corsHeaders,'content-type': 'application/json' }, + status: 200, + }); +} + +// HTTP请求处理主函数 +async function handleRequest(request) { + console.log('[handleRequest] Request received', { method: request.method, url: request.url }); + + // 创建一个新的响应对象,并设置 CORS 头部 + if (request.method === 'OPTIONS') { + console.log('[handleRequest] Preparing CORS preflight response.'); + const response = new Response(null, { + status: 204, // OPTIONS 请求通常返回 204 No Content + headers: corsHeaders + }); + + // 输出响应头部 + console.log('[handleRequest] CORS preflight response headers:', JSON.stringify([...response.headers])); + return response; + } + + // 解析请求 URL 的路径部分 + const url = new URL(request.url); + const path = url.pathname; + + if (path.includes('/v1/models/gemini-pro')) { + console.log('[handleRequest] Handling gemini-pro request.'); + + // 提取 API 键和请求内容 + const api_key = request.headers.get('x-goog-api-key'); + let message; + try { + const requestBody = await request.text(); // 获取请求文本 + console.log('[handleRequest] Request body:', requestBody); + message = JSON.parse(requestBody).contents; // 解析 JSON 内容 + } catch (error) { + console.error('[handleRequest] Error parsing request body:', error.message); + return new Response(JSON.stringify({ error: 'Bad JSON in request' }), { + headers: { ...corsHeaders,'content-type': 'application/json' }, + status: 400, + }); + } + + try { + // 调用 run_conversation 函数并获取响应 + const response = await run_conversation(api_key, message); + + // 检查响应类型并处理 + if (response instanceof Response) { + console.log('[handleRequest] run_conversation provided a response object.'); + return response; + } else { + console.error('[handleRequest] run_conversation returned an unexpected response type.'); + throw new Error('Invalid response type from run_conversation'); + } + } catch (error) { + // 捕获错误并返回错误响应 + console.error('[handleRequest] Error during request handling:', error.message); + return new Response(JSON.stringify({ error: error.message }), { + headers: { ...corsHeaders,'content-type': 'application/json' }, + status: 500, + }); + } + } else { + // 处理不符合特定路径的其他请求 + console.log('[handleRequest] Request not found for path:', path); + return new Response(JSON.stringify({ error: 'Not found' }), { + headers: { ...corsHeaders, + 'content-type': 'application/json' }, + status: 404, + }); + } +} diff --git a/cloudflare-worker b/openai.js similarity index 100% rename from cloudflare-worker rename to openai.js