当前位置: 首页 > news >正文

游戏开发和网站开发那个有前景重庆烤鱼制作

游戏开发和网站开发那个有前景,重庆烤鱼制作,id和wordpress,网站升级改版需要多久​一、软件介绍 文末提供程序和源码下载 node-DeepResearch开源ai程序用于深入调查查询,继续搜索、阅读网页、推理,直到找到答案。 重要提示 与 OpenAI/Gemini/Perfasciity 的“深度研究”不同,我们只专注于通过迭代过程找到正确的答案 。我…

​一、软件介绍

文末提供程序和源码下载

       node-DeepResearch开源ai程序用于深入调查查询,继续搜索、阅读网页、推理,直到找到答案。

重要提示

     与 OpenAI/Gemini/Perfasciity 的“深度研究”不同,我们只专注于通过迭代过程找到正确的答案 。我们不会针对长篇文章进行优化,这是一个完全不同的问题 ——所以如果你需要从深度搜索中获得快速、简洁的答案,你来对地方了。如果您正在寻找像 OpenAI/Gemini/Perplexity 那样的 AI 生成的长报告,那么这不适合您。

二、Try it Yourself 亲自试一试

We host an online deployment of this exact codebase, which allows you to do a vibe-check; or use it as daily productivity tools.
我们托管了这个确切代码库的在线部署,它允许您进行氛围检查;或将其用作日常生产力工具。

https://search.jina.ai

The official API is also available for you to use:
官方 API 也可供您使用:

https://deepsearch.jina.ai/v1/chat/completions
Learn more about the API at https://jina.ai/deepsearch
在 https://jina.ai/deepsearch 了解有关 API 的更多信息

三、Install 安装

代码文末可直接下载更快速

cd node-DeepResearch
npm install

四、Usage 用法

We use Gemini (latest gemini-2.0-flash) / OpenAI / LocalLLM for reasoning, Jina Reader for searching and reading webpages, you can get a free API key with 1M tokens from jina.ai.
我们使用 Gemini(最新 gemini-2.0-flash )/ OpenAI / LocalLLM 进行推理,使用 Jina Reader 搜索和阅读网页,您可以从 jina.ai 获得带有 1M 令牌的免费 API 密钥。

export GEMINI_API_KEY=...  # for gemini
# export OPENAI_API_KEY=... # for openai
# export LLM_PROVIDER=openai # for openai
export JINA_API_KEY=jina_...  # free jina api key, get from https://jina.ai/reader

npm run dev $QUERY
 

您还可以使用我们的官方 DeepSearch API:

https://deepsearch.jina.ai/v1/chat/completions
You can use it with any OpenAI-compatible client.
您可以将它与任何 OpenAI 兼容的客户端一起使用。

For the authentication Bearer, API key, rate limit, get from https://jina.ai/deepsearch.
对于身份验证 Bearer、API 密钥、速率限制,从 https://jina.ai/deepsearch 获取。

Client integration guidelines
客户端集成准则
If you are building a web/local/mobile client that uses Jina DeepSearch API, here are some design guidelines:
如果您正在构建一个使用 Jina DeepSearch API 的 web/本地/移动客户端,以下是一些设计准则:

Our API is fully compatible with OpenAI API schema, this should greatly simplify the integration process. The model name is jina-deepsearch-v1.
我们的 API 与 OpenAI API 架构完全兼容,这应该会大大简化集成过程。型号名称为 jina-deepsearch-v1 .
Our DeepSearch API is a reasoning+search grounding LLM, so it's best for questions that require deep reasoning and search.
我们的 DeepSearch API 是 reasoning+search grounding LLM,因此最适合需要深度推理和搜索的问题。
Two special tokens are introduced <think>...</think>. Please render them with care.
引入了 <think>...</think> 两个特殊令牌。请小心渲染它们。
Citations are often provided, and in Github-flavored markdown footnote format, e.g. [^1], [^2], ...
通常会提供引文,并且采用 Github 风格的 markdown 脚注格式,例如 [^1] , [^2] , ...
Guide the user to get a Jina API key from https://jina.ai, with 1M free tokens for new API key.
引导用户从 https://jina.ai 获取 Jina API 密钥,新 API 密钥有 1M 免费令牌。
There are rate limits, between 10RPM to 30RPM depending on the API key tier.
存在 10RPM 到 30RPM 之间的速率限制,具体取决于 API 密钥层。
Download Jina AI logo here
在此处下载 Jeena Ae 徽标
Demo 演示
was recorded with gemini-1.5-flash, the latest gemini-2.0-flash leads to much better results!
与 一起录制 gemini-1.5-flash ,最新的 gemini-2.0-flash 结果要好得多!

Query: "what is the latest blog post's title from jina ai?" 3 steps; answer is correct!
查询: "what is the latest blog post's title from jina ai?" 3 个步骤;答案是正确的!

Query: "what is the context length of readerlm-v2?" 2 steps; answer is correct!
查询: "what is the context length of readerlm-v2?" 2 个步骤;答案是正确的!

Query: "list all employees from jina ai that u can find, as many as possible" 11 steps; partially correct! but im not in the list :(
查询: "list all employees from jina ai that u can find, as many as possible" 11 个步骤;部分正确!但我不在列表中:(

Query: "who will be the biggest competitor of Jina AI" 42 steps; future prediction kind, so it's arguably correct! atm Im not seeing weaviate as a competitor, but im open for the future "i told you so" moment.
查询: "who will be the biggest competitor of Jina AI" 42 个步骤;未来预测类型,所以可以说是正确的!ATM 我不认为 weaviate 自己是竞争对手,但我对未来“我告诉过你”的时刻持开放态度。

More examples: 更多示例:

# example: no tool calling 
npm run dev "1+1="
npm run dev "what is the capital of France?"

# example: 2-step
npm run dev "what is the latest news from Jina AI?"

# example: 3-step
npm run dev "what is the twitter account of jina ai's founder"

# example: 13-step, ambiguious question (no def of "big")
npm run dev "who is bigger? cohere, jina ai, voyage?"

# example: open question, research-like, long chain of thoughts
npm run dev "who will be president of US in 2028?"
npm run dev "what should be jina ai strategy for 2025?"
Use Local LLM 使用本地LLM
Note, not every LLM works with our reasoning flow, we need those who support structured output (sometimes called JSON Schema output, object output) well. Feel free to purpose a PR to add more open-source LLMs to the working list.
请注意,并非每个LLM都适用于我们的推理流程,我们需要那些支持结构化输出(有时称为 JSON Schema 输出、对象输出)的人。请随意使用 PR 将更多开源LLMs添加到工作列表中。

If you use Ollama or LMStudio, you can redirect the reasoning request to your local LLM by setting the following environment variables:
如果您使用 Ollama 或 LMStudio,则可以通过设置以下环境变量将推理请求重定向到您的本地LLM:

export LLM_PROVIDER=openai  # yes, that's right - for local llm we still use openai client
export OPENAI_BASE_URL=http://127.0.0.1:1234/v1  # your local llm endpoint
export OPENAI_API_KEY=whatever  # random string would do, as we don't use it (unless your local LLM has authentication)
export DEFAULT_MODEL_NAME=qwen2.5-7b  # your local llm model name
OpenAI-Compatible Server API
OpenAI 兼容的服务器 API
If you have a GUI client that supports OpenAI API (e.g. CherryStudio, Chatbox) , you can simply config it to use this server.
如果您有支持 OpenAI API 的 GUI 客户端(例如 CherryStudio、Chatbox),您只需将其配置为使用此服务器即可。

Start the server: 启动服务器:

# Without authentication
npm run serve

# With authentication (clients must provide this secret as Bearer token)
npm run serve --secret=your_secret_token
The server will start on http://localhost:3000 with the following endpoint:
服务器将在 http://localhost:3000 上使用以下端点启动:

POST /v1/chat/completions
# Without authentication
curl http://localhost:3000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "jina-deepsearch-v1",
    "messages": [
      {
        "role": "user",
        "content": "Hello!"
      }
    ]
  }'

# With authentication (when server is started with --secret)
curl http://localhost:3000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer your_secret_token" \
  -d '{
    "model": "jina-deepsearch-v1",
    "messages": [
      {
        "role": "user",
        "content": "Hello!"
      }
    ],
    "stream": true
  }'
Response format: 响应格式:

{
  "id": "chatcmpl-123",
  "object": "chat.completion",
  "created": 1677652288,
  "model": "jina-deepsearch-v1",
  "system_fingerprint": "fp_44709d6fcb",
  "choices": [{
    "index": 0,
    "message": {
      "role": "assistant",
      "content": "YOUR FINAL ANSWER"
    },
    "logprobs": null,
    "finish_reason": "stop"
  }],
  "usage": {
    "prompt_tokens": 9,
    "completion_tokens": 12,
    "total_tokens": 21
  }
}
For streaming responses (stream: true), the server sends chunks in this format:
对于流式响应 (stream: true),服务器按以下格式发送块:

{
  "id": "chatcmpl-123",
  "object": "chat.completion.chunk",
  "created": 1694268190,
  "model": "jina-deepsearch-v1",
  "system_fingerprint": "fp_44709d6fcb",
  "choices": [{
    "index": 0,
    "delta": {
      "content": "..."
    },
    "logprobs": null,
    "finish_reason": null
  }]
}
Note: The think content in streaming responses is wrapped in XML tags:
注意:流式响应中的 think content 包装在 XML 标记中:

<think>
[thinking steps...]
</think>
[final answer]
Docker Setup Docker 设置
Build Docker Image 构建 Docker 镜像
To build the Docker image for the application, run the following command:
要为应用程序构建 Docker 镜像,请运行以下命令:

docker build -t deepresearch:latest .
Run Docker Container 运行 Docker 容器
To run the Docker container, use the following command:
要运行 Docker 容器,请使用以下命令:

docker run -p 3000:3000 --env GEMINI_API_KEY=your_gemini_api_key --env JINA_API_KEY=your_jina_api_key deepresearch:latest
Docker Compose
You can also use Docker Compose to manage multi-container applications. To start the application with Docker Compose, run:
您还可以使用 Docker Compose 来管理多容器应用程序。要使用 Docker Compose 启动应用程序,请运行:

docker-compose up


五、How Does it Work? 它是如何工作的?

Not sure a flowchart helps, but here it is:
不确定流程图是否有帮助,但这里是:

六、软件下载

迅雷网盘

本文信息来源于GitHub作者地址:GitHub - jina-ai/node-DeepResearch: Keep searching, reading webpages, reasoning until it finds the answer (or exceeding the token budget)

http://www.dtcms.com/a/433157.html

相关文章:

  • 广州旅游网站建设设计公司广开网络教学平台
  • 有哪些做的比较精美的网站wordpress还原数据库
  • 网站备案转服务器国外做旅游攻略的网站
  • 如何选择大良网站建设网站备案麻烦
  • 节能环保公司网站建设成都网站快速排名优化
  • 绥化网站建设公司edu域名网站
  • 北京外贸网站设计备案网易严选的网站建设
  • 网站建设与推广的实训报告c2c电子商务网站建设
  • 网站培训费用网站开发中所需要的绘图工具
  • 购物网站建设包括哪些链天网站建设
  • 宁波建网站找哪家北京网站优化推广分析
  • 济源网站维护上海做网站较好的公司
  • 专业公司网站建设滁州网站公司
  • 枣庄建设工程管理局网站新闻列表做的最好的网站
  • 小规模公司怎么注销上海网站优化公司排名
  • 外贸网站如何做的好天蒙旅游区网络营销推广方法
  • 桓台建设局网站石家庄有哪些互联网公司
  • 张家港网站 设计制作北京二级建造师查询系统
  • 承德网站建设有限公司做电影海报在哪个网站好
  • 中国城乡住房建设部网站网站百度统计表格怎么做
  • 景区宣传网站制作模板天元建设集团有限公司六公司
  • 微信网站改版价格网站主页排版
  • 技术好的手机网站建设服务器做网站用什么系统
  • 网站建设找谁无锡网站制作工具
  • 免费用搭建网站购买网站平台如何做分录
  • 千秋网络是家西安做网站的公司成都软件公司前100强
  • 印刷报价网站源码下载网站建设项目规划书目录
  • 全国新增确诊病例广州百度首页优化
  • 鄞州区住房和城乡建设局网站河北建设厅网站上不去
  • 绵阳低价网站建设自动化毕设题目网站开发