自己招一个ai员工-Ubuntu22.04安装Openclaw详细教程-小白可直接上手-持续更新中

自己招一个ai员工-Ubuntu22.04安装Openclaw详细教程-小白可直接上手-持续更新中

Ubuntu22.04安装Openclaw

虽然gnuradio我已经弄完了,项目已经做完。但是还需要完整移植一下,并且要修改射频前端,提高带宽,提高频率切换速度。安装openclaw,其实是想让ai帮我实现gnuradio移植、FPGA程序编写、射频以及ARM里面的linux编程。
一句话验证ai是否能帮我提高效率。

准备工作

还是三哥说的,做什么事情都要找到官网:openclaw的中文官网网址如下:
https://clawd.org.cn/
我们就用中文的,也不翻墙。也不搞什么花里胡哨的东西。

在这里插入图片描述

一键安装

curl -fsSL https://clawd.org.cn/install.sh | sudo bash

在这里插入图片描述


在这里插入图片描述


输入Y

在这里插入图片描述

## 这里面需要deepseek的apikey
我们到https://platform.deepseek.com/sign_in
注册,并获取apikey

在这里插入图片描述


在这里插入图片描述
回到安装终端中,输入key

在这里插入图片描述

设置通道 配置飞书

需要创建飞书机器人。
https://open.feishu.cn,进入开发者后台,这里面需要注册用户,并创建ai机器人。

在这里插入图片描述


创建好后,就能看到app id 和app secret了。

在这里插入图片描述


之后在安装终端中填入app id和secret
在后续也可以单独输入命令
openclaw-cn configure --section channels

在这里插入图片描述
在这里插入图片描述


在这里插入图片描述
在这里插入图片描述


在这里插入图片描述


在这里插入图片描述


如果有问题,运行openclaw-cn onboard --install-daemon,在重新配置一下。
以上gateway没有安装,所以打不开网页端。
执行sudo apt install net-tools 安装必备的工具,之后就可以打开了。

在这里插入图片描述


好了,enjoy吧!

让ai员工更好用

加入免费的模型

配置 GLM-4 .7-Flash官方免费API
https://bigmodel.cn/ 注册个开发者,之后获取apikey

在这里插入图片描述


进入控制台,新建个apikey,之后复制。
之后在openclaw.json文件中写入:
“models”: {
“providers”: {
“glm”: {
“baseUrl”: “https://open.bigmodel.cn/api/paas/v4”,
“apiKey”: “你的apiKey”,
“api”: “openai-completions”,
“models”: [
{
“id”: “glm-4.7-flash”,
“name”: “GLM-4.7 Flash”,
“contextWindow”: 128000,
“maxTokens”: 4096,
“reasoning”: false,
“input”: [
“text”
],
“cost”: {
“input”: 0,
“output”: 0,
“cacheRead”: 0,
“cacheWrite”: 0
}
}
]
}
}
},
“agents”: {
“defaults”: {
“model”: {
“primary”: “glm/glm-4.7-flash”
},
“maxConcurrent”: 4,
“subagents”: {
“maxConcurrent”: 8
}
}
},

可以重新启动openclaw,执行openclaw-cn gateway restart

在这里插入图片描述

配置钉钉

https://open-dev.dingtalk.com 在钉钉开发者平台上拷贝apikey 和secret。
Client ID(即 AppKey)
Client Secret(即 AppSecret)
在“权限管理”中,搜索并添加权限:
Card.Instance.Write(卡片实例写权限)
Card.Streaming.Write(卡片流式写权限)
im:message(消息相关权限,根据需要)
之后安装dingtalk插件
OpenClaw 默认没有内置钉钉插件,需要手动安装社区版本:
openclaw-cn plugins install https://github.com/soimy/clawdbot-channel-dingtalk.git

在这里插入图片描述

在GLM-4 .7-Flash基础上加入deepseek

帮我再添加deepseek,我已经有deepseek的apikey了,当GLM-4.7 Flash免费tocken不够了的时候自动切换到deepseek

在这里插入图片描述


但是ai员工就卡住了。
原来是openclaw.json文件配置错了,在修改一下,加入apikey就可以了。
重启openclaw-cn gateway restart就可以。

在这里插入图片描述


在这里插入图片描述

加入minimax和豆包模型

  1. MiniMax M2.5 API Key获取访问MiniMax平台:https://api.minimax.chat/;
    注册/登录账号,直接获取API Key,复制保存备用。
  2. Seedance2.0 API Key获取访问火山方舟平台:https://console.volcengine.com/ark;
    注册/登录火山引擎账号,开通“火山方舟”服务;
    在“API Key管理”页面创建Key,复制保存备用。
    之后配置openclaw.json
    就可以接入多个大模型了。

配置web搜索

https://exa.ai/
这个可以免费注册

在这里插入图片描述


Sign In
之后填写信息:

在这里插入图片描述
在这里插入图片描述


之后我们copy setup prompt,以便后面使用
之后copy一下apikey,把这个apikey告诉openclaw,让它帮你配置就好了。
输入:
我注册了Exa.ai,并得到apikey:b9fxxxxx,请帮我配置exa.ai,代替Brave Search进行网络搜索。
之后重启openclaw-cn gateway restart,之后就可以。

# Exa API Setup Guide ## Your Configuration | Setting | Value | |---------|-------| | Coding Tool | Codex | | Framework | Other | | Use Case | Web search tool | | Search Type | Auto - Balanced relevance and speed (~1 second) | | Content | Full text | **Project Description:** (Not provided) --- ## API Key Setup ### Environment Variable ```bash export EXA_API_KEY="YOUR_API_KEY" 

.env File

EXA_API_KEY=YOUR_API_KEY 

🔌 Exa MCP Server for OpenAI Codex

Give OpenAI Codex real-time web search, code context, and company research with Exa MCP.

Run in terminal:

codex mcp add exa --url https://mcp.exa.ai/mcp?exaApiKey=f2492f04-b9f0-4c09-86a5-1ce6e1b9a24a 

Tool enablement (optional):
Add a tools= query param to the MCP URL.

Enable specific tools:

https://mcp.exa.ai/mcp?exaApiKey=f2492f04-b9f0-4c09-86a5-1ce6e1b9a24a&tools=web_search_exa,get_code_context_exa,people_search_exa 

Enable all tools:

https://mcp.exa.ai/mcp?exaApiKey=f2492f04-b9f0-4c09-86a5-1ce6e1b9a24a&tools=web_search_exa,web_search_advanced_exa,get_code_context_exa,crawling_exa,company_research_exa,people_search_exa,deep_researcher_start,deep_researcher_check 

Your API key:f2492f04-b9f0-4c09-86a5-1ce6e1b9a24a
Manage keys at dashboard.exa.ai/api-keys.

Troubleshooting: if tools don’t appear, restart your MCP client after updating the config.

📖 Full docs: docs.exa.ai/reference/exa-mcp


Quick Start

cURL

curl-X POST 'https://api.exa.ai/search'\-H'x-api-key: YOUR_API_KEY'\-H'Content-Type: application/json'\-d'{ "query": "latest developments in AI safety research", "type": "auto", "num_results": 10, "contents": { "text": { "max_characters": 20000 } } }'

Function Calling / Tool Use

Function calling (also known as tool use) allows your AI agent to dynamically decide when to search the web based on the conversation context. Instead of searching on every request, the LLM intelligently determines when real-time information would improve its response—making your agent more efficient and accurate.

Why use function calling with Exa?

  • Your agent can ground responses in current, factual information
  • Reduces hallucinations by fetching real sources when needed
  • Enables multi-step reasoning where the agent searches, analyzes, and responds

📚 Full documentation: https://docs.exa.ai/reference/openai-tool-calling

OpenAI Function Calling

import json from openai import OpenAI from exa_py import Exa openai = OpenAI() exa = Exa() tools =[{"type":"function","function":{"name":"exa_search","description":"Search the web for current information.","parameters":{"type":"object","properties":{"query":{"type":"string","description":"Search query"}},"required":["query"]}}}]defexa_search(query:str)->str: results = exa.search_and_contents(query,type="auto", num_results=10, text={"max_characters":20000})return"\n".join([f"{r.title}: {r.url}"for r in results.results]) messages =[{"role":"user","content":"What's the latest in AI safety?"}] response = openai.chat.completions.create(model="gpt-4o", messages=messages, tools=tools)if response.choices[0].message.tool_calls: tool_call = response.choices[0].message.tool_calls[0] search_results = exa_search(json.loads(tool_call.function.arguments)["query"]) messages.append(response.choices[0].message) messages.append({"role":"tool","tool_call_id": tool_call.id,"content": search_results}) final = openai.chat.completions.create(model="gpt-4o", messages=messages)print(final.choices[0].message.content)

Anthropic Tool Use

import anthropic from exa_py import Exa client = anthropic.Anthropic() exa = Exa() tools =[{"name":"exa_search","description":"Search the web for current information.","input_schema":{"type":"object","properties":{"query":{"type":"string","description":"Search query"}},"required":["query"]}}]defexa_search(query:str)->str: results = exa.search_and_contents(query,type="auto", num_results=10, text={"max_characters":20000})return"\n".join([f"{r.title}: {r.url}"for r in results.results]) messages =[{"role":"user","content":"Latest quantum computing developments?"}] response = client.messages.create(model="claude-sonnet-4-20250514", max_tokens=4096, tools=tools, messages=messages)if response.stop_reason =="tool_use": tool_use =next(b for b in response.content if b.type=="tool_use") tool_result = exa_search(tool_use.input["query"]) messages.append({"role":"assistant","content": response.content}) messages.append({"role":"user","content":[{"type":"tool_result","tool_use_id": tool_use.id,"content": tool_result}]}) final = client.messages.create(model="claude-sonnet-4-20250514", max_tokens=4096, tools=tools, messages=messages)print(final.content[0].text)

Search Type Reference

TypeBest ForSpeedDepth
fastReal-time apps, autocomplete, quick lookupsFastestBasic
autoMost queries - balanced relevance & speedMediumSmart
deepResearch, enrichment, thorough resultsSlowDeep
deep-reasoningComplex research, multi-step reasoningSlowestDeepest

Tip:type="auto" works well for most queries. Use type="deep" when you need thorough research results or structured outputs with field-level grounding.


Content Configuration

Choose ONE content type per request (not both):

TypeConfigBest For
Text"text": {"max_characters": 20000}Full content extraction, RAG
Highlights"highlights": {"max_characters": 4000}Snippets, summaries, lower cost

⚠️ Token usage warning: Using text: true (full page text) can significantly increase token count, leading to slower and more expensive LLM calls. To mitigate:

  • Add max_characters limit: "text": {"max_characters": 10000}
  • Use highlights instead if you don’t need contiguous text

When to use text vs highlights:

  • Text - When you need untruncated, contiguous content (e.g., code snippets, full articles, documentation)
  • Highlights - When you need key excerpts and don’t need the full context (e.g., summaries, Q&A, general research)

Domain Filtering (Optional)

Usually not needed - Exa’s neural search finds relevant results without domain restrictions.

When to use:

  • Targeting specific authoritative sources
  • Excluding low-quality domains from results

Example:

{"includeDomains":["arxiv.org","github.com"],"excludeDomains":["pinterest.com"]}

Note:includeDomains and excludeDomains can be used together to include a broad domain while excluding specific subdomains (e.g., "includeDomains": ["vercel.com"], "excludeDomains": ["community.vercel.com"]).


Web Search Tool

{"query":"latest developments in AI safety research","num_results":10,"contents":{"text":{"max_characters":20000}}}

Tips:

  • Use type: "auto" for most queries
  • Great for building search-powered chatbots or agents
  • Combine with contents for RAG workflows

Category Examples

Use category filters to search dedicated indexes. Each category returns only that content type.

Note: Categories can be restrictive. If you’re not getting enough results, try searching without a category first, then add one if needed.

People Search (category: "people")

Find people by role, expertise, or what they work on

curl-X POST 'https://api.exa.ai/search'\-H'x-api-key: YOUR_API_KEY'\-H'Content-Type: application/json'\-d'{ "query": "software engineer distributed systems", "category": "people", "type": "auto", "num_results": 10 }'

Tips:

  • Use SINGULAR form
  • Describe what they work on
  • No date/text filters supported

Company Search (category: "company")

Find companies by industry, criteria, or attributes

curl-X POST 'https://api.exa.ai/search'\-H'x-api-key: YOUR_API_KEY'\-H'Content-Type: application/json'\-d'{ "query": "AI startup healthcare", "category": "company", "type": "auto", "num_results": 10 }'

Tips:

  • Use SINGULAR form
  • Simple entity queries
  • Returns company objects, not articles

News Search (category: "news")

News articles

curl-X POST 'https://api.exa.ai/search'\-H'x-api-key: YOUR_API_KEY'\-H'Content-Type: application/json'\-d'{ "query": "OpenAI announcements", "category": "news", "type": "auto", "num_results": 10, "contents": { "text": { "max_characters": 20000 } } }'

Tips:

  • Use livecrawl: “preferred” for breaking news
  • Avoid date filters unless required

Research Papers (category: "research paper")

Academic papers

curl-X POST 'https://api.exa.ai/search'\-H'x-api-key: YOUR_API_KEY'\-H'Content-Type: application/json'\-d'{ "query": "transformer architecture improvements", "category": "research paper", "type": "auto", "num_results": 10, "contents": { "text": { "max_characters": 20000 } } }'

Tips:

  • Use type: “auto” for most queries
  • Includes arxiv.org, paperswithcode.com, and other academic sources

Tweet Search (category: "tweet")

Twitter/X posts

curl-X POST 'https://api.exa.ai/search'\-H'x-api-key: YOUR_API_KEY'\-H'Content-Type: application/json'\-d'{ "query": "AI safety discussion", "category": "tweet", "type": "auto", "num_results": 10, "contents": { "text": { "max_characters": 20000 } } }'

Tips:

  • Good for real-time discussions
  • Captures public sentiment

Content Freshness (maxAgeHours)

maxAgeHours sets the maximum acceptable age (in hours) for cached content. If the cached version is older than this threshold, Exa will livecrawl the page to get fresh content.

ValueBehaviorBest For
24Use cache if less than 24 hours old, otherwise livecrawlDaily-fresh content
1Use cache if less than 1 hour old, otherwise livecrawlNear real-time data
0Always livecrawl (ignore cache entirely)Real-time data where cached content is unusable
-1Never livecrawl (cache only)Maximum speed, historical/static content
(omit)Default behavior (livecrawl as fallback if no cache exists)Recommended — balanced speed and freshness

When LiveCrawl Isn’t Necessary:
Cached data is sufficient for many queries, especially for historical topics or educational content. These subjects rarely change, so reliable cached results can provide accurate information quickly.

See maxAgeHours docs for more details.


Other Endpoints

Beyond /search, Exa offers these endpoints:

EndpointDescriptionDocs
/contentsGet contents for known URLsDocs
/answerQ&A with citations from web searchDocs

Example - Get contents for URLs:

POST/contents {"urls":["https://example.com/article"],"text":{"max_characters":20000}}

Troubleshooting

Results not relevant?

  1. Try type: "auto" - most balanced option
  2. Try type: "deep" - runs multiple query variations and ranks the combined results
  3. Refine query - use singular form, be specific
  4. Check category matches your use case

Need structured data from search?

  1. Use type: "deep" or type: "deep-reasoning" with outputSchema
  2. Define the fields you need in the schema — Exa returns grounded JSON with citations

Results too slow?

  1. Use type: "fast"
  2. Reduce num_results
  3. Skip contents if you only need URLs

No results?

  1. Remove filters (date, domain restrictions)
  2. Simplify query
  3. Try type: "auto" - has fallback mechanisms

Resources

  • Docs: https://exa.ai/docs
  • Dashboard: https://dashboard.exa.ai
  • API Status: https://status.exa.ai

Read more

一键启动:用vLLM+Open-WebUI快速体验Qwen3-Embedding

一键启动:用vLLM+Open-WebUI快速体验Qwen3-Embedding 1. 引言:为什么选择 Qwen3-Embedding-4B? 在当前大模型驱动的知识库、语义搜索和向量化检索系统中,高质量文本嵌入模型(Text Embedding Model)已成为核心基础设施。阿里通义实验室于2025年8月开源的 Qwen3-Embedding-4B 模型,凭借其“中等体量、长上下文、多语言支持、高精度表现”四大优势,迅速成为开发者构建本地化知识系统的首选。 该模型基于 36层Dense Transformer架构,采用双塔编码结构,默认输出 2560维向量,最大支持 32k token输入长度,覆盖 119种自然语言与编程语言,在MTEB英文、中文、代码三项基准测试中分别达到74.60、68.09、73.50分,显著优于同参数规模的开源竞品。 更重要的是,它支持 Matryoshka Representation Learning (MRL)

【Js逆向 python】Web JS 逆向全体系详细解释

【Js逆向 python】Web JS 逆向全体系详细解释

Web JS 逆向全体系内容 互联网技术安全提示与职业操守 做渗透测试,必须严格遵守以下原则: 1. 合法授权:仅在书面授权的范围内使用逆向技术,禁止未授权测试; 2. 最小影响:避免使用高风险参数(如sqlmap工具的 --risk=3、--os-shell),防止目标服务崩溃; 3. 数据保护:枚举到的敏感数据(如用户密码)需严格保密,测试后立即删除; 4. 留痕清理:测试结束后,协助目标清除测试留下的日志、文件等痕迹。 免责声明 1. 本文所述所有渗透测试技术、工具、命令及实战案例,仅适用于已获得目标系统 / 网络所有者书面授权的测试场景(如企业内部安全评估、甲方委托的红队测试、个人合法拥有的实验环境)。 2. 任何组织或个人若未取得明确书面授权,擅自将本文内容用于对第三方系统 / 网络的扫描、探测、攻击等行为,均属于非法网络活动,涉嫌违反《中华人民共和国网络安全法》《中华人民共和国刑法》(第

服务端之NestJS接口响应message编写规范详解、写给前后端都舒服的接口、API提示信息标准化

服务端之NestJS接口响应message编写规范详解、写给前后端都舒服的接口、API提示信息标准化

MENU * 前言 * 定义 * 提示信息设计原则 * 提示信息风格分类 * 提示信息模板化设计 * 国际化与多语言支持 * 最佳实践 * 参考示例(NestJS响应) * 总结 * 统一风格示例清单推荐 * API响应message清单(可直接使用) 前言 在现代后端开发中,接口响应不仅仅是数据的传递,还承担着向前端或用户传递操作状态和结果的功能。一个规范、统一的message字段设计,可以显著提升系统的可维护性、前端开发效率和用户体验。 定义 响应结构示例(NestJS风格) 各字段作用 提示信息设计原则 简洁明了 1、不宜过长,一般3~12个汉字。 2、避免含糊不清的词,如“完成了”、“OK”等。 统一风格 1、同一项目接口建议使用统一动词+状态组合,例如:获取数据成功、数据加载完成。 上下文清晰 1、提示信息应体现操作对象或类型,如“用户列表获取成功”

springboot+vue基于web的咖啡点单程序设计

springboot+vue基于web的咖啡点单程序设计

目录 * 同行可拿货,招校园代理 ,本人源头供货商 * 功能模块划分 * 核心业务流程 * 技术实现要点 * 数据模型设计 * 前端交互特性 * 安全防护措施 * 扩展性设计 * 项目技术支持 * 源码获取详细视频演示 :文章底部获取博主联系方式!同行可合作 同行可拿货,招校园代理 ,本人源头供货商 功能模块划分 后端(SpringBoot) * 用户认证与授权(JWT/OAuth2) * 商品管理(咖啡品类、配料、价格) * 订单处理(创建/查询/状态更新) * 支付接口集成(微信/支付宝) * 数据统计与分析 前端(Vue) * 响应式用户界面(PC/移动端自适应) * 动态菜单展示与分类筛选 * 购物车实时交互 * 订单状态可视化追踪 * 用户评价系统 核心业务流程 用户侧流程 * 注册/登录 → 浏览菜单