diff --git a/README.md b/README.md index c5da4a3..d121165 100644 --- a/README.md +++ b/README.md @@ -10,6 +10,8 @@ For business cooperation, please contact email 389570357@qq.com ##### `最新`: +- 新增 SimulateDevDesignDiscussions,需要安装[swarm](https://github.com/openai/swarm),工作流下载[./workflow/swarm制作的播客节点workflow.json] + - 新增 SenseVoice - [新增JS-SDK,方便直接在前端项目中使用comfyui](https://github.com/shadowcz007/comfyui-js-sdk) diff --git a/__init__.py b/__init__.py index 9007c2a..bd6a263 100644 --- a/__init__.py +++ b/__init__.py @@ -32,7 +32,7 @@ # except: # print("##nodes.ChatGPT ImportError") -from .nodes.ChatGPT import openai_client +# from .nodes.ChatGPT import openai_client from .nodes.RembgNode import get_rembg_models,U2NET_HOME,run_briarmbg,run_rembg @@ -1261,7 +1261,7 @@ def mix_status(request): # print('\033[91m ### Mixlab Nodes: \033[93mLoaded') try: - from .nodes.ChatGPT import SiliconflowTextToImageNode,JsonRepair,ChatGPTNode,ShowTextForGPT,CharacterInText,TextSplitByDelimiter,SiliconflowFreeNode + from .nodes.ChatGPT import SimulateDevDesignDiscussions,SiliconflowTextToImageNode,JsonRepair,ChatGPTNode,ShowTextForGPT,CharacterInText,TextSplitByDelimiter,SiliconflowFreeNode logging.info('ChatGPT.available True') NODE_CLASS_MAPPINGS_V = { @@ -1271,7 +1271,9 @@ def mix_status(request): "ShowTextForGPT":ShowTextForGPT, "CharacterInText":CharacterInText, "TextSplitByDelimiter":TextSplitByDelimiter, - "JsonRepair":JsonRepair + "JsonRepair":JsonRepair, + + "SimulateDevDesignDiscussions":SimulateDevDesignDiscussions } # 一个包含节点友好/可读的标题的字典 @@ -1282,7 +1284,9 @@ def mix_status(request): "ShowTextForGPT":"Show Text ♾️MixlabApp", "CharacterInText":"Character In Text", "TextSplitByDelimiter":"Text Split By Delimiter", - "JsonRepair":"Json Repair" + "JsonRepair":"Json Repair", + + "SimulateDevDesignDiscussions":"SimulateDevDesignDiscussions ♾️Mixlab Podcast" } diff --git a/nodes/ChatGPT.py b/nodes/ChatGPT.py index 940879b..add5c8c 100644 --- a/nodes/ChatGPT.py +++ b/nodes/ChatGPT.py @@ -1,4 +1,6 @@ import openai +from swarm import Swarm, Agent + import time import urllib.error import re,json,os,string,random @@ -836,4 +838,255 @@ def run(self, json_string,key="",json_string2=None): # 将 Python 对象转换回 JSON 字符串,确保中文字符不被转义 json_str_with_chinese = json.dumps(data, ensure_ascii=False) - return (json_str_with_chinese,v,) \ No newline at end of file + return (json_str_with_chinese,v,) + + +# 以下为固定提示词的LLM节点示例 +class SimulateDevDesignDiscussions: + def __init__(self): + # self.__client = OpenAI() + self.session_history = [] # 用于存储会话历史的列表 + # self.seed=0 + self.system_content="You are ChatGPT, a large language model trained by OpenAI. Answer as concisely as possible." + + @classmethod + def INPUT_TYPES(cls): + + model_list=[ + "gpt-4o", + "gpt-4o-2024-05-13", + "gpt-4", + "gpt-4-0314", + "gpt-4-0613", + "qwen-turbo", + "qwen-plus", + "qwen-long", + "qwen-max", + "qwen-max-longcontext", + "glm-4", + "glm-3-turbo", + "moonshot-v1-8k", + "moonshot-v1-32k", + "moonshot-v1-128k", + "deepseek-chat", + "Qwen/Qwen2-7B-Instruct", + "THUDM/glm-4-9b-chat", + "01-ai/Yi-1.5-9B-Chat-16K" + ] + + return { + "required": { + "subject": ("STRING", {"multiline": True,"dynamicPrompts": False}), + "model": ( model_list, + {"default": model_list[0]}), + "api_url":(list(llm_apis_dict.keys()), + {"default": list(llm_apis_dict.keys())[0]}), + }, + "optional":{ + "api_key":("STRING", {"forceInput": True,}), + "custom_model_name":("STRING", {"forceInput": True,}), #适合自定义model + "custom_api_url":("STRING", {"forceInput": True,}), #适合自定义model + }, + + } + + RETURN_TYPES = ("STRING",) + RETURN_NAMES = ("text",) + FUNCTION = "generate_contextual_text" + CATEGORY = "♾️Mixlab/GPT" + INPUT_IS_LIST = False + OUTPUT_IS_LIST = (False,) + + def generate_contextual_text(self, + subject, + model, + api_url, + api_key=None, + custom_model_name=None, + custom_api_url=None, + ): + + # 设置黄色文本的ANSI转义序列 + YELLOW = "\033[33m" + # 重置文本颜色的ANSI转义序列 + RESET = "\033[0m" + + if custom_model_name!=None: + model=custom_model_name + + api_url=llm_apis_dict[api_url] if api_url in llm_apis_dict else "" + + if custom_api_url!=None: + api_url=custom_api_url + + if api_key==None: + api_key="lm_studio" + + print("api_key,api_url",api_key,api_url) + # + if is_azure_url(api_url): + client=azure_client(api_key,api_url) + else: + # 根据用户选择的模型,设置相应的接口和模型名称 + if model == "glm-4" : + client = ZhipuAI_client(api_key) # 使用 Zhipuai 的接口 + print('using Zhipuai interface') + else : + client = openai_client(api_key,api_url) # 使用 ChatGPT 的接口 + + + + # 以下为多智能体框架 + client = Swarm(client=client) + + # 定义两个代理:软件系统架构师和设计师 + software_architect_agent = Agent( + name="Software Architect", + instructions='''用脱口秀的风格回答编程问题,简短且口语化。 + + 输出格式 + ==== + + * 答案格式:`程序员:xxxxxxxxx` + + 示例 + == + + **输入:** + 如何优化代码性能? + + **输出:** + 程序员:兄弟,先把那些循环里的debug信息删掉,CPU都快哭了。''' + ) + + designer_agent = Agent( + name="Designer", + instructions='''回答问题时,请扮演一位具有多年空间设计和用户体验设计经验的设计师。你的回答应当天马行空,但又富有深度,带有苏格拉底的思考方式,并且使用脱口秀的风格。回答要简短且非常口语化。格式如下: + + 设计师:\[回答内容\] + + Output Format + ============= + + * 回答应当使用“设计师:\[回答内容\]”的格式。 + * 回答应当简短、口语化,富有创意和深度。 + + Examples + ======== + + **Example 1:** + + 主持人:你觉得未来的家会是什么样子? + + 设计师:未来的家?想象一下,房子会像变形金刚一样,随时变形满足你的需求。今天是健身房,明天是电影院,后天是游戏场。家不再是四面墙,而是一个随心所欲的魔法空间。 + + **Example 2:** + + 主持人:你怎么看待极简主义设计? + + 设计师:极简主义?就像吃寿司,去掉所有不必要的装饰,只留下最精华的部分。让空间呼吸,让心灵自由。 + + **Example 3:** + + 主持人:你觉得色彩在设计中有多重要? + + 设计师:色彩?哦,那可是设计的灵魂!就像人生中的调味料,一点红色让你激情澎湃,一点蓝色让你心如止水。色彩决定了空间的情绪基调。''' + ) + + # 定义一个函数,用于转移问题到designer_agent + def transfer_to_designer_agent(): + return designer_agent + + # 将转移函数添加到软件系统架构师和设计师的函数列表中 + software_architect_agent.functions.append(transfer_to_designer_agent) + + # 问题生成 + host_agent = Agent( + name="Host", + instructions=''' + 为播客的主持人生成4到5个问题,这些问题有些是针对设计师问的,有些是针对程序员问的。 + + * 主持人:你知道如何开发一款APP产品,从想法到上线吗? + * 主持人:站在设计师的角度,你怎么看? + * 主持人:不知道程序员又是怎么想的呢? + * 主持人:感谢大家的参与,今天收获蛮大的 + + Steps + ===== + + 1. 确定问题的对象:设计师或程序员。 + 2. 根据对象设计相关的问题,确保问题的多样性和深度。 + 3. 整理问题,使其符合播客主持人的风格和语气。 + + Output Format + ============= + + 问题列表,每个问题以“主持人:”开头,不要出现序号。 + + Examples + ======== + + * 主持人:作为一名设计师,你是如何开始一个新项目的? + * 主持人:程序员在开发过程中遇到的最大挑战是什么? + * 主持人:设计师在团队协作中扮演什么角色? + * 主持人:程序员如何确保代码的质量和稳定性? + * 主持人:感谢大家的参与,今天的讨论非常有意义。 + + Notes + ===== + + * 确保问题针对不同的角色(设计师和程序员)。 + * 保持问题的多样性,涵盖从项目开始到完成的各个阶段。 + * 确保问题能引导出深入的讨论和见解。 + ''') + + + response = client.run(agent=host_agent, messages=[{ + "role":"user", + "content":f"主题是‘{subject}’" + }],model_override=model) + + content=response.messages[-1]["content"] + print(f"{YELLOW}{content}{RESET}") + + texts=content.split("\n") + + # texts=''' + # 主持人:你知道如何开发一款APP产品,从想法到上线吗? + # 主持人:站在设计师的角度,你怎么看? + # 主持人:不知道程序员又是怎么想的呢? + # 主持人:感谢大家的参与,今天收获蛮大的 + # '''.split("\n") + + messages=[] + + texts = [text.strip() for text in texts if text.strip()] + + result=[] + + for text in texts: + messages.append({ + "role": "user", + "content": text + }) + + # 运行客户端,使用软件系统架构师作为初始代理 + response = client.run(agent=software_architect_agent, messages=messages,model_override=model) + + print(f"{text}") + result.append(text) + + # 输出最后一个响应消息的内容 + content=response.messages[-1]["content"] + print(f"{YELLOW}{content}{RESET}") + + result.append(content) + + messages.append({ + "role":"assistant", + "content":content + }) + + + return ("\n".join(result),) + diff --git a/pyproject.toml b/pyproject.toml index d250c71..a307145 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,7 +1,7 @@ [project] name = "comfyui-mixlab-nodes" description = "3D, ScreenShareNode & FloatingVideoNode, SpeechRecognition & SpeechSynthesis, GPT, LoadImagesFromLocal, Layers, Other Nodes, ..." -version = "0.44.0" +version = "0.45.0" license = "MIT" dependencies = ["numpy", "pyOpenSSL", "watchdog", "opencv-python-headless", "matplotlib", "openai", "simple-lama-inpainting", "clip-interrogator==0.6.0", "transformers>=4.36.0", "lark-parser", "imageio-ffmpeg", "rembg[gpu]", "omegaconf==2.3.0", "Pillow>=9.5.0", "einops==0.7.0", "trimesh>=4.0.5", "huggingface-hub", "scikit-image"] diff --git a/requirements.txt b/requirements.txt index 37cc568..7f2a135 100644 --- a/requirements.txt +++ b/requirements.txt @@ -33,3 +33,5 @@ natsort>=8.4.0 git+https://github.com/shadowcz007/SenseVoice-python.git faster_whisper + +git+https://github.com/openai/swarm.git diff --git a/web/javascript/checkVersion_mixlab.js b/web/javascript/checkVersion_mixlab.js index eaf9aa9..09fbaae 100644 --- a/web/javascript/checkVersion_mixlab.js +++ b/web/javascript/checkVersion_mixlab.js @@ -3,7 +3,7 @@ import { app } from '../../../scripts/app.js' const repoOwner = 'shadowcz007' // 替换为仓库的所有者 const repoName = 'comfyui-mixlab-nodes' // 替换为仓库的名称 -const version = 'v0.44.0' +const version = 'v0.45.0' fetch(`https://api.github.com/repos/${repoOwner}/${repoName}/releases/latest`) .then(response => response.json()) diff --git "a/workflow/swarm\345\210\266\344\275\234\347\232\204\346\222\255\345\256\242\350\212\202\347\202\271workflow.json" "b/workflow/swarm\345\210\266\344\275\234\347\232\204\346\222\255\345\256\242\350\212\202\347\202\271workflow.json" new file mode 100644 index 0000000..aab19cd --- /dev/null +++ "b/workflow/swarm\345\210\266\344\275\234\347\232\204\346\222\255\345\256\242\350\212\202\347\202\271workflow.json" @@ -0,0 +1,416 @@ +{ + "last_node_id": 9, + "last_link_id": 8, + "nodes": [ + { + "id": 3, + "type": "TextInput_", + "pos": [ + 137, + 420 + ], + "size": { + "0": 400, + "1": 200 + }, + "flags": {}, + "order": 0, + "mode": 0, + "outputs": [ + { + "name": "STRING", + "type": "STRING", + "links": [ + 2 + ], + "shape": 3, + "slot_index": 0 + } + ], + "title": "使用 Azure OpenAI", + "properties": { + "Node name for S&R": "TextInput_" + }, + "widgets_values": [ + "https://mixcopilot.openai.azure.com" + ] + }, + { + "id": 2, + "type": "KeyInput", + "pos": [ + 144, + 257 + ], + "size": { + "0": 315, + "1": 70 + }, + "flags": {}, + "order": 1, + "mode": 0, + "outputs": [ + { + "name": "key", + "type": "STRING", + "links": [ + 1 + ], + "shape": 3, + "slot_index": 0 + } + ], + "title": "使用你自己的key", + "properties": { + "Node name for S&R": "KeyInput" + }, + "widgets_values": [ + null, + null + ] + }, + { + "id": 6, + "type": "MultiPersonPodcast", + "pos": [ + 1099, + 480 + ], + "size": [ + 481.8963185574753, + 268.61682945154007 + ], + "flags": {}, + "order": 6, + "mode": 0, + "inputs": [ + { + "name": "speaker", + "type": "SPEAKER", + "link": 7, + "slot_index": 0 + }, + { + "name": "text", + "type": "STRING", + "link": 4, + "widget": { + "name": "text" + } + } + ], + "outputs": [ + { + "name": "audio_list", + "type": "AUDIO", + "links": null, + "shape": 3 + }, + { + "name": "audio", + "type": "AUDIO", + "links": [ + 8 + ], + "shape": 3, + "slot_index": 1 + } + ], + "properties": { + "Node name for S&R": "MultiPersonPodcast" + }, + "widgets_values": [ + "小明:大家好,欢迎收听本周的《AI新动态》。我是主持人小明,今天我们有两位嘉宾,分别是小李和小王。大家跟听众打个招呼吧!\n小李:大家好,我是小李,很高兴今天能和大家聊聊最新的AI动态。\n小王:大家好,我是小王,也很期待今天的讨论。", + 0, + 0, + 0, + 0, + false, + 1 + ] + }, + { + "id": 7, + "type": "LoadSpeaker", + "pos": [ + 584, + 567 + ], + "size": { + "0": 315, + "1": 58 + }, + "flags": {}, + "order": 2, + "mode": 0, + "outputs": [ + { + "name": "speaker", + "type": "SPEAKER", + "links": [ + 6 + ], + "shape": 3, + "slot_index": 0 + } + ], + "title": "opus", + "properties": { + "Node name for S&R": "LoadSpeaker" + }, + "widgets_values": [ + "opus_00001" + ] + }, + { + "id": 1, + "type": "SimulateDevDesignDiscussions", + "pos": [ + 611, + 201 + ], + "size": [ + 391.9864763335838, + 217.95792637114943 + ], + "flags": {}, + "order": 3, + "mode": 0, + "inputs": [ + { + "name": "api_key", + "type": "STRING", + "link": 1, + "widget": { + "name": "api_key" + } + }, + { + "name": "custom_model_name", + "type": "STRING", + "link": null, + "widget": { + "name": "custom_model_name" + } + }, + { + "name": "custom_api_url", + "type": "STRING", + "link": 2, + "widget": { + "name": "custom_api_url" + } + } + ], + "outputs": [ + { + "name": "text", + "type": "STRING", + "links": [ + 3, + 4 + ], + "shape": 3, + "slot_index": 0 + } + ], + "properties": { + "Node name for S&R": "SimulateDevDesignDiscussions" + }, + "widgets_values": [ + "数字艺术好看吗?", + "gpt-4o", + "openai", + "", + "", + "" + ] + }, + { + "id": 8, + "type": "RenameSpeaker", + "pos": [ + 593, + 681 + ], + "size": { + "0": 315, + "1": 58 + }, + "flags": {}, + "order": 4, + "mode": 0, + "inputs": [ + { + "name": "speaker", + "type": "SPEAKER", + "link": 6 + } + ], + "outputs": [ + { + "name": "speaker", + "type": "SPEAKER", + "links": [ + 7 + ], + "shape": 3, + "slot_index": 0 + } + ], + "properties": { + "Node name for S&R": "RenameSpeaker" + }, + "widgets_values": [ + "主持人" + ] + }, + { + "id": 5, + "type": "ShowTextForGPT", + "pos": [ + 1071, + 128 + ], + "size": [ + 624.2005965936271, + 279.47889630613906 + ], + "flags": {}, + "order": 5, + "mode": 0, + "inputs": [ + { + "name": "text", + "type": "STRING", + "link": 3, + "widget": { + "name": "text" + } + }, + { + "name": "output_dir", + "type": "STRING", + "link": null, + "widget": { + "name": "output_dir" + } + } + ], + "outputs": [ + { + "name": "STRING", + "type": "STRING", + "links": null, + "shape": 6 + } + ], + "properties": { + "Node name for S&R": "ShowTextForGPT" + }, + "widgets_values": [ + "", + "", + "* 主持人:作为一名设计师,你如何定义“好看”的数字艺术?\n设计师:好看的数字艺术?就像你在沙漠中看到绿洲的那一刻,它能吸引你的眼球,抓住你的心,它能传达情感,让人产生共鸣。可能是颜色的对撞,也可能是形状的魔法,总之,它让你想多看几眼,还想收藏到你的精神博物馆里。\n* 主持人:程序员,你们在开发支持数字艺术的软件时,如何确保用户体验的直观性和美观性?\n程序员:哎呀,这可是门艺术活啊!这时候我们可不像写代码那样呆板,想象力飞起来。我们会尽量让界面简洁好用,不搞那些让人摸不着头脑的功能。动效啥的也要调校好,太多就变花里胡哨了,太少用户觉得干巴巴。最重要的是,多听设计师的,他们可是颜值担当啊!\n* 主持人:站在设计师的角度,你觉得技术如何影响了数字艺术的表现力?\n设计师:技术啊,那可是我们的魔法棒!有了高端的硬件和软件,我们可以在屏幕上玩出各种花样,大到宇宙,小到细胞,想象力在技术的加持下,才能飞得更高更远。不管是3D渲染,还是AR互动,技术就是让我们的创意从草图变成现实的桥梁,让我们画布上的每一个像素都能发光。\n* 主持人:不知道程序员又是怎么看待数字艺术的后台开发和前端展示关系的呢?\n程序员:后端和前端就像魔法师和舞台演员。后端是幕后默默挥舞魔法杖,搞定数据处理啊、服务器啥的,让那台机器运转得顺溜。前端呢,就是站在舞台中央光彩夺目,把数据和功能打包成美美的界面展示给用户。说白了,后端是灵魂,前端是颜值,两个缺一不可,配合得好才是真正的艺术!\n* 主持人:感谢大家的参与,今天关于数字艺术的讨论让我受益匪浅。\n程序员:不客气,代码和艺术的碰撞总是火花四射!\n\n设计师:没错,灵感和技术结合,才能创作出让人惊艳的作品。期待下次再聊!" + ] + }, + { + "id": 9, + "type": "PreviewAudio", + "pos": [ + 1740, + 453 + ], + "size": { + "0": 315, + "1": 76 + }, + "flags": {}, + "order": 7, + "mode": 0, + "inputs": [ + { + "name": "audio", + "type": "AUDIO", + "link": 8 + } + ], + "properties": { + "Node name for S&R": "PreviewAudio" + }, + "widgets_values": [ + null + ] + } + ], + "links": [ + [ + 1, + 2, + 0, + 1, + 0, + "STRING" + ], + [ + 2, + 3, + 0, + 1, + 2, + "STRING" + ], + [ + 3, + 1, + 0, + 5, + 0, + "STRING" + ], + [ + 4, + 1, + 0, + 6, + 1, + "STRING" + ], + [ + 6, + 7, + 0, + 8, + 0, + "SPEAKER" + ], + [ + 7, + 8, + 0, + 6, + 0, + "SPEAKER" + ], + [ + 8, + 6, + 1, + 9, + 0, + "AUDIO" + ] + ], + "groups": [], + "config": {}, + "extra": { + "ds": { + "scale": 1.3310000000000006, + "offset": [ + -361.773434010461, + 27.423855709687306 + ] + } + }, + "version": 0.4 +} \ No newline at end of file