From fdd2739c66b5bf6117ff7e98885d8e208f39285a Mon Sep 17 00:00:00 2001
From: data-infra <825485697@qq.com>
Date: Mon, 11 Dec 2023 10:00:49 +0800
Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E8=84=9A=E6=9C=AC?=
=?UTF-8?q?=E7=9B=AE=E5=BD=95=E5=8F=98=E6=9B=B4?=
MIME-Version: 1.0
Content-Type: text/plain; charset=UTF-8
Content-Transfer-Encoding: 8bit
---
myapp/init-aihub.json | 7397 ----------------------------------
myapp/init-automl.json | 29 -
myapp/init-dataset.csv | 9 -
myapp/init-etl-pipeline.json | 1708 --------
myapp/init-inference.json | 93 -
myapp/init-job-template.json | 3141 ---------------
myapp/init-pipeline.json | 77 -
myapp/init-service.json | 75 -
myapp/init-train-model.json | 38 -
myapp/jinja_context.py | 223 -
10 files changed, 12790 deletions(-)
delete mode 100644 myapp/init-aihub.json
delete mode 100644 myapp/init-automl.json
delete mode 100644 myapp/init-dataset.csv
delete mode 100644 myapp/init-etl-pipeline.json
delete mode 100644 myapp/init-inference.json
delete mode 100644 myapp/init-job-template.json
delete mode 100644 myapp/init-pipeline.json
delete mode 100644 myapp/init-service.json
delete mode 100644 myapp/init-train-model.json
delete mode 100644 myapp/jinja_context.py
diff --git a/myapp/init-aihub.json b/myapp/init-aihub.json
deleted file mode 100644
index 642f9bf1..00000000
--- a/myapp/init-aihub.json
+++ /dev/null
@@ -1,7397 +0,0 @@
-[
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/openjourney",
- "field": "大模型",
- "scenes": "aigc",
- "frameworks": "pytorch",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "openjourney",
- "status": "online",
- "version": "v20221001",
- "uuid": "openjourney-v20221001",
- "label": "openjourney文生图",
- "describe": "Openjourney 是一个基于 Midjourney 图像的开源稳定扩散微调模型,由PromptHero提供",
- "pic": "example.png",
- "hot": "880",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/chat-embedding",
- "field": "自然语言",
- "scenes": "聊天机器人",
- "frameworks": "pytorch",
- "type": "inference,web",
- "name": "chat-embedding",
- "status": "online",
- "version": "v20221001",
- "uuid": "chat-embedding-v20221001",
- "label": "gpt私有知识库",
- "describe": "gpt私有知识库",
- "pic": "example.jpeg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- },
- "inference": {
- "resource_gpu": "0"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/chat-llama",
- "field": "大模型",
- "scenes": "",
- "frameworks": "pytorch",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "chat-llama",
- "status": "online",
- "version": "v20221001",
- "uuid": "llama-v20221001",
- "label": "llama 13B模型微调",
- "describe": "llama 13B模型微调",
- "pic": "example.png",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "resource_memory": "20G",
- "job_template_args": {
- "参数": {
- "--save_model_dir": {
- "type": "str",
- "item_type": "str",
- "label": "模型保存地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型保存地址",
- "editable": 1,
- "condition": ""
- },
- "--model_type": {
- "type": "str",
- "item_type": "str",
- "label": "模型类型",
- "require": 1,
- "choice": [
- "llama",
- "chatglm",
- "bloom"
- ],
- "range": "",
- "default": "llama",
- "placeholder": "",
- "describe": "模型类型",
- "editable": 1,
- "condition": ""
- },
- "--model_name_or_path": {
- "type": "str",
- "item_type": "str",
- "label": "hugging预训练模型名或者地址",
- "require": 1,
- "choice": [
- "decapoda-research/llama-7b-hf",
- "decapoda-research/llama-13b-hf",
- "decapoda-research/llama-65b-hf",
- "decapoda-research/llama-30b-hf",
- "THUDM/chatglm-6b",
- "bigscience/bloomz-3b",
- "bigscience/bloomz-7b1",
- "bigscience/bloomz-7b1-mt"
- ],
- "range": "",
- "default": "decapoda-research/llama-13b-hf",
- "placeholder": "",
- "describe": "hugging预训练模型名或者地址",
- "editable": 1,
- "condition": ""
- },
- "--dataset": {
- "type": "str",
- "item_type": "str",
- "label": "数据集名称或地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/chat-llama/dataset/CoT_Chinese_data.json",
- "placeholder": "",
- "describe": "数据集名称或地址",
- "editable": 1,
- "condition": ""
- },
- "--per_gpu_train_batch_size": {
- "type": "str",
- "item_type": "str",
- "label": "batch size",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "4",
- "placeholder": "",
- "describe": "batch size",
- "editable": 1,
- "condition": ""
- },
- "--learning_rate": {
- "type": "str",
- "item_type": "str",
- "label": "学习率",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "3e-4",
- "placeholder": "",
- "describe": "学习率",
- "editable": 1,
- "condition": ""
- },
- "--epochs": {
- "type": "str",
- "item_type": "str",
- "label": "迭代轮数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "3",
- "placeholder": "",
- "describe": "迭代轮数",
- "editable": 1,
- "condition": ""
- }
- }
- }
- },
- "inference": {
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/deoldify",
- "field": "机器视觉",
- "scenes": "图像合成",
- "type": "dateset,notebook,train,inference",
- "name": "deoldify",
- "status": "online",
- "version": "v20221001",
- "uuid": "deoldify-v20221001",
- "label": "图片上色",
- "describe": "图片上色",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/paddlespeech-tts",
- "field": "听觉",
- "scenes": "语音处理",
- "type": "dateset,notebook,train,inference",
- "name": "paddlespeech-tts",
- "status": "online",
- "version": "v20221114",
- "uuid": "paddle-speech-tts-v20221114",
- "label": "文字转语音",
- "describe": "文字转语音,支持280多种音色模型",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/paddlespeech-asr",
- "field": "听觉",
- "scenes": "语音处理",
- "type": "dateset,notebook,train,inference",
- "name": "paddlespeech-asr",
- "status": "online",
- "version": "v20221116",
- "uuid": "paddle-speech-asr-v20221116",
- "label": "语音转文字",
- "describe": "语音转文字,支持中英文",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/stable-diffusion-zh-en",
- "field": "大模型",
- "scenes": "图像创作",
- "type": "dateset,notebook,train,inference",
- "name": "stable-diffusion-zh-en",
- "status": "online",
- "version": "v20221122",
- "uuid": "stable-diffusion-zh-en-v20221122",
- "label": "文字转图像-中英文混合9种语言",
- "describe": "输入一串文字描述,可生成相应的图片,暂已支持语言:英语(En)、中文(Zh)、西班牙语(Es)、法语(Fr)、俄语(Ru)、日语(Ja)、韩语(Ko)、阿拉伯语(Ar)和意大利语(It)",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/paddleocr",
- "field": "机器视觉",
- "scenes": "图像识别",
- "type": "dateset,notebook,train,inference",
- "name": "paddleocr",
- "status": "online",
- "version": "v20221001",
- "uuid": "paddleocr-v20221001",
- "label": "ocr识别",
- "describe": "paddleocr提供的ocr识别",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/ddddocr",
- "field": "机器视觉",
- "scenes": "图像识别",
- "type": "dateset,notebook,train,inference",
- "name": "ddddocr",
- "status": "online",
- "version": "v20221001",
- "uuid": "ddddocr-v20221001",
- "label": "验证码识别",
- "describe": "ai识别验证码文字和验证码目标",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/gfpgan",
- "field": "机器视觉",
- "scenes": "图像合成",
- "type": "dateset,notebook,train,inference",
- "name": "gfpgan",
- "status": "online",
- "version": "v20221001",
- "uuid": "gfpgan-v20221001",
- "label": "图片修复",
- "describe": "低分辨率照片修复,清晰度增强",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/stable-diffusion",
- "field": "大模型",
- "scenes": "图像创作",
- "type": "dateset,notebook,train,inference",
- "name": "stable-diffusion",
- "status": "online",
- "version": "v20221022",
- "uuid": "stable-diffusion-v20221022",
- "label": "文字转图像",
- "describe": "输入一串文字描述,可生成相应的图片",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/cartoon-sd",
- "field": "机器视觉",
- "scenes": "图像创作",
- "type": "dateset,notebook,train,inference",
- "name": "cartoon-sd",
- "status": "online",
- "version": "v20221125",
- "uuid": "cartoon-sd-v20221125",
- "label": "文字转图像-动漫版",
- "describe": "输入一串文字描述,可生成相应的动漫图片,描述越详细越好哦~",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/paddlespeech-cls",
- "field": "听觉",
- "scenes": "语音处理",
- "type": "dateset,notebook,train,inference",
- "name": "paddlespeech-cls",
- "status": "online",
- "version": "v20221114",
- "uuid": "paddle-speech-cls-v20221114",
- "label": "语音场景分类",
- "describe": "语音场景分类:语种识别等",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/humanseg",
- "field": "机器视觉",
- "scenes": "图像合成",
- "type": "dateset,notebook,train,inference",
- "name": "humanseg",
- "status": "online",
- "version": "v20221001",
- "uuid": "humanseg-v20221001",
- "label": "人体分割背景替换",
- "describe": "人体分割背景替换,视频会议背景替换",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/panoptic",
- "field": "机器视觉",
- "scenes": "目标识别",
- "type": "dateset,notebook,train,inference",
- "name": "panoptic",
- "status": "online",
- "version": "v20221001",
- "uuid": "panoptic-v20221001",
- "label": "图片识别",
- "describe": "resnet50 图像识别",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/animegan",
- "field": "机器视觉",
- "scenes": "图像合成",
- "type": "dateset,notebook,train,inference",
- "name": "animegan",
- "status": "online",
- "version": "v20221001",
- "uuid": "animegan-v20221001",
- "label": "动漫风格化",
- "describe": "图片的全新动漫风格化,宫崎骏或新海诚风格的动漫,以及4种关于人脸的风格转换。",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {},
- "inference": {}
- },
- {
- "price": "1",
- "name": "speech-paraformer-large-asr-nat-zh-cn-16k-common-vocab8404-pytorch",
- "label": "Paraformer语音识别-中文-通用-16k-离线-large-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 8993619,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-large-asr-nat-zh-cn-16k-common-vocab8404-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-fft-inpainting-lama",
- "label": "LaMa图像填充",
- "describe": "针对自然图片进行填充恢复,支持高分辨率图像的输入,同时支持在线refinement,使得高分辨率图片恢复出更加真实的内容细节",
- "hot": 1344866,
- "pic": "example.jpg",
- "uuid": "cv-fft-inpainting-lama",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像填充",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-unet-person-image-cartoon-compound-models",
- "label": "DCT-Net人像卡通化",
- "describe": "该模型采用全新的DCT-Net(Domain-Calibrated Translation) 域校准图像翻译模型,利用小样本的风格数据,即可得到高保真、强鲁棒、易拓展的人像风格转换模型。",
- "hot": 505792,
- "pic": "example.jpg",
- "uuid": "cv-unet-person-image-cartoon-compound-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像卡通化",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-news",
- "label": "RaNER命名实体识别-中文-新闻领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文MSRA数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 436414,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-news",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-text-generation-2-7b",
- "label": "GPT-3预训练生成模型-中文-2.7B",
- "describe": "2.7B参数量的中文GPT-3文本生成模型",
- "hot": 243767,
- "pic": "example.jpg",
- "uuid": "nlp-gpt3-text-generation-2-7b",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-resnet-carddetection-scrfd34gkps",
- "label": "卡证检测矫正模型",
- "describe": "输入一张图片,检测其中是否出现卡证,如有则返回卡证的矩形框和角点,以及矫正后的卡证图像。",
- "hot": 239476,
- "pic": "example.jpg",
- "uuid": "cv-resnet-carddetection-scrfd34gkps",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "卡证检测矫正",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet18-human-detection",
- "label": "人体检测-通用-Base",
- "describe": "给定一张输入图像,输出图像中人体的坐标。",
- "hot": 223218,
- "pic": "example.jpeg",
- "uuid": "cv-resnet18-human-detection",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人体检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet50-face-detection-retinaface",
- "label": "RetinaFace人脸检测关键点模型",
- "describe": "给定一张图片,返回图片中人脸区域的位置和五点关键点。RetinaFace为当前学术界和工业界精度较高的人脸检测和人脸关键点定位二合一的方法,被CVPR 2020 录取。该方法的主要贡献是: 引入关键点分支,可以在训练阶段引入关键点预测分支进行多任务学习,提供额外的互补特征,inference去掉关键点分支即可,并不会引入额外的计算量。",
- "hot": 150064,
- "pic": "example.png",
- "uuid": "cv-resnet50-face-detection-retinaface",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-zh2en",
- "label": "CSANMT连续语义增强机器翻译-中英-通用领域-large",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 143809,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-zh2en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-structbert-nli-chinese-tiny",
- "label": "StructBERT自然语言推理-中文-通用-tiny",
- "describe": "StructBERT自然语言推理-中文-通用-tiny是在structbert-tiny-chinese预训练模型的基础上,用CMNLI、OCNLI两个数据集(45.8w条数据)训练出来的自然语言推理模型",
- "hot": 126986,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-nli-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "自然语言推理",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-unet-person-image-cartoon-handdrawn-compound-models",
- "label": "DCT-Net人像卡通化-手绘",
- "describe": "该模型采用全新的DCT-Net(Domain-Calibrated Translation) 域校准图像翻译模型,利用小样本的风格数据,即可得到高保真、强鲁棒、易拓展的人像手绘风格转换模型。",
- "hot": 117239,
- "pic": "example.jpg",
- "uuid": "cv-unet-person-image-cartoon-handdrawn-compound-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像卡通化",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-paraformer-large-vad-punc-asr-nat-zh-cn-16k-common-vocab8404-pytorch",
- "label": "Paraformer语音识别-中文-通用-16k-离线-large-长音频版",
- "describe": "Paraformer-large长音频模型集成VAD、ASR、标点与时间戳功能,可直接对时长为数小时音频进行识别,并输出带标点文字与时间戳",
- "hot": 113854,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-large-vad-punc-asr-nat-zh-cn-16k-common-vocab8404-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-resnet-image-quality-assessment-mos-youtubeugc",
- "label": "图像质量MOS评估",
- "describe": "通过模型预测图像MOS分",
- "hot": 102942,
- "pic": "example.jpg",
- "uuid": "cv-resnet-image-quality-assessment-mos-youtubeugc",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像质量MOS评估",
- "field": "机器视觉"
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/cv-unet-person-image-cartoon-3d-compound-models",
- "field": "机器视觉",
- "scenes": "",
- "frameworks": "TensorFlow",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "cv-unet-person-image-cartoon-3d-compound-models",
- "status": "online",
- "version": "v20221001",
- "uuid": "cv-unet-person-image-cartoon-3d-compound-models-v20221001",
- "label": "DCT-Net人像卡通化-3D",
- "describe": "该模型采用全新的DCT-Net(Domain-Calibrated Translation) 域校准图像翻译模型,利用小样本的风格数据,即可得到高保真、强鲁棒、易拓展的人像3D风格转换模型。",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "job_template_args": {
- "参数": {
- "--save_model_dir": {
- "type": "str",
- "item_type": "str",
- "label": "模型保存地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型保存地址",
- "editable": 1,
- "condition": ""
- },
- "--data_photo": {
- "type": "str",
- "item_type": "str",
- "label": "源图片的目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/cv-unet-person-image-cartoon-3d-compound-models/dataset/face_photo",
- "placeholder": "",
- "describe": "源图片的目录",
- "editable": 1,
- "condition": ""
- },
- "--data_cartoon": {
- "type": "str",
- "item_type": "str",
- "label": "卡通图片的目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/cv-unet-person-image-cartoon-3d-compound-models/dataset/face_cartoon",
- "placeholder": "",
- "describe": "卡通图片的目录",
- "editable": 1,
- "condition": ""
- }
- }
- }
- },
- "inference": {
- "resource_memory": "4G",
- "resource_cpu": "7",
- "resource_gpu": "1"
- }
- },
- {
- "price": "1",
- "name": "cv-unet-person-image-cartoon-artstyle-compound-models",
- "label": "DCT-Net人像卡通化-艺术",
- "describe": "该模型采用全新的DCT-Net(Domain-Calibrated Translation) 域校准图像翻译模型,利用小样本的风格数据,即可得到高保真、强鲁棒、易拓展的人像艺术风格转换模型。",
- "hot": 84664,
- "pic": "example.jpg",
- "uuid": "cv-unet-person-image-cartoon-artstyle-compound-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像卡通化",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-unet-person-image-cartoon-sketch-compound-models",
- "label": "DCT-Net人像卡通化-素描",
- "describe": "该模型采用全新的DCT-Net(Domain-Calibrated Translation) 域校准图像翻译模型,利用小样本的风格数据,即可得到高保真、强鲁棒、易拓展的人像素描风格转换模型。",
- "hot": 84574,
- "pic": "example.jpg",
- "uuid": "cv-unet-person-image-cartoon-sketch-compound-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像卡通化",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "punc-ct-transformer-zh-cn-common-vocab272727-pytorch",
- "label": "CT-Transformer标点-中文-通用-pytorch",
- "describe": "中文标点通用模型:可用于语音识别模型输出文本的标点预测。",
- "hot": 80822,
- "pic": "example.jpg",
- "uuid": "punc-ct-transformer-zh-cn-common-vocab272727-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "标点预测",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-text-generation-chinese-base",
- "label": "GPT-3预训练生成模型-中文-base",
- "describe": "1亿参数量的中文GPT-3文本生成模型",
- "hot": 77547,
- "pic": "example.jpg",
- "uuid": "nlp-gpt3-text-generation-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-unet-image-matting",
- "label": "BSHM人像抠图",
- "describe": "人像抠图对输入含有人像的图像进行处理,无需任何额外输入,实现端到端人像抠图,输出四通道人像抠图结果。",
- "hot": 73859,
- "pic": "example.jpg",
- "uuid": "cv-unet-image-matting",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像抠图",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet-face-recognition-facemask",
- "label": "口罩人脸识别模型FaceMask",
- "describe": "",
- "hot": 73459,
- "pic": "example.png",
- "uuid": "cv-resnet-face-recognition-facemask",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-zh-cn-16k-common-vocab8358-tensorflow1",
- "label": "Paraformer语音识别-中文-通用-16k-离线",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 69982,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-zh-cn-16k-common-vocab8358-tensorflow1",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/cv-tinynas-object-detection-damoyolo",
- "field": "机器视觉",
- "scenes": "",
- "type": "dateset,notebook,train,inference",
- "name": "cv-tinynas-object-detection-damoyolo",
- "status": "online",
- "version": "v20221001",
- "uuid": "cv-tinynas-object-detection-damoyolo-v20221001",
- "label": "DAMOYOLO-高性能通用检测模型-S",
- "describe": "DAMOYOLO是一款面向工业落地的高性能检测框架,精度和速度超越当前的一众典型YOLO框架(YOLOE、YOLOv6、YOLOv7)。基于TinyNAS技术,DAMOYOLO能够针对不同的硬件算力,进行低成本的模型定制化搜索。这里仅提供DAMOYOLO-S模型,更多模型请参考README。",
- "pic": "example.jpg",
- "hot": "3344866",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "resource_gpu": "1",
- "resource_memory": "40G",
- "resource_cpu": "10",
- "job_template_args": {
- "参数": {
- "--save_model_dir": {
- "type": "str",
- "item_type": "str",
- "label": "模型保存地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型保存地址",
- "editable": 1,
- "condition": ""
- },
- "--train_image_dir": {
- "type": "str",
- "item_type": "str",
- "label": "训练集图片的目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/cv-tinynas-object-detection-damoyolo/dataset/train_image_dir",
- "placeholder": "",
- "describe": "训练图片的目录",
- "editable": 1,
- "condition": ""
- },
- "--val_image_dir": {
- "type": "str",
- "item_type": "str",
- "label": "校验集图片的目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/cv-tinynas-object-detection-damoyolo/dataset/val_image_dir",
- "placeholder": "",
- "describe": "校验集图片的目录",
- "editable": 1,
- "condition": ""
- },
- "--train_ann": {
- "type": "str",
- "item_type": "str",
- "label": "训练集标注文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/cv-tinynas-object-detection-damoyolo/dataset/visdrone_train.json",
- "placeholder": "",
- "describe": "训练集标注文件路径",
- "editable": 1,
- "condition": ""
- },
- "--val_ann": {
- "type": "str",
- "item_type": "str",
- "label": "校验集标注文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/{{creator}}/cube-studio/aihub/deep-learning/cv-tinynas-object-detection-damoyolo/dataset/visdrone_val.json",
- "placeholder": "",
- "describe": "校验集标注文件路径",
- "editable": 1,
- "condition": ""
- },
- "--max_epochs": {
- "type": "str",
- "item_type": "str",
- "label": "最大迭代数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "3",
- "placeholder": "",
- "describe": "最大迭代数",
- "editable": 1,
- "condition": ""
- },
- "--num_classes": {
- "type": "str",
- "item_type": "str",
- "label": "分类类型个数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "80",
- "placeholder": "",
- "describe": "分类类型个数",
- "editable": 1,
- "condition": ""
- }
- }
- }
- },
- "inference": {
- }
- },
- {
- "price": "1",
- "name": "ofa-image-caption-coco-large-en",
- "label": "OFA图像描述-英文-通用领域-large",
- "describe": "根据用户输入的任意图片,AI智能创作模型3秒内快速写出“一句话描述”,可用于图像标签和图像简介。",
- "hot": 67927,
- "pic": "example.jpg",
- "uuid": "ofa-image-caption-coco-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-fsmn-vad-zh-cn-16k-common-pytorch",
- "label": "FSMN语音端点检测-中文-通用-16k",
- "describe": "FSMN-Monophone VAD模型,可用于检测长语音片段中有效语音的起止时间点。",
- "hot": 65670,
- "pic": "example.png",
- "uuid": "speech-fsmn-vad-zh-cn-16k-common-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音端点检测",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-en2zh",
- "label": "CSANMT连续语义增强机器翻译-英中-通用领域-large",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 57507,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-en2zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-unet-skin-retouching",
- "label": "ABPN人像美肤",
- "describe": "人像美肤模型对输入含有人像的图像进行处理,无需任何额外输入,实现脸部皮肤区域匀肤(处理痘印、肤色不均等)、去瑕疵(脂肪粒、斑点、痣等)及全身皮肤区域美白。模型仅对皮肤区域进行处理,不影响其他区域。",
- "hot": 56970,
- "pic": "example.jpg",
- "uuid": "cv-unet-skin-retouching",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像美肤",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "mgeo-geographic-entity-alignment-chinese-base",
- "label": "MGeo地址相似度匹配实体对齐-中文-地址领域-base",
- "describe": "模型判断两条地址是否指代同一道路、村庄、POI等。将两条地址的关系分为完全对齐、部分对齐、不对齐。该任务是构建地理信息知识库的核心技术。",
- "hot": 50831,
- "pic": "example.jpg",
- "uuid": "mgeo-geographic-entity-alignment-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet18-ocr-detection-line-level-damo",
- "label": "读光-文字检测-行检测模型-中英-通用领域",
- "describe": "给定一张图片,检测出图中所含文字的外接框的端点的坐标值。",
- "hot": 49225,
- "pic": "example.jpg",
- "uuid": "cv-resnet18-ocr-detection-line-level-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-convnexttiny-ocr-recognition-general-damo",
- "label": "读光-文字识别-行识别模型-中英-通用领域",
- "describe": "给定一张图片,识别出图中所含文字并输出字符串。",
- "hot": 48023,
- "pic": "example.jpg",
- "uuid": "cv-convnexttiny-ocr-recognition-general-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-swinl-panoptic-segmentation-cocopan",
- "label": "Mask2Former-SwinL全景分割",
- "describe": "基于Mask2Former架构,SwinL为backbone的全景分割模型。训练数据库为COCO-Panoptic",
- "hot": 45790,
- "pic": "example.jpg",
- "uuid": "cv-swinl-panoptic-segmentation-cocopan",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-word-segmentation-chinese-base",
- "label": "BAStructBERT分词-中文-新闻领域-base",
- "describe": "基于预训练语言模型的新闻领域中文分词模型,根据用户输入的中文句子产出分词结果。",
- "hot": 43605,
- "pic": "example.png",
- "uuid": "nlp-structbert-word-segmentation-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-gpen-image-portrait-enhancement",
- "label": "GPEN人像修复增强",
- "describe": "GPEN将预训练好的StyleGAN2网络作为decoder嵌入到人像修复模型中,并通过finetune的方式最终实现修复功能,在多项指标上达到行业领先的效果。",
- "hot": 41899,
- "pic": "example.jpg",
- "uuid": "cv-gpen-image-portrait-enhancement",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像增强",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentiment-classification-chinese-base",
- "label": "StructBERT情感分类-中文-通用-base",
- "describe": "StructBERT情感分类-中文-通用-base是基于bdci、dianping、jd binary、waimai-10k四个数据集(11.5w条数据)训练出来的情感分类模型。",
- "hot": 41828,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentiment-classification-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-structbert-zero-shot-classification-chinese-base",
- "label": "StructBERT零样本分类-中文-base",
- "describe": "该模型使用StructBERT-base在xnli_zh数据集(将英文数据集重新翻译得到中文数据集)上面进行了训练得到。",
- "hot": 41454,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-zero-shot-classification-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "零样本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-f3net-product-segmentation",
- "label": "图像分割-商品展示图场景的商品分割-电商领域",
- "describe": "通用商品分割模型,适用于商品展示图场景",
- "hot": 41106,
- "pic": "example.jpg",
- "uuid": "cv-f3net-product-segmentation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用商品分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-ecom-50cls",
- "label": "RaNER命名实体识别-中文-电商领域-细粒度-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文细粒度电商数据集训练的模型。本方法采用Transformer-CRF模型,使用sbert-base作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 38901,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-ecom-50cls",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "promptclue",
- "label": "全中文任务支持零样本学习模型",
- "describe": "支持近20中文任务,并具有零样本学习能力。 针对理解类任务,如分类、情感分析、抽取等,可以自定义标签体系;针对生成任务,可以进行采样自由生成。使用1000亿中文token(字词级别)进行大规模预训练,累计学习1.5万亿中文token,并且在100+任务上进行多任务学习获得。",
- "hot": 34930,
- "pic": "example.png",
- "uuid": "promptclue",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "端到端文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zhiyan-emo-zh-cn-16k",
- "label": "语音合成-中文-多情感领域-16k-发音人Zhiyan",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 34612,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zhiyan-emo-zh-cn-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "ofa-image-caption-muge-base-zh",
- "label": "OFA图像描述-中文-电商领域-base",
- "describe": "根据用户输入的任意商品图片,AI智能创作模型3秒内快速写出“商品描述”。",
- "hot": 34157,
- "pic": "example.png",
- "uuid": "ofa-image-caption-muge-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-frcrn-ans-cirm-16k",
- "label": "FRCRN语音降噪-单麦-16k",
- "describe": "支持音频通话场景和各种噪声环境下语音音频录音的单通道语音智能降噪模型算法。模型输入和输出均为16kHz采样率单通道语音时域波形信号,输入信号可有单通道麦克风直接进行录制,输出为噪声抑制后的语音音频信号[1]。",
- "hot": 33671,
- "pic": "example.png",
- "uuid": "speech-frcrn-ans-cirm-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音降噪",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-unet-image-face-fusion-damo",
- "label": "图像人脸融合",
- "describe": "给定一张模板图像和一张用户图像,图像人脸融合模型能够自动地将用户图中的人脸融合到模板人脸图像中,生成一张包含用户图人脸特征的新图像。",
- "hot": 31748,
- "pic": "example.jpg",
- "uuid": "cv-unet-image-face-fusion-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像人脸融合",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-ddcolor-image-colorization",
- "label": "DDColor图像上色",
- "describe": "DDColor 是最新的图像上色算法,输入一张黑白图像,返回上色处理后的彩色图像,并能够实现自然生动的上色效果。",
- "hot": 31108,
- "pic": "example.jpg",
- "uuid": "cv-ddcolor-image-colorization",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像上色",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "mgeo-geographic-elements-tagging-chinese-base",
- "label": "MGeo门址地址结构化要素解析-中文-地址领域-base",
- "describe": "模型用于识别门址地址中的常见要素,例如:行政区划信息、路网信息、POI (兴趣点)、楼栋号、户室号等。",
- "hot": 30450,
- "pic": "example.jpg",
- "uuid": "mgeo-geographic-elements-tagging-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "序列标注",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-nafnet-image-denoise-sidd",
- "label": "NAFNet图像去噪",
- "describe": "NAFNet(Nonlinear Activation Free Network)提出了一个简单的基线,计算效率高。其不需要使用非线性激活函数(Sigmoid、ReLU、GELU、Softmax等),可以达到SOTA性能。",
- "hot": 29541,
- "pic": "example.png",
- "uuid": "cv-nafnet-image-denoise-sidd",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像降噪",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "chatyuan-large",
- "label": "元语功能型对话大模型",
- "describe": "元语功能型对话大模型这个模型可以用于问答、结合上下文做对话、做各种生成任务,包括创意性写作,也能回答一些像法律、新冠等领域问题。它基于PromptCLUE-large结合数亿条功能对话多轮对话数据进一步训练得到。",
- "hot": 28628,
- "pic": "example.jpeg",
- "uuid": "chatyuan-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "端到端文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-crnn-ocr-recognition-general-damo",
- "label": "读光-文字识别-CRNN模型-中英-通用领域",
- "describe": "",
- "hot": 27699,
- "pic": "example.jpg",
- "uuid": "cv-crnn-ocr-recognition-general-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-convai-text2sql-pretrain-cn",
- "label": "SPACE-T表格问答预训练模型-中文-通用领域-base",
- "describe": "SPACE-T表格问答预训练模型-中文-通用领域-base大规模预训练模型,基于transformers架构,在千万级中文表格,亿级中文表格训练数据上进行预训练,在中文跨领域、多轮、Text-to-SQL语义解析等任务上能取得很好的效果。",
- "hot": 27121,
- "pic": "example.jpeg",
- "uuid": "nlp-convai-text2sql-pretrain-cn",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "表格问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-zh-cn-16k-common-vocab8358-tensorflow1-online",
- "label": "UniASR语音识别-中文-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 25820,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-zh-cn-16k-common-vocab8358-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "mplug-image-captioning-coco-base-en",
- "label": "mPLUG图像描述模型-英文-base",
- "describe": "达摩MPLUG英文图像描述base模型",
- "hot": 25449,
- "pic": "example.jpg",
- "uuid": "mplug-image-captioning-coco-base-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-poetry-generation-chinese-large",
- "label": "GPT-3诗词生成模型-中文-large",
- "describe": "3亿参数量的中文GPT-3诗词生成模型",
- "hot": 24427,
- "pic": "example.jpeg",
- "uuid": "nlp-gpt3-poetry-generation-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zh-cn-16k",
- "label": "语音合成-中文-多情感领域-16k-多发音人",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 24327,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zh-cn-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-resnet-facedetection-scrfd10gkps",
- "label": "SCRFD人脸检测关键点模型",
- "describe": "输入图片,检测其中的人脸区域及5点关键点,支持检测极大/极小脸和任意角度人脸。",
- "hot": 24183,
- "pic": "example.jpg",
- "uuid": "cv-resnet-facedetection-scrfd10gkps",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-cmeee",
- "label": "RaNER命名实体识别-中文-医疗领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文CMeEE数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 23528,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-cmeee",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "ofa-visual-grounding-refcoco-large-zh",
- "label": "OFA通过描述定位图像物体-中文-通用领域-large",
- "describe": "中文视觉定位任务:给定一张图片,一段描述,通过描述找到图片对应的物体。",
- "hot": 22558,
- "pic": "example.jpg",
- "uuid": "ofa-visual-grounding-refcoco-large-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉定位",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-palm2-0-text-generation-chinese-base",
- "label": "PALM 2.0摘要生成模型-中文-base",
- "describe": "本任务是PALM通用预训练生成模型,在英文CNN/Dail Mail和中文LCSTS上进行finetune的文本摘要生成下游任务。",
- "hot": 22276,
- "pic": "example.jpeg",
- "uuid": "nlp-palm2-0-text-generation-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "mplug-visual-question-answering-coco-large-en",
- "label": "mPLUG视觉问答模型-英文-large",
- "describe": "本任务是mPLUG,在英文VQA数据集进行finetune的视觉问答下游任务。给定一个问题和图片,通过图片信息来给出答案。",
- "hot": 22072,
- "pic": "example.jpg",
- "uuid": "mplug-visual-question-answering-coco-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉问答",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-csrnet-image-color-enhance-models",
- "label": "CSRNet图像调色",
- "describe": "基于CSRNet实现的图像色彩增强算法,输入待增强图像,输出色彩增强后的图像。CSRNet通过计算全局调整参数并将之作用于条件网络得到的特征,保证效果的基础之上实现轻便高效的训练和推理。",
- "hot": 21993,
- "pic": "example.jpg",
- "uuid": "cv-csrnet-image-color-enhance-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像颜色增强",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-vit-base-image-classification-dailylife-labels",
- "label": "ViT图像分类-中文-日常物品",
- "describe": "自建1300类常见物体标签体系,覆盖常见的日用品,动物,植物,家具,设备,食物等物体,标签从海量中文互联网社区语料进行提取,保留了出现频率较高的常见物体名称。模型结构采用最新的ViT-Base结构。",
- "hot": 21950,
- "pic": "example.jpg",
- "uuid": "cv-vit-base-image-classification-dailylife-labels",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用分类",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "multi-modal-clip-vit-large-patch14-336-zh",
- "label": "CLIP模型-中文-通用领域-large-336分辨率",
- "describe": "本项目为CLIP模型的中文版本,使用大规模中文数据进行训练(~2亿图文对),旨在帮助用户实现中文领域的跨模态检索、图像表示等。视觉encoder采用vit结构,文本encoder采用roberta结构。 模型在多个中文图文检索数据集上进行了效果测试。",
- "hot": 21116,
- "pic": "example.jpg",
- "uuid": "multi-modal-clip-vit-large-patch14-336-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "多模态表征",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-resnest101-general-recognition",
- "label": "万物识别-中文-通用领域",
- "describe": "本模型是对包含主体物体的图像进行标签识别,无需任何额外输入,输出主体物体的类别标签,目前已经覆盖了5W多类的物体类别,几乎囊括了日常所有物体。",
- "hot": 19389,
- "pic": "example.jpg",
- "uuid": "cv-resnest101-general-recognition",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "万物识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet50-video-category",
- "label": "短视频内容分类模型-中文-通用领域",
- "describe": "本模型采用ResNet-50网络结构提取视觉特征,并利用NextVLAD网络对连续视频帧进行特征聚合。本模型是对短视频进行内容分类,输入视频片段,输出视频内容分类,目前已经覆盖了23个一级类目/160个二级类目。",
- "hot": 19022,
- "pic": "example.jpg",
- "uuid": "cv-resnet50-video-category",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "短视频内容识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "mplug-image-captioning-coco-large-en",
- "label": "mPLUG图像描述模型-英文-large",
- "describe": "达摩MPLUG英文图像描述large模型",
- "hot": 17743,
- "pic": "example.jpg",
- "uuid": "mplug-image-captioning-coco-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zhitian-emo-zh-cn-16k",
- "label": "语音合成-中文-多情感领域-16k-发音人Zhitian",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 17437,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zhitian-emo-zh-cn-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-word-segmentation-chinese-base-ecommerce",
- "label": "BAStructBERT分词-中文-电商领域-base",
- "describe": "基于预训练语言模型的电商领域中文分词模型,根据用户输入的中文句子产出分词结果。",
- "hot": 17284,
- "pic": "example.png",
- "uuid": "nlp-structbert-word-segmentation-chinese-base-ecommerce",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-convnexttiny-ocr-recognition-document-damo",
- "label": "读光-文字识别-行识别模型-中英-文档印刷体文本领域",
- "describe": "给定一张文档印刷体图片,识别出图中所含文字并输出字符串。",
- "hot": 17181,
- "pic": "example.jpg",
- "uuid": "cv-convnexttiny-ocr-recognition-document-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-bart-text-error-correction-chinese",
- "label": "BART文本纠错-中文-通用领域-large",
- "describe": "我们采用seq2seq方法建模文本纠错任务。模型训练上,我们使用中文BART作为预训练模型,然后在Lang8和HSK训练数据上进行finetune。不引入额外资源的情况下,本模型在NLPCC18测试集上达到了SOTA。",
- "hot": 17104,
- "pic": "example.png",
- "uuid": "nlp-bart-text-error-correction-chinese",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本纠错",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-rom-passage-ranking-chinese-base",
- "label": "ROM语义相关性-中文-通用领域-base",
- "describe": "基于ROM-Base预训练模型的通用领域中文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 17073,
- "pic": "example.png",
- "uuid": "nlp-rom-passage-ranking-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "promptclue-base-v1-5",
- "label": "全中文任务支持零样本学习模型v1.5",
- "describe": "支持近20中文任务,并具有零样本学习能力。 针对理解类任务,如分类、情感分析、抽取等,可以自定义标签体系;针对生成任务,可以进行采样自由生成。使用1000亿中文token(字词级别)进行大规模预训练,累计学习1.5万亿中文token,并且在100+任务上进行多任务学习获得。",
- "hot": 17035,
- "pic": "example.png",
- "uuid": "promptclue-base-v1-5",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "端到端文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-hrnetv2w32-body-2d-keypoints-image",
- "label": "HRNet人体关键点-2D",
- "describe": "输入一张人物图像,实现端到端的人体关键点检测,输出图像中所有人体的15点人体关键点、点位置信度和人体检测框。",
- "hot": 16873,
- "pic": "example.jpeg",
- "uuid": "cv-hrnetv2w32-body-2d-keypoints-image",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人体2D关键点",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-word-segmentation-chinese-lite",
- "label": "BAStructBERT分词-中文-新闻领域-lite",
- "describe": "基于预训练语言模型的新闻领域中文分词模型,根据用户输入的中文句子产出分词结果。",
- "hot": 16733,
- "pic": "example.png",
- "uuid": "nlp-structbert-word-segmentation-chinese-lite",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-convnexttiny-ocr-recognition-handwritten-damo",
- "label": "读光-文字识别-行识别模型-中英-手写文本领域",
- "describe": "给定一张手写体图片,识别出图中所含文字并输出字符串。",
- "hot": 16719,
- "pic": "example.jpg",
- "uuid": "cv-convnexttiny-ocr-recognition-handwritten-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-nanodet-face-human-hand-detection",
- "label": "目标检测-人脸人体人手-通用领域",
- "describe": "通用场景下的,人脸-人体-人手三合一目标检测",
- "hot": 16165,
- "pic": "example.jpg",
- "uuid": "cv-nanodet-face-human-hand-detection",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸人体人手三合一检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-ja-16k-common-vocab93-tensorflow1-offline",
- "label": "UniASR语音识别-日语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 15999,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-ja-16k-common-vocab93-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-chinese-base",
- "label": "CoROM文本向量-中文-通用领域-base",
- "describe": "基于CoROM-base预训练语言模型的通用领域中文文本表示模型,基于输入的句子产出对应的文本向量,文本向量可以使用在下游的文本检索、句子相似度计算、文本聚类等任务中。",
- "hot": 15883,
- "pic": "example.png",
- "uuid": "nlp-corom-sentence-embedding-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-large-generic",
- "label": "RaNER命名实体识别-中文-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在中文数据集MultiCoNER-ZH-Chinese训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 15855,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "ofa-visual-question-answering-pretrain-large-en",
- "label": "OFA视觉问答-英文-通用领域-large",
- "describe": "视觉问答任务:给定一张图片和一个关于图片的问题,要求模型正确作答。",
- "hot": 15507,
- "pic": "example.jpg",
- "uuid": "ofa-visual-question-answering-pretrain-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉问答",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-googlenet-pgl-video-summarization",
- "label": "PGL_SUM视频摘要-Web视频领域",
- "describe": "视频摘要,输入一段长视频,算法对视频进行镜头切割得到视频片段,评估视频帧的重要性,输出重要视频帧的帧号,根据帧号可以合成一段短视频(摘要视频)。",
- "hot": 15496,
- "pic": "example.jpg",
- "uuid": "cv-googlenet-pgl-video-summarization",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频摘要",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-r50-panoptic-segmentation-cocopan",
- "label": "Mask2Former-R50全景分割",
- "describe": "基于Mask2Former架构,resnet50为backbone的全景分割模型。训练数据库为COCO-Panoptic。支持finetune。",
- "hot": 15238,
- "pic": "example.jpg",
- "uuid": "cv-r50-panoptic-segmentation-cocopan",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-visual-grounding-refcoco-large-en",
- "label": "OFA通过描述定位图像物体-英文-通用领域-large",
- "describe": "视觉定位任务:给定一张图片,一段描述,通过描述找到图片对应的物体。",
- "hot": 15110,
- "pic": "example.jpg",
- "uuid": "ofa-visual-grounding-refcoco-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉定位",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-palm2-0-text-generation-commodity-chinese-base",
- "label": "PALM 2.0商品文案生成-中文-base",
- "describe": "达摩PALM 2.0中文商品文案生成base模型",
- "hot": 15044,
- "pic": "example.jpg",
- "uuid": "nlp-palm2-0-text-generation-commodity-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-corom-passage-ranking-chinese-base-ecom",
- "label": "CoROM语义相关性-中文-电商领域-base",
- "describe": "基于ROM-Base预训练模型的电商领域中文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 14893,
- "pic": "example.jpg",
- "uuid": "nlp-corom-passage-ranking-chinese-base-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "spring-couplet-generation",
- "label": "春联生成模型-中文-base",
- "describe": "春联生成模型是达摩院AliceMind团队利用基础生成大模型在春联场景的应用,该模型可以通过输入两字随机祝福词,生成和祝福词相关的春联。",
- "hot": 14747,
- "pic": "example.jpg",
- "uuid": "spring-couplet-generation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-chinese-base-ecom",
- "label": "CoROM文本向量-中文-电商领域-base",
- "describe": "基于CoROM-base预训练语言模型的电商领域中文文本表示模型,基于输入的句子产出对应的文本向量,文本向量可以使用在下游的文本检索、句子相似度计算、文本聚类等任务中。",
- "hot": 14303,
- "pic": "example.jpg",
- "uuid": "nlp-corom-sentence-embedding-chinese-base-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-gan-face-image-generation",
- "label": "StyleGAN2人脸生成",
- "describe": "StyleGAN是图像生成领域的代表性工作,StyleGAN2在StyleGAN的基础上,采用Weight Demodulation取代AdaIN等改进极大的减少了water droplet artifacts等,生成结果有了质的提升,甚至能达到以假乱真的程度。",
- "hot": 14190,
- "pic": "example.jpg",
- "uuid": "cv-gan-face-image-generation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸生成",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-text-generation-13b",
- "label": "GPT-3预训练生成模型-中文-13B",
- "describe": "13B参数量的中文GPT-3文本生成模型",
- "hot": 14182,
- "pic": "example.jpg",
- "uuid": "nlp-gpt3-text-generation-13b",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "nlp-structbert-word-segmentation-chinese-lite-ecommerce",
- "label": "BAStructBERT分词-中文-电商领域-lite",
- "describe": "基于预训练语言模型的电商领域中文分词模型,根据用户输入的中文句子产出分词结果。",
- "hot": 13880,
- "pic": "example.png",
- "uuid": "nlp-structbert-word-segmentation-chinese-lite-ecommerce",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-mt5-zero-shot-augment-chinese-base",
- "label": "全任务零样本学习-mT5分类增强版-中文-base",
- "describe": "该模型在mt5模型基础上使用了大量中文数据进行训练,并引入了零样本分类增强的技术,使模型输出稳定性大幅提升。支持任务包含:分类、摘要、翻译、阅读理解、问题生成等等。",
- "hot": 13750,
- "pic": "example.jpg",
- "uuid": "nlp-mt5-zero-shot-augment-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "端到端文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zh-cn-multisp-pretrain-16k",
- "label": "SambertHifigan语音合成-中文-多人预训练-16k",
- "describe": "中文语音合成16k采样率多人预训练模型",
- "hot": 13580,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zh-cn-multisp-pretrain-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-aams-style-transfer-damo",
- "label": "AAMS图像风格迁移",
- "describe": "给定内容图像和风格图像作为输入,风格迁移模型会自动地将内容图像的风格、纹理特征变换为风格图像的类型,同时保证图像的内容特征不变",
- "hot": 13553,
- "pic": "example.jpg",
- "uuid": "cv-aams-style-transfer-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "风格迁移",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-rrdb-image-super-resolution",
- "label": "RealESRGAN图像超分辨率-x4",
- "describe": "RealESRGAN提出了通过多次降质的方式来模拟真实复杂降质,相比较于之前的简单下采样,能够更好处理真实的低分辨率场景。",
- "hot": 13192,
- "pic": "example.png",
- "uuid": "cv-rrdb-image-super-resolution",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像超分辨",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-charctc-kws-phone-xiaoyun",
- "label": "CTC语音唤醒-移动端-单麦-16k-小云小云",
- "describe": "移动端语音唤醒模型,检测关键词为“小云小云”。模型主体为4层FSMN结构,使用CTC训练准则,参数量750K,适用于移动端设备运行。",
- "hot": 13079,
- "pic": "example.jpg",
- "uuid": "speech-charctc-kws-phone-xiaoyun",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音唤醒",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-convnexttiny-ocr-recognition-licenseplate-damo",
- "label": "读光-文字识别-行识别模型-中英-车牌文本领域",
- "describe": "给定一张车牌图片,识别出图中所含文字并输出字符串。",
- "hot": 12755,
- "pic": "example.jpg",
- "uuid": "cv-convnexttiny-ocr-recognition-licenseplate-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-ocr-recognition-general-base-zh",
- "label": "OFA文字识别-中文-通用场景-base",
- "describe": "基于OFA模型的finetune后的OCR文字识别任务,基于通用数据集训练,比特定数据集finetune效果相差不大。",
- "hot": 12734,
- "pic": "example.jpg",
- "uuid": "ofa-ocr-recognition-general-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "ofa-ocr-recognition-scene-base-zh",
- "label": "OFA文字识别-中文-日常场景-base",
- "describe": "基于OFA模型的finetune后的OCR文字识别任务,可有效识别日常场景的文字内容,比如广告牌、店铺名等等",
- "hot": 12566,
- "pic": "example.jpg",
- "uuid": "ofa-ocr-recognition-scene-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-swin-b-image-instance-segmentation-coco",
- "label": "CascadeMaskRCNN-SwinB图像实例分割",
- "describe": "基于Cascade mask rcnn架构,backbone为swin transformer模型。",
- "hot": 11968,
- "pic": "example.jpeg",
- "uuid": "cv-swin-b-image-instance-segmentation-coco",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-image-caption-coco-huge-en",
- "label": "OFA图像描述-英文-通用领域-huge",
- "describe": "根据用户输入的任意图片,AI智能创作模型3秒内快速写出“一句话描述”,可用于图像标签和图像简介。",
- "hot": 11960,
- "pic": "example.jpg",
- "uuid": "ofa-image-caption-coco-huge-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zhibei-emo-zh-cn-16k",
- "label": "语音合成-中文-多情感领域-16k-发音人Zhibei",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 11915,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zhibei-emo-zh-cn-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zhizhe-emo-zh-cn-16k",
- "label": "语音合成-中文-多情感领域-16k-发音人Zhizhe",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 11643,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zhizhe-emo-zh-cn-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-kuakua-robot-chinese-large",
- "label": "GPT-3夸夸机器人-中文-large",
- "describe": "GPT-3夸夸机器人,主要用于夸夸场景,我们训练的机器人可以针对用户的不同输入进行全方位无死角的夸,同时针对相同的输入重复调用模型会得到不同的夸奖词",
- "hot": 11532,
- "pic": "example.jpg",
- "uuid": "nlp-gpt3-kuakua-robot-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-convnexttiny-ocr-recognition-scene-damo",
- "label": "读光-文字识别-行识别模型-中英-自然场景文本领域",
- "describe": "给定一张自然场景图片,识别出图中所含文字并输出字符串。",
- "hot": 11512,
- "pic": "example.jpg",
- "uuid": "cv-convnexttiny-ocr-recognition-scene-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-lstmcrf-word-segmentation-chinese-news",
- "label": "LSTM分词-中文-新闻领域",
- "describe": "char-BiLSTM-CRF中文新闻领域分词模型",
- "hot": 11184,
- "pic": "example.jpg",
- "uuid": "nlp-lstmcrf-word-segmentation-chinese-news",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-vitb16-segmentation-shop-seg",
- "label": "商品显著性图像分割-电商领域",
- "describe": "商品显著性分割模型,对商品图像提取显著性区域mask",
- "hot": 11016,
- "pic": "example.jpg",
- "uuid": "cv-vitb16-segmentation-shop-seg",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "商品显著性分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-passage-ranking-english-base",
- "label": "CoROM语义相关性-英文-通用领域-base",
- "describe": "基于CoROM-Base预训练模型的通用领域英文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 10988,
- "pic": "example.jpg",
- "uuid": "nlp-corom-passage-ranking-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-mfcca-asr-zh-cn-16k-alimeeting-vocab4950",
- "label": "MFCCA多通道多说话人语音识别-中文-AliMeeting-16k-离线",
- "describe": "考虑到麦克风阵列不同麦克风接收信号的差异,该模型采用了一种多帧跨通道注意力机制,该方法对相邻帧之间的跨通道信息进行建模,以利用帧级和通道级信息的互补性。此外,还引入了一种多层卷积模块以融合多通道输出和一种通道掩码策略以解决训练和推理之间的音频通道数量不匹配的问题。在ICASSP2022 M2MeT竞赛上发布的真实会议场景语料库AliMeeting上进行了相关实验,该多通道模型在Eval和Test集上比单通道模型CER分别相对降低了39.9%和37.0%。此外,在同等的模型参数和训练数据下,本文提出的模型获得的识别性能超越竞赛期间最佳结果,在AliMeeting上实现了目前最新的SOTA性能。",
- "hot": 10800,
- "pic": "example.jpg",
- "uuid": "speech-mfcca-asr-zh-cn-16k-alimeeting-vocab4950",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-lstmcrf-word-segmentation-chinese-ecommerce",
- "label": "LSTM分词-中文-电商领域",
- "describe": "char-biLSTM-CRF中文电商领域分词模型",
- "hot": 10781,
- "pic": "example.jpg",
- "uuid": "nlp-lstmcrf-word-segmentation-chinese-ecommerce",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-unet-image-colorization",
- "label": "DeOldify图像上色",
- "describe": "DeOldify是图像上色领域比较有名的开源算法,模型利用resnet作为encoder构建一个unet结构的网络,并提出了多个不同的训练版本,在效果、效率、鲁棒性等等方面有良好的综合表现。",
- "hot": 10429,
- "pic": "example.jpg",
- "uuid": "cv-unet-image-colorization",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像上色",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-bert-relation-extraction-chinese-base-commerce",
- "label": "DIRECT商品评价解析-中文-电商-base",
- "describe": "",
- "hot": 10210,
- "pic": "example.jpg",
- "uuid": "nlp-bert-relation-extraction-chinese-base-commerce",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "关系抽取",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-mobilenet-face-2d-keypoints-alignment",
- "label": "106点人脸关键点-通用领域-2D",
- "describe": "人脸2d关键点对齐模型",
- "hot": 10205,
- "pic": "example.jpg",
- "uuid": "cv-mobilenet-face-2d-keypoints-alignment",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸2D关键点",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-m",
- "label": "DAMOYOLO-高性能通用检测模型-M",
- "describe": "DAMOYOLO是一款面向工业落地的高性能检测框架,精度和速度超越当前的一众典型YOLO框架(YOLOE、YOLOv6、YOLOv7)。基于TinyNAS技术,DAMOYOLO能够针对不同的硬件算力,进行低成本的模型定制化搜索。这里仅提供DAMOYOLO-M模型,更多模型请参考README。",
- "hot": 10138,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-m",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-english-base",
- "label": "CoROM文本向量-英文-通用领域-base",
- "describe": "基于CoROM-Base预训练语言模型的通用领域英文文本表示模型,基于输入的句子产出对应的连续文本向量,改文本向量可以使用在下游的文本检索、句子相似度计算、文本聚类等任务中。",
- "hot": 10019,
- "pic": "example.png",
- "uuid": "nlp-corom-sentence-embedding-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-paraformer-large-asr-nat-zh-cn-16k-common-vocab8358-tensorflow1",
- "label": "Paraformer语音识别-中文-通用-16k-离线-large",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 9786,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-large-asr-nat-zh-cn-16k-common-vocab8358-tensorflow1",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentence-similarity-chinese-large",
- "label": "StructBERT文本相似度-中文-通用-large",
- "describe": "StructBERT文本相似度-中文-通用-large是在structbert-large-chinese预训练模型的基础上,用atec、bq_corpus、chineseSTS、lcqmc、paws-x-zh五个数据集(52.5w条数据,正负比例0.48:0.52)训练出来的相似度匹配模型。",
- "hot": 9713,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentence-similarity-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnetc3d-action-detection-detection2d",
- "label": "日常动作检测",
- "describe": "输入视频文件,输出该段时间内视频所包含的动作,当前支持9中常见动作识别",
- "hot": 9648,
- "pic": "example.jpg",
- "uuid": "cv-resnetc3d-action-detection-detection2d",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "动作检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-flow-based-body-reshaping-damo",
- "label": "FBBR人体美型",
- "describe": "给定一张单个人物图像(半身或全身),无需任何额外输入,人体美型模型能够端到端地实现对人物身体区域(肩部,腰部,腿部等)的自动化美型处理。",
- "hot": 9427,
- "pic": "example.gif",
- "uuid": "cv-flow-based-body-reshaping-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人体美型",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-generic",
- "label": "RaNER命名实体识别-中文-通用领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文Ontonotes4.0数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 9398,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-tinynas-classification",
- "label": "TinyNAS高性能图像分类网络结构模型",
- "describe": "ZenNet 是基于 Tiny-NAS (Zen-NAS) 算法设计出的高效的卷积网络结构。 本 demo 只提供 zennet_imagenet1k_latency12ms_res22 backbone,其它网络结构可以从README 中获取。",
- "hot": 9348,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-classification",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用分类",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentiment-classification-chinese-ecommerce-base",
- "label": "StructBERT情感分类-中文-电商-base",
- "describe": "StructBERT中文情感分类模型是基于百万电商评价数据训练出来的情感分类模型",
- "hot": 9332,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentiment-classification-chinese-ecommerce-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-u2net-salient-detection",
- "label": "U2Net图像显著性检测",
- "describe": "给定一张输入图像,输出视觉显著注意力程度图(归一化至0~255)。",
- "hot": 9325,
- "pic": "example.jpg",
- "uuid": "cv-u2net-salient-detection",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-bert-document-segmentation-chinese-base",
- "label": "BERT文本分割-中文-通用领域",
- "describe": "该模型基于wiki-zh公开语料训练,对未分割的长文本进行段落分割。提升未分割文本的可读性以及下游NLP任务的性能。",
- "hot": 9324,
- "pic": "example.jpg",
- "uuid": "nlp-bert-document-segmentation-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分割",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-ecom",
- "label": "RaNER命名实体识别-中文-电商领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文电商数据集训练的模型。本方法采用Transformer-CRF模型,使用sbert-base作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 9250,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-chinese-base-medical",
- "label": "CoROM文本向量-中文-医疗领域-base",
- "describe": "基于ROM-Base预训练模型的医疗领域中文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 9192,
- "pic": "example.jpg",
- "uuid": "nlp-corom-sentence-embedding-chinese-base-medical",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-en2zh-base",
- "label": "CSANMT连续语义增强机器翻译-英中-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 8981,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-en2zh-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-hrnet-crowd-counting-dcanet",
- "label": "DCANet人群密度估计-多域",
- "describe": "采用单一模型就可以同时针对多个不同域的数据进行精确预测,是multidomain crowd counting中经典的方法",
- "hot": 8952,
- "pic": "example.jpeg",
- "uuid": "cv-hrnet-crowd-counting-dcanet",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人群密度估计",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-vitb-video-single-object-tracking-ostrack",
- "label": "OSTrack视频单目标跟踪-通用领域",
- "describe": "该模型采用基于OSTrack的Transformer方案,输入视频和对应第一帧的待跟踪目标物体矩形框,可端对端推理得到待跟踪目标物体在每一帧图片的跟踪矩形框。",
- "hot": 8918,
- "pic": "example.gif",
- "uuid": "cv-vitb-video-single-object-tracking-ostrack",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频单目标跟踪",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-passvitb-image-reid-person-market",
- "label": "行人图像特征表示提取-Market1501",
- "describe": "基于图片的行人图像特征表示(image embedding)提取模型。输入图像,可提取并输出图像的特征表示,后续能够利用该特征表示进行后续的相似度计算和图像排序。",
- "hot": 8814,
- "pic": "example.jpg",
- "uuid": "cv-passvitb-image-reid-person-market",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "行人重识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentiment-classification-chinese-large",
- "label": "StructBERT情感分类-中文-通用-large",
- "describe": "StructBERT情感分类-中文-通用-large是基于bdci、dianping、jd binary、waimai-10k四个数据集(11.5w条数据)训练出来的情感分类模型",
- "hot": 8780,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentiment-classification-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-palm2-0-text-generation-chinese-large",
- "label": "PALM 2.0摘要生成模型-中文-large",
- "describe": "PALM 2.0中文摘要生成large模型",
- "hot": 8648,
- "pic": "example.jpeg",
- "uuid": "nlp-palm2-0-text-generation-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "u2pp-conformer-asr-cn-16k-online",
- "label": "WeNet-U2pp_Conformer-语音识别-中文-16k-实时",
- "describe": "WeNet 是一款面向工业落地应用的语音识别工具包,提供了从语音识别模型的训练到部署的一条龙服务。我们使用 conformer 网络结构和 CTC/attention loss 联合优化方法,统一的流式/非流式语音识别方案,具有业界一流的识别效果;提供云上和端上直接部署的方案,最小化模型训练和产品落地之间的工程工作;框架简洁,模型训练部分完全基于 pytorch 生态,不依赖于 kaldi 等复杂的工具。 详细的注释和文档,非常适合用于学习端到端语音识别的基础知识和实现细节。 支持时间戳,对齐,端点检测,语言模型等相关功能。",
- "hot": 8565,
- "pic": "example.png",
- "uuid": "u2pp-conformer-asr-cn-16k-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-text-generation-chinese-large",
- "label": "GPT-3预训练生成模型-中文-large",
- "describe": "3亿参数量的中文GPT-3文本生成模型",
- "hot": 8543,
- "pic": "example.jpg",
- "uuid": "nlp-gpt3-text-generation-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "erlangshen-roberta-330m-sentiment",
- "label": "二郎神-RoBERTa-330M-情感分类",
- "describe": "二郎神-RoBERTa-330M-情感分类",
- "hot": 8526,
- "pic": "example.jpg",
- "uuid": "erlangshen-roberta-330m-sentiment",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-stable-diffusion-v2-image-inpainting-base",
- "label": "StableDiffusionV2图像填充",
- "describe": "借助Stable Diffusion强大的生成能力,StableDiffusionv2图像填充模型能够补全缺失的图像区域,生成效果自然而真实;不仅如此,除了自适应填充背景内容外,用户还可以通过指定引导文字在缺失区域生成指定内容,畅享AI生成的乐趣。",
- "hot": 8478,
- "pic": "example.gif",
- "uuid": "cv-stable-diffusion-v2-image-inpainting-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像填充",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-resnet101-face-detection-cvpr22papermogface",
- "label": "MogFace人脸检测模型-large",
- "describe": "Wider Face榜单冠军模型。给定一张图片,检测图片中的人脸区域,支持小脸检测。",
- "hot": 7965,
- "pic": "example.jpg",
- "uuid": "cv-resnet101-face-detection-cvpr22papermogface",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-zh-cn-multisp-pretrain-24k",
- "label": "SambertHifigan语音合成-中文-多人预训练-24k",
- "describe": "语音合成中文24k采样率多人预训练模型",
- "hot": 7940,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-zh-cn-multisp-pretrain-24k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-zh-cn-16k-common-vocab8358-tensorflow1-offline",
- "label": "UniASR语音识别-中文-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 7908,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-zh-cn-16k-common-vocab8358-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentence-similarity-chinese-base",
- "label": "StructBERT文本相似度-中文-通用-base",
- "describe": "StructBERT文本相似度-中文-通用-base是在structbert-base-chinese预训练模型的基础上,用atec、bq_corpus、chineseSTS、lcqmc、paws-x-zh五个数据集(52.5w条数据,正负比例0.48:0.52)训练出来的相似度匹配模型。由于license权限问题,目前只上传了BQ_Corpus、chineseSTS、LCQMC这三个数据集。",
- "hot": 7897,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentence-similarity-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-vit-base-image-classification-imagenet-labels",
- "label": "ViT图像分类-通用",
- "describe": "本模型适用范围较广,支持ImageNet 1000类物体识别,也可作为下游任务的预训练backbone。",
- "hot": 7787,
- "pic": "example.jpg",
- "uuid": "cv-vit-base-image-classification-imagenet-labels",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用分类",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-ocr-recognition-handwriting-base-zh",
- "label": "OFA文字识别-中文-手写体-base",
- "describe": "基于OFA模型的finetune后的OCR文字识别任务,可有效识别手写体文字。",
- "hot": 7753,
- "pic": "example.jpg",
- "uuid": "ofa-ocr-recognition-handwriting-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "ofa-image-caption-coco-distilled-en",
- "label": "OFA图像描述-英文-通用领域-蒸馏33M",
- "describe": "根据用户输入的任意图片,AI智能创作模型3秒内快速写出“一句话描述”,可用于图像标签和图像简介。",
- "hot": 7718,
- "pic": "example.jpg",
- "uuid": "ofa-image-caption-coco-distilled-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-plug-text-generation-27b",
- "label": "PLUG预训练生成模型-中文-27B",
- "describe": "PLUG是一个270亿参数的大规模中文理解和生成联合预训练模型,由海量高质量中文文本预训练得到,在中文的多个下游理解和生成任务上,该模型效果达到state-of-the-art水平,且具有零样本生成能力。",
- "hot": 7634,
- "pic": "example.jpg",
- "uuid": "nlp-plug-text-generation-27b",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-dfsmn-aec-psm-16k",
- "label": "DFSMN回声消除-单麦单参考-16k",
- "describe": "支持音频通话场景的单通道回声消除模型算法。模型接受单通道麦克风信号和单通道参考信号作为输入,输出回声消除和残余抑制后的音频信号[1]。模型采用Deep FSMN结构,提取原始观测信号以及线性滤波后信号的Fbank特征作为输入,预测输出目标语音的Phase senstive mask。",
- "hot": 7546,
- "pic": "example.png",
- "uuid": "speech-dfsmn-aec-psm-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "回声消除",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-cn-en-moe-16k-vocab8358-tensorflow1-offline",
- "label": "UniASR语音识别-中文英文混-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 7433,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-cn-en-moe-16k-vocab8358-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-tinynas-human-detection-damoyolo",
- "label": "实时人体检测-通用",
- "describe": "本模型为高性能热门应用系列检测模型中的实时人体检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有人体的坐标信息。更多具体信息请参考Model card。",
- "hot": 7408,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-human-detection-damoyolo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-en2fr",
- "label": "CSANMT连续语义增强机器翻译-英法-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 7235,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-en2fr",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-cspnet-image-object-detection-yolox",
- "label": "实时目标检测-通用领域",
- "describe": "基于yolox小模型的通用检测模型",
- "hot": 7153,
- "pic": "example.jpg",
- "uuid": "cv-cspnet-image-object-detection-yolox",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-book",
- "label": "RaNER命名实体识别-中文-小说领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文Book9小说领域数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 7138,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-book",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnest101-animal-recognition",
- "label": "动物识别-中文-通用领域",
- "describe": "本模型是对含有动物的图像进行标签识别,无需任何额外输入,输出动物的类别标签,目前已经覆盖了8K多类的细粒度的动物类别。",
- "hot": 7085,
- "pic": "example.jpg",
- "uuid": "cv-resnest101-animal-recognition",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "动物识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-vitb16-classification-vision-efficient-tuning-adapter",
- "label": "基础视觉模型高效调优-Adapter",
- "describe": "",
- "hot": 7021,
- "pic": "example.jpg",
- "uuid": "cv-vitb16-classification-vision-efficient-tuning-adapter",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "基础模型调优",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "multi-modal-clip-vit-base-patch16-zh",
- "label": "CLIP模型-中文-通用领域-base",
- "describe": "本项目为CLIP模型的中文版本,使用大规模中文数据进行训练(~2亿图文对),旨在帮助用户实现中文领域的跨模态检索、图像表示等。视觉encoder采用vit结构,文本encoder采用roberta结构。 模型在多个中文图文检索数据集上进行了效果测试。",
- "hot": 6890,
- "pic": "example.jpg",
- "uuid": "multi-modal-clip-vit-base-patch16-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "多模态表征",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-vit-object-detection-coco",
- "label": "VitDet图像目标检测",
- "describe": "输入一张图片,输出图像中较通用目标(COCO-80类范围)的位置及类别。",
- "hot": 6889,
- "pic": "example.jpg",
- "uuid": "cv-vit-object-detection-coco",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-passage-ranking-chinese-base-medical",
- "label": "CoROM语义相关性-中文-医疗领域-base",
- "describe": "基于CoROM-Base预训练模型的医疗领域中文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 6842,
- "pic": "example.jpg",
- "uuid": "nlp-corom-passage-ranking-chinese-base-medical",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-ru-16k-common-vocab1664-tensorflow1-offline",
- "label": "UniASR语音识别-俄语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 6839,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-ru-16k-common-vocab1664-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-dla34-table-structure-recognition-cycle-centernet",
- "label": "读光-表格结构识别-有线表格",
- "describe": "有线表格结构识别,输入图像,检测出单元格bbox并将其拼接起来得到精准而完整的表格。",
- "hot": 6842,
- "pic": "example.jpg",
- "uuid": "cv-dla34-table-structure-recognition-cycle-centernet",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "表格结构识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-dfsmn-kws-char-farfield-16k-nihaomiya",
- "label": "FSMN远场唤醒-双麦-16k-你好米雅",
- "describe": "远场唤醒模型,输入为双麦克风阵列的双通道音频加一路音箱播放的参考音频,适用于智能音箱、故事机等智能设备场景。此demo使用开源数据训练,唤醒词为“你好米雅”,用户可使用我们提供的训练套件基于自有数据训练新唤醒词。",
- "hot": 6777,
- "pic": "example.jpg",
- "uuid": "speech-dfsmn-kws-char-farfield-16k-nihaomiya",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音唤醒",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-cn-dialect-16k-vocab8358-tensorflow1-offline",
- "label": "UniASR语音识别-中文方言-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 6690,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-cn-dialect-16k-vocab8358-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-ecom",
- "label": "RaNER命名实体识别-英语-电商领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英语电商query和商品标题数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 6611,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-manual-face-detection-mtcnn",
- "label": "Mtcnn人脸检测关键点模型",
- "describe": "给定一张图片,返回图片中人脸区域的位置和五点关键点。MTCNN是工业界广泛应用的检测关键点二合一模型。",
- "hot": 6559,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-detection-mtcnn",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-visual-question-answering-pretrain-huge-en",
- "label": "OFA视觉问答模型-英文-通用领域-huge",
- "describe": "视觉问答任务:给定一张图片和一个关于图片的问题,要求模型正确作答。",
- "hot": 6504,
- "pic": "example.jpg",
- "uuid": "ofa-visual-question-answering-pretrain-huge-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉问答",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-paraformer-large-contextual-asr-nat-zh-cn-16k-common-vocab8404",
- "label": "Paraformer语音识别-中文-通用-16k-离线-large-热词版",
- "describe": "基于Paraformer-large的热词版本模型,可实现对热词的定制化,基于提供的热词列表对热词进行激励增强,提升模型对热词的召回",
- "hot": 6456,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-large-contextual-asr-nat-zh-cn-16k-common-vocab8404",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-vit-b32-retrieval-vop",
- "label": "VoP: 通用跨模态视频检索模型",
- "describe": "VoP是基于CLIP的快速跨模态检索微调框架,可以适用于任何需要做视频文本跨模态检索的“视频-文本对”数据当中。",
- "hot": 6377,
- "pic": "example.jpg",
- "uuid": "cv-vit-b32-retrieval-vop",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频文本跨模态检索",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-pathshift-action-recognition",
- "label": "PST动作识别模型-tiny",
- "describe": "Patch Shift Transformer(PST)是把2D Transformer 模型在不增加参数量的情况下转换成适应视频多帧输入的动作识别模型",
- "hot": 6372,
- "pic": "example.png",
- "uuid": "cv-pathshift-action-recognition",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "动作识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-mossformer-separation-temporal-8k",
- "label": "MossFormer语音分离-单麦-8k",
- "describe": "基于MossFormer的语音分离模型,可以把混杂在一起的两人语音分离开来,输入为一路混合音频,输出为两路分离后的音频,格式均为8000Hz单通道。",
- "hot": 6355,
- "pic": "example.jpg",
- "uuid": "speech-mossformer-separation-temporal-8k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音分离",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-yolopv2-image-driving-perception-bdd100k",
- "label": "YOLOPV2车辆检测车道线分割-自动驾驶领域",
- "describe": "YOLOPv2 适用于自动驾驶场景下的实时全景驾驶感知, 同时执行三种不同的任务,分别为车辆检测,可行驶区域分割以及车道线分割。",
- "hot": 6343,
- "pic": "example.jpeg",
- "uuid": "cv-yolopv2-image-driving-perception-bdd100k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像驾驶感知",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-nextvit-small-image-classification-dailylife-labels",
- "label": "NextViT实时图像分类-中文-日常物品",
- "describe": "采用基于Transformer的第一个实现工业TensorRT实时落地的Next-ViT模型结构,对自建1300类常见物体标签体系进行分类。",
- "hot": 6263,
- "pic": "example.jpg",
- "uuid": "cv-nextvit-small-image-classification-dailylife-labels",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用分类",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-large-asr-2pass-zh-cn-16k-common-vocab8358-tensorflow1-offline",
- "label": "UniASR语音识别-中文-通用-16k-离线-large",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 6248,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-large-asr-2pass-zh-cn-16k-common-vocab8358-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-swinl-semantic-segmentation-cocopanmerge",
- "label": "Mask2Former-SwinL语义分割",
- "describe": "基于Mask2Former架构,SwinL为backbone的语义分割模型",
- "hot": 6225,
- "pic": "example.jpg",
- "uuid": "cv-swinl-semantic-segmentation-cocopanmerge",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-cspnet-video-object-detection-streamyolo",
- "label": "StreamYOLO实时视频目标检测-自动驾驶领域",
- "describe": "实时视频目标检测模型",
- "hot": 6181,
- "pic": "example.jpg",
- "uuid": "cv-cspnet-video-object-detection-streamyolo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-manual-face-detection-ulfd",
- "label": "ULFD人脸检测模型-tiny",
- "describe": "1M轻量级人脸检测模型。给定一张图片,返回图片中人脸位置的坐标。ULFD为轻量级人脸检测算法, 基于SSD框架手工设计了backbone结构,是业界开源的第一个1M人脸检测模型。当输入320x240分辨率的图片且未使用onnxruntime加速时,在CPU上跑需要50-60ms,当使用onnxruntime加速后,在CPU上仅需要8-11ms。",
- "hot": 6171,
- "pic": "example.jpeg",
- "uuid": "cv-manual-face-detection-ulfd",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-ocr-recognition-document-base-zh",
- "label": "OFA文字识别-中文-印刷体-base",
- "describe": "基于OFA模型的finetune后的OCR文字识别任务,可有效识别印刷体文字。",
- "hot": 6118,
- "pic": "example.jpg",
- "uuid": "ofa-ocr-recognition-document-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-en-16k-common-vocab1080-tensorflow1-offline",
- "label": "UniASR语音识别-英语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 6101,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-en-16k-common-vocab1080-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-aishell1-pytorch",
- "label": "Paraformer语音识别-中文-aishell1-16k-离线",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 6068,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-aishell1-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-resnet18-ocr-detection-db-line-level-damo",
- "label": "读光-文字检测-DBNet行检测模型-中英-通用领域",
- "describe": "给定一张图片,检测出图中所含文字的外接框的端点的坐标值。",
- "hot": 5952,
- "pic": "example.jpg",
- "uuid": "cv-resnet18-ocr-detection-db-line-level-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-andy-en-us-16k",
- "label": "语音合成-美式英文-通用领域-16k-发音人Andy",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 5882,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-andy-en-us-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-lstm-named-entity-recognition-chinese-generic",
- "label": "LSTM命名实体识别-中文-通用领域",
- "describe": "本方法采用char-BiLSTM-CRF模型",
- "hot": 5841,
- "pic": "example.jpg",
- "uuid": "nlp-lstm-named-entity-recognition-chinese-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-canonical-body-3d-keypoints-video",
- "label": "人体关键点检测-通用领域-3D",
- "describe": "输入一段单人视频,实现端到端的3D人体关键点检测,输出视频中每一帧的3D人体关键点坐标。",
- "hot": 5798,
- "pic": "example.jpg",
- "uuid": "cv-canonical-body-3d-keypoints-video",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人体3D关键点",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-unet-universal-matting",
- "label": "BSHM通用抠图",
- "describe": "",
- "hot": 5818,
- "pic": "example.png",
- "uuid": "cv-unet-universal-matting",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用抠图",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-newcrfs-image-depth-estimation-indoor",
- "label": "基于神经窗口全连接CRFs的单目深度估计",
- "describe": "单目深度估计是从单张RGB图预测场景深度,是一个很具有挑战性的任务。现在做这个任务的方法大都是设计越来越复杂的网络来简单粗暴地回归深度图,但我们采取了一个更具可解释性的路子,就是使用优化方法中的条件随机场(CRFs)。由于CRFs的计算量很大,通常只会用于计算相邻节点的能量,而很难用于计算整个图模型中所有节点之间的能量。为了借助这种全连接CRFs的强大表征力,我们采取了一种折中的方法,即将整个图模型划分为一个个小窗口,在每个窗口里面进行全连接CRFs的计算,这样就可以大大减少计算量,使全连接CRFs在深度估计这一任务上成为了可能。同时,为了更好地在节点之间进行信息传递,我们利用多头注意力机制计算了多头能量函数,然后用网络将这个能量函数优化到一个精确的深度图。基于此,我们用视觉transformer作为encoder,神经窗口全连接条件随机场作为decoder,构建了一个bottom-up-top-down的网络架构,这个网络在KITTI、NYUv2上都取得了SOTA的性能,同时可以应用于全景图深度估计任务,在MatterPort3D上也取得了SOTA的性能。",
- "hot": 5789,
- "pic": "example.png",
- "uuid": "cv-newcrfs-image-depth-estimation-indoor",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像深度估计",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-faq-question-answering-chinese-base",
- "label": "StructBERT FAQ问答-中文-通用领域-base",
- "describe": "FAQ问答模型以StructBERT预训练模型-中文-base为基础,使用简单的原型网络,通过小样本meta-learning的方式在海量业务数据预训练(亿级)、微调(百万级),在多个公开数据上取得了非常好的效果,适用于FAQ问答任务和小样本分类任务;",
- "hot": 5727,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-faq-question-answering-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "FAQ问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-structbert-outbound-intention-chinese-tiny",
- "label": "StructBERT意图识别-中文-外呼-tiny",
- "describe": "本模型基于StructBERT-tiny模型,使用外呼场景下的对话意图识别数据进行微调得到的。",
- "hot": 5581,
- "pic": "example.png",
- "uuid": "nlp-structbert-outbound-intention-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "taiyi-stable-diffusion-1b-anime-chinese-v0-1",
- "label": "太乙-Stable-Diffusion-1B-动漫-中文-v0.1",
- "describe": "首个开源的中文Stable Diffusion动漫模型,基于100万筛选过的动漫中文图文对训练。",
- "hot": 5550,
- "pic": "example.jpg",
- "uuid": "taiyi-stable-diffusion-1b-anime-chinese-v0-1",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-convnext-base-image-classification-garbage",
- "label": "ConvNeXt图像分类-中文-垃圾分类",
- "describe": "自建265类常见的生活垃圾标签体系,15w张图片数据,包含可回收垃圾、厨余垃圾、有害垃圾、其他垃圾4个标准垃圾大类,覆盖常见的食品,厨房用品,家具,家电等生活垃圾,标签从海量中文互联网社区语料进行提取,整理出了频率较高的常见生活垃圾名称。模型结构采用ConvNeXt-Base结构, 经过大规模数据集ImageNet-22K预训练后,在数据集上进行微调。",
- "hot": 5363,
- "pic": "example.jpg",
- "uuid": "cv-convnext-base-image-classification-garbage",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用分类",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-ja-16k-common-vocab93-tensorflow1-online",
- "label": "UniASR语音识别-日语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 5350,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-ja-16k-common-vocab93-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-cn-dialect-16k-vocab8358-tensorflow1-online",
- "label": "UniASR语音识别-中文方言-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 5314,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-cn-dialect-16k-vocab8358-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-en2es",
- "label": "CSANMT连续语义增强机器翻译-英西-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 5284,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-en2es",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-luna-en-gb-16k",
- "label": "语音合成-英式英文-通用领域-16k-发音人Luna",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 5260,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-luna-en-gb-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-cspnet-image-object-detection-yolox-nano-coco",
- "label": "实时目标检测-通用领域-移动端",
- "describe": "通用实时检测超轻量级模型",
- "hot": 5182,
- "pic": "example.jpg",
- "uuid": "cv-cspnet-image-object-detection-yolox-nano-coco",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-image-classification-imagenet-large-en",
- "label": "OFA图像分类-数据集imagenet1k-large",
- "describe": "ImageNet-1K图片分类任务:给定一张图片,要求模型从1K个候选类别中正确给出图片分类标签。",
- "hot": 5177,
- "pic": "example.jpg",
- "uuid": "ofa-image-classification-imagenet-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用分类",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "ofa-mmspeech-asr-aishell1-large-zh",
- "label": "OFA-MMSpeech语音识别-中文-aishell1-large",
- "describe": "对比SOTA,MMSpeech字错误率降低了48.3%/42.4%,效果达到1.6%/1.9%,远超SOTA 3.1%/3.3%(benchmark为AIShell1 dev/test)。",
- "hot": 5132,
- "pic": "example.jpg",
- "uuid": "ofa-mmspeech-asr-aishell1-large-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-palm2-0-text-generation-weather-chinese-base",
- "label": "PALM 2.0天气播报模型-中文-base",
- "describe": "达摩PALM 2.0中文天气播报base模型",
- "hot": 4984,
- "pic": "example.jpg",
- "uuid": "nlp-palm2-0-text-generation-weather-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-mt5-dialogue-rewriting-chinese-base",
- "label": "MT5开放域多轮对话改写-中文-通用-base",
- "describe": "开放域多轮对话改写模型主要解决开放域对话中的指代和省略问题,输入对话上下文,输出改写后的语义完整的问题。该模型基于google/mt5-base基座在公开数据和业务数据集上finetune而得,适用于开放域对话场景。",
- "hot": 4999,
- "pic": "example.jpg",
- "uuid": "nlp-mt5-dialogue-rewriting-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "端到端文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-segformer-b0-image-semantic-segmentation-coco-stuff164k",
- "label": "Segformer-B0实时语义分割",
- "describe": "Neurips2021文章SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers在COCO_Stuff164K数据集上的复现。官方源码暂没有提供COCO_Stuff164K的相关实现。本模型基于Segformer分割框架所配置训练的实时语义分割框架。使用了一个GPU上实时运行的配置结构。在CoCo-Stuff-164的数据集上进行了172类的分类",
- "hot": 4933,
- "pic": "example.jpg",
- "uuid": "cv-segformer-b0-image-semantic-segmentation-coco-stuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet18-license-plate-detection-damo",
- "label": "读光-车牌检测-通用",
- "describe": "",
- "hot": 4927,
- "pic": "example.jpg",
- "uuid": "cv-resnet18-license-plate-detection-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "车牌检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-abuse-detect-chinese-tiny",
- "label": "StructBERT辱骂风险识别-中文-外呼-tiny",
- "describe": "本模型基于StructBERT-tiny模型,使用外呼场景下的辱骂风险识别数据集训练得到。",
- "hot": 4909,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-abuse-detect-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-bert-sentiment-analysis-english-base",
- "label": "BERT情感分析-英文-base-TweetEval数据集",
- "describe": "该模型基于bert-base-uncased,在TweetEval数据集上fine-tune得到",
- "hot": 4878,
- "pic": "example.png",
- "uuid": "nlp-bert-sentiment-analysis-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-vitb16-classification-vision-efficient-tuning-prompt",
- "label": "基础视觉模型高效调优-Prompt",
- "describe": "",
- "hot": 4831,
- "pic": "example.jpg",
- "uuid": "cv-vitb16-classification-vision-efficient-tuning-prompt",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "基础模型调优",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-unet-person-image-cartoon-sd-design-compound-models",
- "label": "DCT-Net人像卡通化-扩散模型-插画",
- "describe": "该模型采用全新的DCT-Net(Domain-Calibrated Translation) 域校准图像翻译模型,结合Stable-Diffusion模型进行小样本风格数据生成,从而训练得到高保真、强鲁棒、易拓展的人像风格转换模型。",
- "hot": 4818,
- "pic": "example.jpg",
- "uuid": "cv-unet-person-image-cartoon-sd-design-compound-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像卡通化",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-tinynas-detection",
- "label": "AIRDet-高性能检测模型-S",
- "describe": "AIRDet高性能检测模型是基于Tiny-NAS技术设计出的卷积神经网络,具有精度高、速度快的特点,这里只提供S模型。借助Tiny-NAS技术能够实现针对硬件算力的最优模型搜索能力,更多模型请参考README获取。",
- "hot": 4772,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-detection",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-effnetv2-video-human-matting",
- "label": "视频人像抠图模型-通用领域",
- "describe": "输入一段视频,返回视频中人像的alpha序列",
- "hot": 4742,
- "pic": "example.jpg",
- "uuid": "cv-effnetv2-video-human-matting",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频人像抠图",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-resume",
- "label": "RaNER命名实体识别-中文-简历领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文Resume数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 4723,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-resume",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-vitl16-segmentation-text-driven-seg",
- "label": "文本指导的图像分割-英文-通用领域",
- "describe": "vitl backbone,输入英文文本描述和图像,根据英文描述对图像进行语义分割",
- "hot": 4693,
- "pic": "example.jpg",
- "uuid": "cv-vitl16-segmentation-text-driven-seg",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本指导的图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-manual-face-quality-assessment-fqa",
- "label": "人脸质量模型FQA",
- "describe": "人脸质量模型FQA",
- "hot": 4621,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-quality-assessment-fqa",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸质量评估",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-realbasicvsr-video-super-resolution-videolq",
- "label": "RealBasicVSR视频超分辨率",
- "describe": "长期传播会在轻度退化的情况下提高性能,但严重的退化可能会通过传播被放大,从而损害输出视频的质量。为了平衡细节合成和退化抑制之间的权衡,RealBasicVSR引入了图像预清理模块,在质量和效率上都优于现有方法。",
- "hot": 4583,
- "pic": "example.png",
- "uuid": "cv-realbasicvsr-video-super-resolution-videolq",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频超分辨率",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-dut-raft-video-stabilization-base",
- "label": "DUT-RAFT视频稳像",
- "describe": "",
- "hot": 4571,
- "pic": "example.gif",
- "uuid": "cv-dut-raft-video-stabilization-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频稳像",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-hrnetocr-skychange",
- "label": "图像天空替换模型",
- "describe": "",
- "hot": 4516,
- "pic": "example.jpg",
- "uuid": "cv-hrnetocr-skychange",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "天空替换",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab3445-pytorch-online",
- "label": "UniASR语音识别-中文-通用-8k-实时-pytorch",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 4510,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab3445-pytorch-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-zh-cn-8k-common-vocab8358-tensorflow1",
- "label": "Paraformer语音识别-中文-通用-8k-离线",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 4401,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-zh-cn-8k-common-vocab8358-tensorflow1",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "ofa-ocr-recognition-web-base-zh",
- "label": "OFA文字识别-中文-网络场景-base",
- "describe": "基于OFA模型的finetune后的OCR文字识别任务,可有效识别网络场景的文字内容。",
- "hot": 4386,
- "pic": "example.jpg",
- "uuid": "ofa-ocr-recognition-web-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-ru-16k-common-vocab1664-tensorflow1-online",
- "label": "UniASR语音识别-俄语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 4368,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-ru-16k-common-vocab1664-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-lstm-named-entity-recognition-chinese-resume",
- "label": "LSTM命名实体识别-中文-简历领域",
- "describe": "本方法采用char-BiLSTM-CRF模型",
- "hot": 4325,
- "pic": "example.jpg",
- "uuid": "nlp-lstm-named-entity-recognition-chinese-resume",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-hrnetw48-human-wholebody-keypoint-image",
- "label": "全身关键点检测-通用领域-2D",
- "describe": "输入一张人物图像,端到端检测全身133点关键点,输出人体框和对应的全身关键点,包含68个人脸关键点、42个手势关键点、17个骨骼关键点和6个脚步关键点。",
- "hot": 4316,
- "pic": "example.jpg",
- "uuid": "cv-hrnetw48-human-wholebody-keypoint-image",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "全身关键点检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-palm2-0-text-generation-english-base",
- "label": "PALM 2.0摘要生成模型-英文-base",
- "describe": "本任务是PALM通用预训练生成模型,在英文CNN/Dail Mail和中文LCSTS上进行finetune的文本摘要生成下游任务。",
- "hot": 4242,
- "pic": "example.jpg",
- "uuid": "nlp-palm2-0-text-generation-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "mplug-visual-question-answering-coco-base-zh",
- "label": "mPLUG视觉问答模型-中文-base",
- "describe": "本任务是mPLUG在中文VQA进行finetune的视觉问答下游任务,给定一个问题和图片,通过图片信息来给出答案。",
- "hot": 4236,
- "pic": "example.jpg",
- "uuid": "mplug-visual-question-answering-coco-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉问答",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-segformer-b5-image-semantic-segmentation-coco-stuff164k",
- "label": "Segformer-B5实时语义分割",
- "describe": "Neurips2021文章SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers在COCO_Stuff164K数据集上的复现。官方源码暂没有提供COCO_Stuff164K的相关实现。本模型基于Segformer分割框架所配置训练的实时语义分割框架。使用了一个GPU上运行精度最高的配置结构。在CoCo-Stuff-164的数据集上进行了172类的分类",
- "hot": 4232,
- "pic": "example.jpg",
- "uuid": "cv-segformer-b5-image-semantic-segmentation-coco-stuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-social-media",
- "label": "RaNER命名实体识别-中文-社交媒体领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文Weibo数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 4202,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-social-media",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "taiyi-stable-diffusion-1b-chinese-v0-1",
- "label": "太乙-Stable-Diffusion-1B-中文-v0.1",
- "describe": "首个开源的中文Stable Diffusion模型,基于0.2亿筛选过的中文图文对训练。",
- "hot": 4183,
- "pic": "example.jpg",
- "uuid": "taiyi-stable-diffusion-1b-chinese-v0-1",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-unet-person-image-cartoon-sd-illustration-compound-models",
- "label": "DCT-Net人像卡通化-扩散模型-漫画",
- "describe": "",
- "hot": 4180,
- "pic": "example.jpg",
- "uuid": "cv-unet-person-image-cartoon-sd-illustration-compound-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像卡通化",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-yolox-image-object-detection-auto",
- "label": "实时目标检测-自动驾驶领域",
- "describe": "检测自动驾驶场景图片的目标,支持车辆检测。",
- "hot": 4173,
- "pic": "example.jpeg",
- "uuid": "cv-yolox-image-object-detection-auto",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-visual-grounding-refcoco-distilled-en",
- "label": "OFA通过描述定位图像物体-英文-通用领域-蒸馏33M",
- "describe": "视觉定位任务:给定一张图片,一段描述,通过描述找到图片对应的物体。",
- "hot": 4150,
- "pic": "example.jpg",
- "uuid": "ofa-visual-grounding-refcoco-distilled-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉定位",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-xvector-sv-zh-cn-cnceleb-16k-spk3465-pytorch",
- "label": "xvector说话人确认-中文-cnceleb-16k-离线-pytorch",
- "describe": "该模型是使用CN-Celeb 1&2以及AliMeeting数据集预训练得到的说话人嵌入码(speaker embedding)提取模型。可以直接用于通用和会议场景的说话人确认和说话人日志等任务。在CN-Celeb语音测试集上EER为9.00%,在AliMeeting测试集上的EER为1.45%。",
- "hot": 4148,
- "pic": "example.jpg",
- "uuid": "speech-xvector-sv-zh-cn-cnceleb-16k-spk3465-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "说话人确认",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-id-16k-common-vocab1067-tensorflow1-online",
- "label": "UniASR语音识别-印尼语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 4145,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-id-16k-common-vocab1067-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-part-of-speech-chinese-base",
- "label": "BAStructBERT词性标注-中文-新闻领域-base",
- "describe": "基于预训练语言模型的新闻领域中文词性标注模型,根据用户输入的中文句子产出词性标注结果。",
- "hot": 4099,
- "pic": "example.png",
- "uuid": "nlp-structbert-part-of-speech-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "词性标注",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-ko-16k-common-vocab6400-tensorflow1-online",
- "label": "UniASR语音识别-韩语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 4090,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-ko-16k-common-vocab6400-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-id-16k-common-vocab1067-tensorflow1-offline",
- "label": "UniASR语音识别-印尼语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 4089,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-id-16k-common-vocab1067-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "erlangshen-roberta-110m-sentiment",
- "label": "二郎神-RoBERTa-110M-情感分类",
- "describe": "",
- "hot": 4075,
- "pic": "example.jpg",
- "uuid": "erlangshen-roberta-110m-sentiment",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet18-ocr-detection-word-level-damo",
- "label": "读光-文字检测-单词检测模型-英文-通用领域",
- "describe": "本模型是以自底向上的方式,先检测文本块和文字行之间的吸引排斥关系,然后对文本块聚类成行,最终输出单词的外接框的坐标值。",
- "hot": 4058,
- "pic": "example.jpg",
- "uuid": "cv-resnet18-ocr-detection-word-level-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab8358-tensorflow1-offline",
- "label": "UniASR语音识别-中文-通用-8k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 4051,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab8358-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-facemask",
- "label": "实时口罩检测-通用",
- "describe": "本模型为高性能热门应用系列检测模型中的实时口罩检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有人脸的坐标信息,以及是否佩戴口罩。更多具体信息请参考Model card。",
- "hot": 4037,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-facemask",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-generic",
- "label": "RaNER命名实体识别-英语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在MultiCoNER领域数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 4010,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-chuangirl-sichuan-16k",
- "label": "语音合成-四川话-通用领域-16k-发音人chuangirl",
- "describe": "四川话语音合成女声16k模型,本模型使用Sambert-hifigan网络结构。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 4003,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-chuangirl-sichuan-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab8358-tensorflow1-online",
- "label": "UniASR语音识别-中文-通用-8k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 3996,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab8358-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-multilingual-large-generic",
- "label": "RaNER命名实体识别-多语言统一-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在多语言数据集MultiCoNER-MULTI-Multilingual训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3985,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-multilingual-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet50-bert-video-scene-segmentation-movienet",
- "label": "BaSSL视频场景分割-长视频领域",
- "describe": "针对长视频进行场景分割,也可按照镜头进行分割,有助于进行视频拆条和视频理解等。该模型支持分割结果的本地保存,同时可以支持微调操作。",
- "hot": 3942,
- "pic": "example.jpg",
- "uuid": "cv-resnet50-bert-video-scene-segmentation-movienet",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频场景分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-cantonese-chs-16k-common-vocab1468-tensorflow1-online",
- "label": "UniASR语音识别-粤语简体-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 3938,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-cantonese-chs-16k-common-vocab1468-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-annie-en-us-16k",
- "label": "语音合成-美式英文-通用领域-16k-发音人Annie",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 3936,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-annie-en-us-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-ptts-autolabel-16k",
- "label": "个性化语音合成-自动标注模型-16k",
- "describe": "用于训练个性化语音合成模型的自动标注工具依赖的模型资源",
- "hot": 3884,
- "pic": "example.jpeg",
- "uuid": "speech-ptts-autolabel-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-bank",
- "label": "RaNER命名实体识别-中文-银行领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文Bank数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3883,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-bank",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentiment-classification-chinese-tiny",
- "label": "StructBERT情感分类-中文-通用-tiny",
- "describe": "StructBERT情感分类-中文-通用-tiny是基于bdci、dianping、jd binary、waimai-10k四个数据集(11.5w条数据)训练出来的情感分类模型。",
- "hot": 3844,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentiment-classification-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-luca-en-gb-16k",
- "label": "语音合成-英式英文-通用领域-16k-发音人Luca",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 3844,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-luca-en-gb-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-news",
- "label": "RaNER命名实体识别-英文-新闻领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文conll03/conllpp领域数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3842,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-news",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-hrnetw18-hand-pose-keypoints-coco-wholebody",
- "label": "手部关键点检测-通用领域-2D",
- "describe": "该模型采用自顶向下的Heatmap手部关键点检测框架,通过端对端的快速推理可以得到图像中的全部手部关键点。",
- "hot": 3821,
- "pic": "example.jpg",
- "uuid": "cv-hrnetw18-hand-pose-keypoints-coco-wholebody",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "手部2D关键点",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "taiyi-stable-diffusion-1b-chinese-en-v0-1",
- "label": "太乙-Stable-Diffusion-1B-中英双语-v0.1",
- "describe": "首个开源的中英双语Stable Diffusion模型,基于0.2亿筛选过的中文图文对训练。",
- "hot": 3796,
- "pic": "example.jpg",
- "uuid": "taiyi-stable-diffusion-1b-chinese-en-v0-1",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-game",
- "label": "RaNER命名实体识别-中文-游戏领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文游戏领域数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3793,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-game",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-en-us-16k",
- "label": "语音合成-美式英文-通用领域-16k-多发音人",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 3780,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-en-us-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-en-16k-common-vocab1080-tensorflow1-online",
- "label": "UniASR语音识别-英语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 3759,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-en-16k-common-vocab1080-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-literature",
- "label": "RaNER命名实体识别-中文-文学领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在中文Literature数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3751,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-literature",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-manual-facial-landmark-confidence-flcm",
- "label": "FLCM人脸关键点置信度模型",
- "describe": "",
- "hot": 3749,
- "pic": "example.jpg",
- "uuid": "cv-manual-facial-landmark-confidence-flcm",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸2D关键点",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-lstm-named-entity-recognition-chinese-news",
- "label": "LSTM命名实体识别-中文-新闻领域",
- "describe": "本方法采用char-BiLSTM-CRF模型",
- "hot": 3717,
- "pic": "example.jpg",
- "uuid": "nlp-lstm-named-entity-recognition-chinese-news",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-zh-cn-16k-common-vocab3444-tensorflow1-online",
- "label": "Paraformer语音识别-中文-通用-16k-实时",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 3682,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-zh-cn-16k-common-vocab3444-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-nafnet-image-deblur-gopro",
- "label": "NAFNet图像去模糊",
- "describe": "NAFNet(Nonlinear Activation Free Network)提出了一个简单的基线,计算效率高。其不需要使用非线性激活函数(Sigmoid、ReLU、GELU、Softmax等),可以达到SOTA性能。",
- "hot": 3667,
- "pic": "example.jpg",
- "uuid": "cv-nafnet-image-deblur-gopro",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像去模糊",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "erlangshen-roberta-110m-similarity",
- "label": "二郎神-RoBERTa-110M-文本相似度",
- "describe": "",
- "hot": 3642,
- "pic": "example.jpg",
- "uuid": "erlangshen-roberta-110m-similarity",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-structbert-part-of-speech-chinese-lite",
- "label": "BAStructBERT词性标注-中文-新闻领域-lite",
- "describe": "基于预训练语言模型的新闻领域中文词性标注模型,根据用户输入的中文句子产出词性标注结果。",
- "hot": 3634,
- "pic": "example.png",
- "uuid": "nlp-structbert-part-of-speech-chinese-lite",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "词性标注",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-ir101-facerecognition-cfglint",
- "label": "CurricularFace人脸识别模型",
- "describe": "输入一张图片,检测矫正人脸区域后提取特征,两个人脸特征可用于人脸比对,多个人脸特征可用于人脸检索。",
- "hot": 3522,
- "pic": "example.jpg",
- "uuid": "cv-ir101-facerecognition-cfglint",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-clip-it-video-summarization-language-guided-en",
- "label": "CLIP_It自然语言引导的视频摘要-Web视频领域-英文",
- "describe": "自然语言引导的视频摘要,用户根据自己的需求输入一段自然语言和一个长视频,算法根据用户输入自然语言的内容对输入视频进行自适应的视频摘要。",
- "hot": 3479,
- "pic": "example.jpg",
- "uuid": "cv-clip-it-video-summarization-language-guided-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本指导的视频摘要",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "mgeo-geographic-textual-similarity-rerank-chinese-base",
- "label": "MGeo地址QueryPOI相关性排序-中文-地址领域-base",
- "describe": "模型对用户输入的地址query以及候选POI列表(包括每个POI包括POI的地址描述以及POI位置)进行相关性排序。",
- "hot": 3453,
- "pic": "example.jpg",
- "uuid": "mgeo-geographic-textual-similarity-rerank-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-chinese-base-finance",
- "label": "RaNER命名实体识别-中文-金融领域-base",
- "describe": "该模型是基于检索增强(RaNer)方法在CCKS2021中文金融案件要素抽取数据训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3436,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-chinese-base-finance",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-lstmcrf-part-of-speech-chinese-news",
- "label": "LSTM词性标注-中文-新闻领域",
- "describe": "",
- "hot": 3386,
- "pic": "example.jpg",
- "uuid": "nlp-lstmcrf-part-of-speech-chinese-news",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "词性标注",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-t",
- "label": "DAMOYOLO-高性能通用检测模型-T",
- "describe": "DAMOYOLO是一款面向工业落地的高性能检测框架,精度和速度超越当前的一众典型YOLO框架(YOLOE、YOLOv6、YOLOv7)。基于TinyNAS技术,DAMOYOLO能够针对不同的硬件算力,进行低成本的模型定制化搜索。这里仅提供DAMOYOLO-T模型,更多模型请参考README。",
- "hot": 3379,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-t",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-masts-sentence-similarity-clue-chinese-large",
- "label": "MaSTS文本相似度-中文-搜索-CLUE语义匹配-large",
- "describe": "MaSTS中文文本相似度-CLUE语义匹配模型是在MaSTS预训练模型-CLUE语义匹配的基础上,在QBQTC数据集上训练出来的相似度匹配模型。在CLUE语义匹配榜上通过集成此模型获得了第一名的成绩。",
- "hot": 3341,
- "pic": "example.jpg",
- "uuid": "nlp-masts-sentence-similarity-clue-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-music",
- "label": "RaNER命名实体识别-英文-音乐领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文Music数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3324,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-music",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-segformer-b4-image-semantic-segmentation-coco-stuff164k",
- "label": "Segformer-B4实时语义分割",
- "describe": "Neurips2021文章SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers在COCO_Stuff164K数据集上的复现。官方源码暂没有提供COCO_Stuff164K的相关实现。本模型基于Segformer分割框架所配置训练的语义分割框架。网络模型结构的具体超参数为论文中的B4配置。在CoCo-Stuff-164的数据集上进行了172类的分类",
- "hot": 3316,
- "pic": "example.jpg",
- "uuid": "cv-segformer-b4-image-semantic-segmentation-coco-stuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "mgeo-geographic-composition-analysis-chinese-base",
- "label": "MGeo地址Query成分分析要素识别-中文-地址领域-base",
- "describe": "模型用于识别地址query中的区划、路网、POI、户室号、公交地铁、品牌商圈等元素。",
- "hot": 3299,
- "pic": "example.jpg",
- "uuid": "mgeo-geographic-composition-analysis-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "序列标注",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-lstm-named-entity-recognition-chinese-social-media",
- "label": "LSTM命名实体识别-中文-社交媒体领域",
- "describe": "本方法采用char-BiLSTM-CRF模型。",
- "hot": 3276,
- "pic": "example.jpg",
- "uuid": "nlp-lstm-named-entity-recognition-chinese-social-media",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-turkish-large-generic",
- "label": "RaNER命名实体识别-土耳其语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在土耳其语数据集MultiCoNER-TR-Turkish训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3248,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-turkish-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-cigarette",
- "label": "实时香烟检测-通用",
- "describe": "本模型为高性能热门应用系列检测模型中的实时香烟检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有香烟的坐标信息,并可用于吸烟检测等后续应用场景。更多具体信息请参考Model card。",
- "hot": 3245,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-cigarette",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-ai",
- "label": "RaNER命名实体识别-英文-人工智能领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文AI数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3228,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-ai",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-russian-large-generic",
- "label": "RaNER命名实体识别-俄语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在俄语数据集MultiCoNER-RU-Russian训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3223,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-russian-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "mgeo-geographic-where-what-cut-chinese-base",
- "label": "MGeo地址地点WhereWhat切分-中文-地址领域-base",
- "describe": "模型提供将一条地址切分为门址+POI描述的功能。当一条地址包含多个地点描述时,通常需要对其进行切分,将原始地址切为where和what两部分。",
- "hot": 3168,
- "pic": "example.jpg",
- "uuid": "mgeo-geographic-where-what-cut-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "序列标注",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-xiaoda-wuushanghai-16k",
- "label": "语音合成-上海话-通用领域-16k-发音人xiaoda",
- "describe": "上海话语音合成女声16k模型,本模型使用Sambert-hifigan网络结构。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 3154,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-xiaoda-wuushanghai-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-gpen-image-portrait-enhancement-hires",
- "label": "GPEN人像增强修复-大分辨率人脸",
- "describe": "GPEN通过将预训练的人像生成网络嵌入到Unet网络中联合微调的方式在人像修复任务的多项指标中上达到了sota的结果。",
- "hot": 3114,
- "pic": "example.jpg",
- "uuid": "cv-gpen-image-portrait-enhancement-hires",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人像增强",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-dutch-large-generic",
- "label": "RaNER命名实体识别-荷兰语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在荷兰语数据集MultiCoNER-NL-Dutch训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 3089,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-dutch-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-es-16k-common-vocab3445-tensorflow1-online",
- "label": "UniASR语音识别-西班牙语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 3079,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-es-16k-common-vocab3445-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-ko-16k-common-vocab6400-tensorflow1-offline",
- "label": "UniASR语音识别-韩语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 3077,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-ko-16k-common-vocab6400-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-yolox-pai-hand-detection",
- "label": "YOLOX-PAI手部检测模型",
- "describe": "输入一张图像,并对其中手部区域进行检测,输出所有手部区域检测框、置信度和标签。",
- "hot": 3056,
- "pic": "example.jpeg",
- "uuid": "cv-yolox-pai-hand-detection",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-cantonese-chs-16k-common-vocab1468-tensorflow1-offline",
- "label": "UniASR语音识别-粤语简体-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 3042,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-cantonese-chs-16k-common-vocab1468-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-korean-large-generic",
- "label": "RaNER命名实体识别-韩语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在韩语数据集MultiCoNER-KO-Korean训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2987,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-korean-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-farsi-large-generic",
- "label": "RaNER命名实体识别-波斯语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在波斯语数据集MultiCoNER-FA-Farsi训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2944,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-farsi-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "mplug-image-captioning-coco-base-zh",
- "label": "mPLUG图像描述模型-中文-base",
- "describe": "mPLUG中文图像描述base模型",
- "hot": 2935,
- "pic": "example.png",
- "uuid": "mplug-image-captioning-coco-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-segformer-b2-image-semantic-segmentation-coco-stuff164k",
- "label": "Segformer-B2实时语义分割",
- "describe": "Neurips2021文章SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers在COCO_Stuff164K数据集上的复现。官方源码暂没有提供COCO_Stuff164K的相关实现。本模型基于Segformer分割框架所配置训练的语义分割框架。网络模型结构的具体超参数为论文中的B2配置。在CoCo-Stuff-164的数据集上进行了172类的分类",
- "hot": 2878,
- "pic": "example.jpg",
- "uuid": "cv-segformer-b2-image-semantic-segmentation-coco-stuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-fsmn-vad-zh-cn-8k-common",
- "label": "FSMN语音端点检测-中文-通用-8k",
- "describe": "FSMN-Monophone VAD模型,可用于检测长语音片段中有效语音的起止时间点。",
- "hot": 2870,
- "pic": "example.png",
- "uuid": "speech-fsmn-vad-zh-cn-8k-common",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音端点检测",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-manual-face-detection-tinymog",
- "label": "TinyMog人脸检测器-tiny",
- "describe": "",
- "hot": 2864,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-detection-tinymog",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-mmspeech-asr-aishell1-base-zh",
- "label": "OFA-MMSpeech语音识别-中文-aishell1-base",
- "describe": "对比SOTA,MMSpeech字错误率降低了48.3%/42.4%,效果达到1.6%/1.9%,远超SOTA 3.1%/3.3%(benchmark为AIShell1 dev/test)。",
- "hot": 2820,
- "pic": "example.jpg",
- "uuid": "ofa-mmspeech-asr-aishell1-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-raft-video-frame-interpolation",
- "label": "VFI-RAFT视频插帧",
- "describe": "给定一段低帧率视频,模型通过对帧间的光流和运动估计生成中间帧,最终输出一段高帧率视频,从而提升视频的流畅度。",
- "hot": 2806,
- "pic": "example.png",
- "uuid": "cv-raft-video-frame-interpolation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频插帧",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-german-large-generic",
- "label": "RaNER命名实体识别-德语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在德语数据集MultiCoNER-DE-German训练的模型。本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2793,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-german-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-segformer-b3-image-semantic-segmentation-coco-stuff164k",
- "label": "Segformer-B3实时语义分割",
- "describe": "Neurips2021文章SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers在COCO_Stuff164K数据集上的复现。官方源码暂没有提供COCO_Stuff164K的相关实现。本模型基于Segformer分割框架所配置训练的语义分割框架。网络模型结构的具体超参数为论文中的B3配置。在CoCo-Stuff-164的数据集上进行了172类的分类",
- "hot": 2770,
- "pic": "example.jpg",
- "uuid": "cv-segformer-b3-image-semantic-segmentation-coco-stuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-hindi-large-generic",
- "label": "RaNER命名实体识别-印地语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在MultiCoNER领域数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2769,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-hindi-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-res2net-camouflaged-detection",
- "label": "图像伪装色目标检测",
- "describe": "给定一张输入图像,输出视觉显著注意力程度图(归一化至0~255)。",
- "hot": 2766,
- "pic": "example.jpg",
- "uuid": "cv-res2net-camouflaged-detection",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-fa-16k-common-vocab1257-pytorch-offline",
- "label": "UniASR语音识别-波斯语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 2752,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-fa-16k-common-vocab1257-pytorch-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-en-gb-16k",
- "label": "语音合成-英式英文-通用领域-16k-多发音人",
- "describe": "本模型是一种应用于参数TTS系统的后端声学模型及声码器模型。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 2746,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-en-gb-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-social-media",
- "label": "RaNER命名实体识别-英文-社交媒体领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文wnut17领域数据集训练的模型。本方法采用Transformer-CRF模型,使用StructBERT作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2715,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-social-media",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-wiki",
- "label": "RaNER命名实体识别-英语-wiki领域-large",
- "describe": "本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2680,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-wiki",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-es-16k-common-vocab3445-tensorflow1-offline",
- "label": "UniASR语音识别-西班牙语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 2673,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-es-16k-common-vocab3445-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-spanish-large-generic",
- "label": "RaNER命名实体识别-西班牙语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在西班牙语数据集MultiCoNER-ES-Spanish训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2657,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-spanish-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-en2ru-base",
- "label": "CSANMT连续语义增强机器翻译-英俄-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 2642,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-en2ru-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-jiajia-cantonese-16k",
- "label": "语音合成-广东粤语-通用领域-16k-发音人jiajia",
- "describe": "广东粤语语音合成女声16k模型,本模型使用Sambert-hifigan网络结构。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 2629,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-jiajia-cantonese-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-vitb16-classification-vision-efficient-tuning-lora",
- "label": "基础视觉模型高效调优-LoRA",
- "describe": "",
- "hot": 2605,
- "pic": "example.jpg",
- "uuid": "cv-vitb16-classification-vision-efficient-tuning-lora",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "基础模型调优",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-literature",
- "label": "RaNER命名实体识别-英文-文学领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文Literature数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2602,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-literature",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-zh-cn-8k-common-vocab3444-tensorflow1-online",
- "label": "Paraformer语音识别-中文-通用-8k-实时",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 2600,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-zh-cn-8k-common-vocab3444-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-politics",
- "label": "RaNER命名实体识别-英文-政治领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文Politics数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2547,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-politics",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-bangla-large-generic",
- "label": "RaNER命名实体识别-孟加拉语-通用领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在孟加拉语数据集MultiCoNER-BN-Bangla训练的模型。 本方法采用Transformer-CRF模型,使用XLM-RoBERTa作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2506,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-bangla-large-generic",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-segformer-b1-image-semantic-segmentation-coco-stuff164k",
- "label": "Segformer-B1实时语义分割",
- "describe": "Neurips2021文章SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers在COCO_Stuff164K数据集上的复现。官方源码暂没有提供COCO_Stuff164K的相关实现。本模型基于Segformer分割框架所配置训练的实时语义分割框架。网络模型结构的具体超参数为论文中的B1配置。在CoCo-Stuff-164的数据集上进行了172类的分类",
- "hot": 2501,
- "pic": "example.jpg",
- "uuid": "cv-segformer-b1-image-semantic-segmentation-coco-stuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-mmspeech-pretrain-base-zh",
- "label": "OFA-MMSpeech语音识别预训练-中文-通用领域-base",
- "describe": "对比SOTA,MMSpeech字错误率降低了48.3%/42.4%,效果达到1.6%/1.9%,远超SOTA 3.1%/3.3%(benchmark为AIShell1 dev/test)。",
- "hot": 2442,
- "pic": "example.jpg",
- "uuid": "ofa-mmspeech-pretrain-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-zh-cn-16k-aishell1-vocab4234-pytorch",
- "label": "Paraformer语音识别-中文-aishell1-16k-离线-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 2431,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-zh-cn-16k-aishell1-vocab4234-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-diffusion-text-to-image-synthesis-tiny",
- "label": "文本到图像生成扩散模型-中英文-通用领域-tiny",
- "describe": "文本到图像生成扩散模型-中英文-通用领域-tiny",
- "hot": 2429,
- "pic": "example.jpg",
- "uuid": "cv-diffusion-text-to-image-synthesis-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet34-face-attribute-recognition-fairface",
- "label": "人脸属性识别模型FairFace",
- "describe": "给定一张带人脸的图片,返回其性别和年龄范围。",
- "hot": 2398,
- "pic": "example.png",
- "uuid": "cv-resnet34-face-attribute-recognition-fairface",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸属性识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-conformer-asr-nat-zh-cn-16k-aishell2-vocab5212-pytorch",
- "label": "Conformer语音识别-中文-aishell2-16k-离线-pytorch",
- "describe": "Conformer模型通过在self-attenion基础上叠加卷积模块来加强模型的局部信息建模能力,进一步提升了模型的效果。Conformer已经在AISHELL-1、AISHELL-2、LibriSpeech等多个开源数据上取得SOTA结果。",
- "hot": 2361,
- "pic": "example.jpeg",
- "uuid": "speech-conformer-asr-nat-zh-cn-16k-aishell2-vocab5212-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-nafnet-image-deblur-reds",
- "label": "NAFNet图像去模糊压缩",
- "describe": "NAFNet(Nonlinear Activation Free Network)提出了一个简单的基线,计算效率高。其不需要使用非线性激活函数(Sigmoid、ReLU、GELU、Softmax等),可以达到SOTA性能。",
- "hot": 2341,
- "pic": "example.jpg",
- "uuid": "cv-nafnet-image-deblur-reds",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像去模糊",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentence-similarity-chinese-tiny",
- "label": "StructBERT文本相似度-中文-通用-tiny",
- "describe": "StructBERT文本相似度-中文-通用-tiny是在structbert-tiny-chinese预训练模型的基础上,用atec、bq_corpus、chineseSTS、lcqmc、paws-x-zh五个数据集(52.5w条数据,正负比例0.48:0.52)训练出来的相似度匹配模型",
- "hot": 2325,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentence-similarity-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab3445-pytorch-offline",
- "label": "UniASR语音识别-中文-通用-8k-离线-pytorch",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 2323,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-zh-cn-8k-common-vocab3445-pytorch-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-safety-helmet",
- "label": "实时安全帽检测-通用",
- "describe": "本模型为高性能热门应用系列检测模型中的实时安全帽(头盔)检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有人头的坐标信息,以及是否佩戴安全帽(头盔)。更多具体信息请参考Model card。",
- "hot": 2301,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-safety-helmet",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-mmspeech-pretrain-large-zh",
- "label": "OFA-MMSpeech语音识别预训练-中文-通用领域-large",
- "describe": "对比SOTA,MMSpeech字错误率降低了48.3%/42.4%,效果达到1.6%/1.9%,远超SOTA 3.1%/3.3%(benchmark为AIShell1 dev/test)。",
- "hot": 2301,
- "pic": "example.jpg",
- "uuid": "ofa-mmspeech-pretrain-large-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-russian-large-ecom",
- "label": "RaNER命名实体识别-俄语-电商领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在俄语电商query数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2269,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-russian-large-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-english-large-science",
- "label": "RaNER命名实体识别-英文-科学领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在英文Science数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2245,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-english-large-science",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-tinynas-head-detection-damoyolo",
- "label": "实时人头检测-通用",
- "describe": "本模型为高性能热门应用系列检测模型中的实时人头检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有人头的坐标信息,并可用于行人计数等后续应用场景。更多具体信息请参考Model card。",
- "hot": 2238,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-head-detection-damoyolo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-fr2en",
- "label": "CSANMT连续语义增强机器翻译-法英-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 2220,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-fr2en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-paraformerbert-asr-nat-zh-cn-16k-aishell2-vocab5212-pytorch",
- "label": "ParaformerBert语音识别-中文-aishell2-16k-离线-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 2126,
- "pic": "example.jpg",
- "uuid": "speech-paraformerbert-asr-nat-zh-cn-16k-aishell2-vocab5212-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-spanish-large-ecom",
- "label": "RaNER命名实体识别-西班牙语-电商领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在西班牙语电商query数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 2125,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-spanish-large-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-chinese-tiny",
- "label": "CoROM文本向量-中文-通用领域-tiny",
- "describe": "基于CoROM-base预训练语言模型的通用领域中文文本表示模型,基于输入的句子产出对应的文本向量,文本向量可以使用在下游的文本检索、句子相似度计算、文本聚类等任务中。",
- "hot": 2116,
- "pic": "example.png",
- "uuid": "nlp-corom-sentence-embedding-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "ofa-text-to-image-synthesis-coco-large-en",
- "label": "OFA文生图模型-英文-通用领域-large",
- "describe": "文本到图像生成任务:输入一句英文描述文本,模型会返回一张符合文本描述的256*256分辨率图像。",
- "hot": 2108,
- "pic": "example.jpg",
- "uuid": "ofa-text-to-image-synthesis-coco-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-structbert-siamese-uie-chinese-base",
- "label": "SiameseUIE通用信息抽取-中文-base",
- "describe": "",
- "hot": 2097,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-siamese-uie-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "孪生通用信息抽取",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-vitb16-classification-vision-efficient-tuning-prefix",
- "label": "基础视觉模型高效调优-Prefix",
- "describe": "",
- "hot": 2027,
- "pic": "example.jpg",
- "uuid": "cv-vitb16-classification-vision-efficient-tuning-prefix",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "基础模型调优",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-ponet-fill-mask-chinese-base",
- "label": "PoNet预训练模型-中文-base",
- "describe": "nlp_ponet_fill-mask_chinese-base是用中文wiki训练的预训练PoNet模型。",
- "hot": 2024,
- "pic": "example.jpg",
- "uuid": "nlp-ponet-fill-mask-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "预训练",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "multi-modal-clip-vit-huge-patch14-zh",
- "label": "CLIP模型-中文-通用领域-huge",
- "describe": "本项目为CLIP模型的中文版本,使用大规模中文数据进行训练(~2亿图文对),旨在帮助用户实现中文领域的跨模态检索、图像表示等。视觉encoder采用vit结构,文本encoder采用roberta结构。 模型在多个中文图文检索数据集上进行了效果测试。",
- "hot": 1967,
- "pic": "example.jpg",
- "uuid": "multi-modal-clip-vit-huge-patch14-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "多模态表征",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-raner-named-entity-recognition-french-large-ecom",
- "label": "RaNER命名实体识别-法语-电商领域-large",
- "describe": "该模型是基于检索增强(RaNer)方法在法语电商query数据集训练的模型。本方法采用Transformer-CRF模型,使用xlm-roberta-large作为预训练模型底座,结合使用外部工具召回的相关句子作为额外上下文,使用Multi-view Training方式进行训练。",
- "hot": 1924,
- "pic": "example.jpg",
- "uuid": "nlp-raner-named-entity-recognition-french-large-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-paraformerbert-asr-nat-zh-cn-16k-aishell1-vocab4234-pytorch",
- "label": "ParaformerBert语音识别-中文-aishell1-16k-离线-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 1902,
- "pic": "example.jpg",
- "uuid": "speech-paraformerbert-asr-nat-zh-cn-16k-aishell1-vocab4234-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-palm2-0-pretrained-chinese-base",
- "label": "PALM 2.0预训练生成模型-中文-base",
- "describe": "达摩PALM 2.0中文Base预训练模型",
- "hot": 1890,
- "pic": "example.png",
- "uuid": "nlp-palm2-0-pretrained-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-bart-text-error-correction-chinese-law",
- "label": "BART文本纠错-中文-法律领域-large",
- "describe": "法研杯2022文书校对赛道冠军纠错模型(单模型)。",
- "hot": 1867,
- "pic": "example.png",
- "uuid": "nlp-bart-text-error-correction-chinese-law",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本纠错",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-paraformer-asr-nat-zh-cn-16k-aishell2-vocab5212-pytorch",
- "label": "Paraformer语音识别-中文-aishell2-16k-离线-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 1812,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-asr-nat-zh-cn-16k-aishell2-vocab5212-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-raft-video-frame-interpolation-practical",
- "label": "VFI-RAFT视频插帧-应用型",
- "describe": "偏实际应用的视频插帧模型,相较原版模型,该模型能支持任意倍率的帧率转换,同时在各种困难场景下如大运动、重复纹理、台标字幕等有更好更稳定的插帧效果。",
- "hot": 1810,
- "pic": "example.jpeg",
- "uuid": "cv-raft-video-frame-interpolation-practical",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频插帧",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-en",
- "label": "语音识别-英语-后处理- ITN模型",
- "describe": "英语文本反正则化。Inverse Text Processing for English.",
- "hot": 1787,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-maskdino-swin-l-image-instance-segmentation-coco",
- "label": "MaskDINO-SwinL图像实例分割",
- "describe": "SOTA通用目标检测和实例分割模型,backbone使用Swin transformer large。",
- "hot": 1781,
- "pic": "example.jpeg",
- "uuid": "cv-maskdino-swin-l-image-instance-segmentation-coco",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-zero-shot-classification-chinese-large",
- "label": "StructBERT零样本分类-中文-large",
- "describe": "该模型使用StructBERT-large在xnli_zh数据集(将英文数据集重新翻译得到中文数据集)上面进行了训练得到。",
- "hot": 1739,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-zero-shot-classification-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "零样本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet50-image-quality-assessment-degradation",
- "label": "图像画质损伤分析",
- "describe": "",
- "hot": 1721,
- "pic": "example.jpg",
- "uuid": "cv-resnet50-image-quality-assessment-degradation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像画质损伤分析",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-sentence-similarity-chinese-retail-base",
- "label": "StructBERT文本相似度-中文-电商-base",
- "describe": "StructBERT中文电商域文本相似度模型是在structbert-base-chinese预训练模型的基础上,用电商域标注数据训练出来的相似度匹配模型。",
- "hot": 1702,
- "pic": "example.png",
- "uuid": "nlp-structbert-sentence-similarity-chinese-retail-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-bert-document-segmentation-english-base",
- "label": "BERT文本分割-英文-通用领域",
- "describe": "该模型基于wiki-en公开语料训练,对未分割的长文本进行段落分割。提升未分割文本的可读性以及下游NLP任务的性能。",
- "hot": 1680,
- "pic": "example.jpg",
- "uuid": "nlp-bert-document-segmentation-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分割",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-ja",
- "label": "语音识别-日语-后处理- ITN模型",
- "describe": "日语文本反正则化。Inverse Text Processing for Japanese.",
- "hot": 1675,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-ja",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-kyong-korean-16k",
- "label": "语音合成-韩语-通用领域-16k-发音人kyong",
- "describe": "韩语语音合成女声16k模型,本模型使用Sambert-hifigan网络结构。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 1642,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-kyong-korean-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "multi-modal-clip-vit-large-patch14-zh",
- "label": "CLIP模型-中文-通用领域-large",
- "describe": "本项目为CLIP模型的中文版本,使用大规模中文数据进行训练(~2亿图文对),旨在帮助用户实现中文领域的跨模态检索、图像表示等。视觉encoder采用vit结构,文本encoder采用roberta结构。 模型在多个中文图文检索数据集上进行了效果测试。",
- "hot": 1633,
- "pic": "example.jpg",
- "uuid": "multi-modal-clip-vit-large-patch14-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "多模态表征",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-resnet50-face-reconstruction",
- "label": "人脸重建模型",
- "describe": "单图人脸重建榜单REALY冠军模型,相关论文被CVPR2023收录。",
- "hot": 1626,
- "pic": "example.jpg",
- "uuid": "cv-resnet50-face-reconstruction",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸重建",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-controlnet-controllable-image-generation-nine-annotators",
- "label": "ControlNet可控图像生成",
- "describe": "输入一张图像,指定控制类别并提供期望生成图像的描述prompt,模型会根据输入图像抽取相应的控制信息并生成精美图像。",
- "hot": 1614,
- "pic": "example.jpg",
- "uuid": "cv-controlnet-controllable-image-generation-nine-annotators",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "guohua-diffusion",
- "label": "国画Diffusion模型",
- "describe": "这是在国画上训练的微调Stable Diffusion模型",
- "hot": 1597,
- "pic": "example.jpg",
- "uuid": "guohua-diffusion",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "multilingual-glm-summarization-zh",
- "label": "mGLM多语言大模型-生成式摘要-中文",
- "describe": "mGLM多语言大模型可从大段文本中提取关键信息,为你生成简短的中文摘要,支持多种语言输入",
- "hot": 1589,
- "pic": "example.png",
- "uuid": "multilingual-glm-summarization-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本摘要",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-chinese-tiny-ecom",
- "label": "CoROM文本向量-中文-电商领域-tiny",
- "describe": "基于CoROM-tiny预训练语言模型的电商领域中文文本表示模型,基于输入的句子产出对应的文本向量,文本向量可以使用在下游的文本检索、句子相似度计算、文本聚类等任务中。",
- "hot": 1582,
- "pic": "example.jpg",
- "uuid": "nlp-corom-sentence-embedding-chinese-tiny-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-ir50-face-recognition-arcface",
- "label": "ArcFace人脸识别模型",
- "describe": "输入一张图片,检测矫正人脸区域后提取特征,两个人脸特征可用于人脸比对,多个人脸特征可用于人脸检索。",
- "hot": 1561,
- "pic": "example.jpg",
- "uuid": "cv-ir50-face-recognition-arcface",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-charctc-kws-phone-xiaoyun-commands",
- "label": "CTC语音唤醒-移动端-单麦-16k-小云-多命令词",
- "describe": "移动端语音多命令词模型,我们根据以往项目积累,挑选了多个场景常用命令词数据进行模型迭代,所得单一模型支持30+关键词的快速检测。",
- "hot": 1556,
- "pic": "example.jpg",
- "uuid": "speech-charctc-kws-phone-xiaoyun-commands",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音唤醒",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-uhdm-image-demoireing",
- "label": "uhdm图像去摩尔纹",
- "describe": "图像去摩尔纹",
- "hot": 1549,
- "pic": "example.jpg",
- "uuid": "cv-uhdm-image-demoireing",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像去摩尔纹",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-chinese-tiny-medical",
- "label": "CoROM文本向量-中文-医疗领域-tiny",
- "describe": "基于CoROM-tiny预训练语言模型的电商领域中文文本表示模型,基于输入的句子产出对应的文本向量,文本向量可以使用在下游的文本检索、句子相似度计算、文本聚类等任务中。",
- "hot": 1505,
- "pic": "example.jpg",
- "uuid": "nlp-corom-sentence-embedding-chinese-tiny-medical",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-corom-sentence-embedding-english-tiny",
- "label": "CoROM文本向量-英文-通用领域-tiny",
- "describe": "基于CoROM-Base预训练模型的通用领域英文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 1489,
- "pic": "example.png",
- "uuid": "nlp-corom-sentence-embedding-english-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-base-zh",
- "label": "OFA预训练模型-中文-通用领域-base",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 1488,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-base-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-es2en",
- "label": "CSANMT连续语义增强机器翻译-西英-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 1488,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-es2en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-masha-russian-16k",
- "label": "语音合成-俄语-通用领域-16k-发音人masha",
- "describe": "俄语语音合成女声16k模型,本模型使用Sambert-hifigan网络结构。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 1348,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-masha-russian-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-paraformer-tiny-commandword-asr-nat-zh-cn-16k-vocab544-pytorch",
- "label": "Paraformer语音识别-中文-端上指令词-16k-离线-tiny",
- "describe": "轻量化小词表Paraformer中文指令词识别模型,参数量控制在5M左右,支持通用智能家居交互等常规指令词,并且使用share embedding策略进一步缩小参数量。",
- "hot": 1312,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-tiny-commandword-asr-nat-zh-cn-16k-vocab544-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-swin-t-referring-video-object-segmentation",
- "label": "MTTR文本指导的视频目标分割-英文",
- "describe": "",
- "hot": 1298,
- "pic": "example.jpg",
- "uuid": "cv-swin-t-referring-video-object-segmentation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本指导的视频目标分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-swinl-image-object-detection-dino",
- "label": "DINO-高精度目标检测模型",
- "describe": "本模型是DINO高精度目标检测模型,采用SwinL主干网络,在COCO验证集精度可达63.39%。",
- "hot": 1276,
- "pic": "example.jpg",
- "uuid": "cv-swinl-image-object-detection-dino",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-de",
- "label": "语音识别-德语-后处理- ITN模型",
- "describe": "德语文本反正则化。Inverse Text Processing for German.",
- "hot": 1274,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-de",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-resnet101-image-single-human-parsing",
- "label": "M2FP单人人体解析",
- "describe": "",
- "hot": 1273,
- "pic": "example.jpg",
- "uuid": "cv-resnet101-image-single-human-parsing",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-unet-video-colorization",
- "label": "DeOldify视频上色",
- "describe": "DeOldify 是上色领域比较有名的开源算法,模型利用 ResNet 作为 encoder 构建一个 UNet 结构的网络,并提出了多个不同的训练版本,本模型是用于视频上色的版本。",
- "hot": 1261,
- "pic": "example.jpg",
- "uuid": "cv-unet-video-colorization",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频上色",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "chatyuan-large-v2",
- "label": "元语功能型对话大模型v2",
- "describe": "元语功能型对话大模型这个模型可以用于问答、结合上下文做对话、做各种生成任务,包括创意性写作,也能回答一些像法律、新冠等领域问题。它基于PromptCLUE-large结合数亿条功能对话多轮对话数据进一步训练得到。是元语功能型对话大模型v1的升级版",
- "hot": 1254,
- "pic": "example.jpeg",
- "uuid": "chatyuan-large-v2",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "端到端文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-vitadapter-semantic-segmentation-cocostuff164k",
- "label": "Mask2Former-ViTAdapter语义分割",
- "describe": "该语义分割模型基于Mask2Former架构,ViTAdapter为backbone,训练数据库为COCO-Stuff164k。",
- "hot": 1244,
- "pic": "example.jpg",
- "uuid": "cv-vitadapter-semantic-segmentation-cocostuff164k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-bert-relation-extraction-chinese-base",
- "label": "RoBERTa关系抽取-中文-通用-base",
- "describe": "百科关系抽取模型是在hfl/chinese-roberta-wwm-ext预训练模型的基础上,用duie数据集训练出来的关系抽取模型。",
- "hot": 1237,
- "pic": "example.jpg",
- "uuid": "nlp-bert-relation-extraction-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "关系抽取",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet101-image-multiple-human-parsing",
- "label": "M2FP多人人体解析",
- "describe": "",
- "hot": 1223,
- "pic": "example.jpg",
- "uuid": "cv-resnet101-image-multiple-human-parsing",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "通用图像分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-paraformer-large-asr-nat-zh-cn-16k-aishell1-vocab8404-pytorch",
- "label": "Paraformer语音识别-中文-aishell1-16k-离线-large-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 1194,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-large-asr-nat-zh-cn-16k-aishell1-vocab8404-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-nli-chinese-base",
- "label": "StructBERT自然语言推理-中文-通用-base",
- "describe": "StructBERT自然语言推理-中文-通用-base是在structbert-base-chinese预训练模型的基础上,用CMNLI、OCNLI两个数据集(45.8w条数据)训练出来的自然语言推理模型。",
- "hot": 1175,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-nli-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "自然语言推理",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-tadaconv-action-recognition",
- "label": "TAdaConv动作识别模型-英文-通用领域",
- "describe": "TAdaConv是一种在动作识别模型中即插即用的时序自适应卷积(Temporally-Adaptive Convolutions)。可以明显提升SlowFast、R2D和R3D等模型性能。",
- "hot": 1166,
- "pic": "example.png",
- "uuid": "cv-tadaconv-action-recognition",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "动作识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-fact-checking-chinese-base",
- "label": "StructBERT事实准确性检测-中文-电商-base",
- "describe": "StructBERT事实准确性检测-中文-电商-base是在structbert-base-chinese预训练模型的基础上,使用业务数据训练出的自然语言推理模型,用于事实准确性检测,输入两个句子,判断两个句子描述的事实是否一致。",
- "hot": 1154,
- "pic": "example.jpeg",
- "uuid": "nlp-structbert-fact-checking-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "自然语言推理",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-gpt3-text-generation-30b",
- "label": "GPT-3预训练生成模型-中文-30B",
- "describe": "",
- "hot": 1104,
- "pic": "example.jpg",
- "uuid": "nlp-gpt3-text-generation-30b",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-fr",
- "label": "语音识别-法语-后处理- ITN模型",
- "describe": "法语文本反正则化。Inverse Text Processing for French.",
- "hot": 1093,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-fr",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-ru",
- "label": "语音识别-俄语-后处理- ITN模型",
- "describe": "俄语文本反正则化。Inverse Text Processing for Russian.",
- "hot": 1060,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-ru",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-zero-shot-classification-chinese-tiny",
- "label": "StructBERT零样本分类-中文-tiny",
- "describe": "该模型使用StructBERT-base在xnli_zh数据集(将英文数据集重新翻译得到中文数据集)上面进行了训练得到。",
- "hot": 1050,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-zero-shot-classification-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "零样本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-traffic-sign",
- "label": "实时交通标识检测-自动驾驶领域",
- "describe": "本模型为高性能热门应用系列检测模型中的交通标识检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有交通标识的坐标信息。更多具体信息请参考Model card。",
- "hot": 1024,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-traffic-sign",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-data2vec-pretrain-zh-cn-aishell2-16k-pytorch",
- "label": "Data2vec语音识别-预训练-中文-aishell2-16k-pytorch",
- "describe": "近年来,随着预训练的流行,许多研究致力于利用预训练的方式来充分利用大量的无监督数据,帮助提升在有监督语音数据有限情况下的语音识别的性能。wav2vec,HuBERT,WavLM等方法,都通过无监督预训练的方式在语音识别任务上取得了不错的识别率。2022年,Meta AI在ICML上提出了data2vec,具体结构如下图琐事,其能同时应用于语音、视觉、自然语言处理等不同模态,且都取得了不错的性能。",
- "hot": 1018,
- "pic": "example.jpeg",
- "uuid": "speech-data2vec-pretrain-zh-cn-aishell2-16k-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-mobilenet-v2-bad-image-detecting",
- "label": "异常图像检测",
- "describe": "基于mobilenet-v2的简化版网络,检测图像是否为花屏、绿屏或者正常图像。",
- "hot": 1003,
- "pic": "example.jpg",
- "uuid": "cv-mobilenet-v2-bad-image-detecting",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "异常图像检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet50-ocr-detection-vlpt",
- "label": "读光-文字检测-单词检测模型-英文-VLPT预训练",
- "describe": "给定一张图片,检测出图内文字并给出多边形包围框。检测模型使用DB,backbone初始化参数基于多模态交互预训练方法VLPT。",
- "hot": 991,
- "pic": "example.jpg",
- "uuid": "cv-resnet50-ocr-detection-vlpt",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文字检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "speech-sambert-hifigan-tts-ainan-zh-cn-16k",
- "label": "语音合成-中文-通用领域-16k-发音人ainan",
- "describe": "中文语音合成男声16k模型,本模型使用Sambert-hifigan网络结构。其中后端声学模型的SAM-BERT,将时长模型和声学模型联合进行建模。声码器在HIFI-GAN开源工作的基础上,我们针对16k, 48k采样率下的模型结构进行了调优设计,并提供了基于因果卷积的低时延流式生成和chunk流式生成机制,可与声学模型配合支持CPU、GPU等硬件条件下的实时流式合成。",
- "hot": 978,
- "pic": "example.jpg",
- "uuid": "speech-sambert-hifigan-tts-ainan-zh-cn-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音合成",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-id",
- "label": "语音识别-印尼语-后处理- ITN模型",
- "describe": "印尼语文本反正则化。Inverse Text Processing for Indonesian.",
- "hot": 960,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-id",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-pt",
- "label": "语音识别-葡萄牙语-后处理- ITN模型",
- "describe": "葡萄语文本反正则化。Inverse Text Processing for Portuguese.",
- "hot": 930,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-pt",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-nli-chinese-large",
- "label": "StructBERT自然语言推理-中文-通用-large",
- "describe": "StructBERT自然语言推理-中文-通用-large是在structbert-large-chinese预训练模型的基础上,用CMNLI、OCNLI两个数据集(45.8w条数据)训练出来的自然语言推理模型。",
- "hot": 920,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-nli-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "自然语言推理",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-es",
- "label": "语音识别-西班牙-后处理- ITN模型",
- "describe": "西班牙文本反正则化。Inverse Text Processing for Spanish.",
- "hot": 919,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-es",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-vi",
- "label": "语音识别-越南语-后处理- ITN模型",
- "describe": "越南语文本反正则化。Inverse Text Processing for Vietnamese.",
- "hot": 911,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-vi",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-passage-ranking-chinese-tiny",
- "label": "CoROM语义相关性-中文-通用领域-tiny",
- "describe": "基于ROM-tiny预训练模型的通用领域中文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 899,
- "pic": "example.jpg",
- "uuid": "nlp-corom-passage-ranking-chinese-tiny",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "multi-modal-clip-vtretrival-msrvtt-53",
- "label": "视频文本表征模型-英文-通用领域",
- "describe": "该模型是在10亿公开英文图文数据训练的多模态模型。视觉encoder采用vit-large结构,文本encoder采用bert-base结构。 模型在视频文本检索等数据集上进行了zero-shot和finetune效果测试,能够在msrvtt上达到sota结果。",
- "hot": 889,
- "pic": "example.jpg",
- "uuid": "multi-modal-clip-vtretrival-msrvtt-53",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频多模态表征",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-star-conversational-text-to-sql",
- "label": "SPACE-T表格问答预训练模型-英文-通用领域-Large",
- "describe": "本项目是多轮Text-to-SQL模型,可针对不同领域数据库和用户直接进行多轮对话,生成相应的SQL查询语句。用户可以在对话过程中表达自己对数据库模式的查询要求,并在系统的帮助下生成符合要求的SQL查询语句。",
- "hot": 867,
- "pic": "example.jpg",
- "uuid": "nlp-star-conversational-text-to-sql",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "表格问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-ko",
- "label": "语音识别-韩语-后处理- ITN模型",
- "describe": "韩语文本反正则化。Inverse Text Processing for Korean.",
- "hot": 861,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-ko",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-ponet-extractive-summarization-topic-level-chinese-base",
- "label": "PoNet抽取式话题摘要模型-中文-base-ICASSP2023-MUG-Track2",
- "describe": "该模型基于PoNet模型架构,在AliMeeting4MUG Corpus训练,进行抽取式话题摘要任务。",
- "hot": 859,
- "pic": "example.jpg",
- "uuid": "nlp-ponet-extractive-summarization-topic-level-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "抽取式摘要",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-inverse-text-processing-fun-text-processing-itn-tl",
- "label": "语音识别-菲律宾语-后处理- ITN模型",
- "describe": "菲律宾语文本反正则化。Inverse Text Processing for Tagalog.",
- "hot": 846,
- "pic": "example.jpg",
- "uuid": "speech-inverse-text-processing-fun-text-processing-itn-tl",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "逆文本正则化",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "mgeo-backbone-chinese-base",
- "label": "MGeo多任务多模态地址预训练底座-中文-base",
- "describe": "MGeo是适用于多种地址任务的预训练底座模型,该模型基于地图-文本多模态架构,使用多任务预训练(MOMETAS)技术融合了注意力对抗预训练(ASA)、句子对预训练(MaSTS)、多模态预训练,训练得到适合于多类地址任务的预训练底座,能够下游广泛的地址处理任务带来性能提升。",
- "hot": 819,
- "pic": "example.jpg",
- "uuid": "mgeo-backbone-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-conformer-asr-nat-zh-cn-16k-aishell1-vocab4234-pytorch",
- "label": "Conformer语音识别-中文-aishell1-16k-离线-pytorch",
- "describe": "Conformer模型通过在self-attenion基础上叠加卷积模块来加强模型的局部信息建模能力,进一步提升了模型的效果。Conformer已经在AISHELL-1、AISHELL-2、LibriSpeech等多个开源数据上取得SOTA结果。",
- "hot": 807,
- "pic": "example.jpeg",
- "uuid": "speech-conformer-asr-nat-zh-cn-16k-aishell1-vocab4234-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-stable-diffusion-paint-by-example",
- "label": "图像示例替换",
- "describe": "给定一张原图和对应的mask区域,将给定的参考图自适应地替换上去",
- "hot": 801,
- "pic": "example.jpg",
- "uuid": "cv-stable-diffusion-paint-by-example",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像示例替换",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "nlp-structbert-faq-question-answering-chinese-finance-base",
- "label": "StructBERT FAQ问答-中文-金融领域-base",
- "describe": "金融领域FAQ问答模型以StructBERT FAQ问答-中文-通用领域-base模型为基础,在金融领域数据上微调得到,适用于金融领域FAQ问答任务,包括但不局限于:银行、保险等场景;",
- "hot": 750,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-faq-question-answering-chinese-finance-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "FAQ问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-daflow-virtual-try-on-base",
- "label": "DAFlow虚拟试衣模型-VITON数据",
- "describe": "DAFlow是一种单阶段虚拟试衣框架,无需中间分割结果作为label,直接用模特上身图作为监督。同时本工作提出一种新的空间变换结构,在虚拟试衣和一些变换任务上达到SOTA.",
- "hot": 748,
- "pic": "example.jpg",
- "uuid": "cv-daflow-virtual-try-on-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "虚拟试衣",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-emotion-classification-chinese-base",
- "label": "StructBERT情绪分类-中文-七分类-base",
- "describe": "",
- "hot": 738,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-emotion-classification-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet50-product-bag-embedding-models",
- "label": "商品图像同款特征",
- "describe": "本模型是对商品图像进行表征向量提取,用户可基于表征向量进行大规模的同款/相似款商品搜索;无需额外输入,模型可自动进行箱包商品的主体抠图,并基于主体提取结果完成表征向量提取。",
- "hot": 731,
- "pic": "example.jpeg",
- "uuid": "cv-resnet50-product-bag-embedding-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "商品图片特征",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-bert-sentence-similarity-english-base",
- "label": "BERT文本相似度-英文-base-学术数据集paws",
- "describe": "该模型是在bert-base-uncased预训练模型的基础上,用paws数据集训练出来的文本相似度匹配模型。",
- "hot": 724,
- "pic": "example.jpg",
- "uuid": "nlp-bert-sentence-similarity-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-ecbsr-image-super-resolution-mobile",
- "label": "ECBSR端上图像超分模型",
- "describe": "ECBSR模型基于Edgeoriented Convolution Block (ECB)模块构建,完整模型可导出为简洁的CNN网络结构,适用于移动端、嵌入式等严格限制算力的场景。",
- "hot": 715,
- "pic": "example.jpg",
- "uuid": "cv-ecbsr-image-super-resolution-mobile",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像超分辨",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-domain-classification-chinese",
- "label": "FastText文本领域分类-中文-国民经济行业18大类",
- "describe": "用于中文的文本领域分类,分类依据为国民经济行业分类(GB/T 4754—2017),原分类标准有20大类,目前支持18个行业的分类:交通运输仓储邮政\\住宿餐饮\\信息软件\\农业\\制造业\\卫生医疗\\国际组织\\建筑\\房地产\\政府组织\\教育\\文体娱乐\\水利环境\\电力燃气水生产\\科学技术\\租赁法律\\采矿\\金融。",
- "hot": 715,
- "pic": "example.jpg",
- "uuid": "nlp-domain-classification-chinese",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-yolov5-video-multi-object-tracking-fairmot",
- "label": "视频多目标跟踪-行人",
- "describe": "该模型采用基于FairMOT的方案,输入待跟踪视频,可端对端推理得到视频中的所有行人的运动轨迹。",
- "hot": 710,
- "pic": "example.gif",
- "uuid": "cv-yolov5-video-multi-object-tracking-fairmot",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频多目标跟踪",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-tinynas-object-detection-damoyolo-phone",
- "label": "实时手机检测-通用",
- "describe": "本模型为高性能热门应用系列检测模型中的实时手机检测模型,基于面向工业落地的高性能检测框架DAMOYOLO,其精度和速度超越当前经典的YOLO系列方法。用户使用的时候,仅需要输入一张图像,便可以获得图像中所有手机的坐标信息,并可用于打电话检测等后续应用场景。更多具体信息请参考Model card。",
- "hot": 682,
- "pic": "example.jpg",
- "uuid": "cv-tinynas-object-detection-damoyolo-phone",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "垂类目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-vgg19-facial-expression-recognition-fer",
- "label": "人脸表情识别模型FER",
- "describe": "给定一张带人脸的图片,返回图片中人脸的表情。目前支持7种表情:愤怒,厌恶,恐惧,快乐,悲伤,惊讶,中立。",
- "hot": 660,
- "pic": "example.jpg",
- "uuid": "cv-vgg19-facial-expression-recognition-fer",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸表情识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-xlmr-named-entity-recognition-indo-ecommerce-title",
- "label": "XLM-R命名实体识别-印尼语-电商领域(商品标题)-base",
- "describe": "XLM-R命名实体识别-印尼语-电商领域(商品标题)-base是基于20K电商领域商品标题数据训练得到的印尼语命名实体识别模型,可根据用户输入的印尼语商品标题文本产出命名实体识别结果。",
- "hot": 655,
- "pic": "example.png",
- "uuid": "nlp-xlmr-named-entity-recognition-indo-ecommerce-title",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-structbert-keyphrase-extraction-base-icassp2023-mug-track4-baseline",
- "label": "StructBert关键词抽取-中文-base-ICASSP2023-MUG-Track4",
- "describe": "ICASSP2023 MUG Track4 关键词抽取Baseline",
- "hot": 620,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-keyphrase-extraction-base-icassp2023-mug-track4-baseline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "glm130b",
- "label": "GLM130B-中英大模型",
- "describe": "模型集成中",
- "hot": 588,
- "pic": "example.jpg",
- "uuid": "glm130b",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-tiny-en",
- "label": "OFA预训练模型-英文-通用领域-tiny",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 569,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-tiny-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-paraformer-large-asr-nat-zh-cn-16k-aishell2-vocab8404-pytorch",
- "label": "Paraformer语音识别-中文-aishell2-16k-离线-large-pytorch",
- "describe": "Paraformer是一种非自回归端到端语音识别模型。非自回归模型相比于目前主流的自回归模型,可以并行的对整条句子输出目标文字,特别适合利用GPU进行并行推理。Paraformer是目前已知的首个在工业大数据上可以获得和自回归端到端模型相同性能的非自回归模型。配合GPU推理,可以将推理效率提升10倍,从而将语音识别云服务的机器成本降低接近10倍。",
- "hot": 559,
- "pic": "example.jpg",
- "uuid": "speech-paraformer-large-asr-nat-zh-cn-16k-aishell2-vocab8404-pytorch",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-corom-passage-ranking-chinese-tiny-ecom",
- "label": "CoROM语义相关性-中文-电商领域-tiny",
- "describe": "基于CoROM-Base预训练模型的电商领域中文语义相关性模型,模型以一个source sentence以及一个句子列表作为输入,最终输出source sentence与列表中每个句子的相关性得分(0-1,分数越高代表两者越相关)。",
- "hot": 551,
- "pic": "example.jpg",
- "uuid": "nlp-corom-passage-ranking-chinese-tiny-ecom",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-xlmr-word-segmentation-thai",
- "label": "XLM-R分词-泰语-通用领域-base",
- "describe": "XLM-R分词-泰语-通用领域-base是基于BEST-2010数据训练得到的泰语分词模型,可根据用户输入的泰语文本产出分词结果。",
- "hot": 543,
- "pic": "example.png",
- "uuid": "nlp-xlmr-word-segmentation-thai",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "分词",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-resnet101-detection-fewshot-defrcn",
- "label": "defrcn少样本目标检测",
- "describe": "",
- "hot": 539,
- "pic": "example.jpg",
- "uuid": "cv-resnet101-detection-fewshot-defrcn",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像少样本目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-cartoon-stable-diffusion-clipart",
- "label": "卡通系列文生图模型-剪贴画",
- "describe": "",
- "hot": 529,
- "pic": "example.jpg",
- "uuid": "cv-cartoon-stable-diffusion-clipart",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "nlp-bert-zero-shot-english-base",
- "label": "BERT零样本分类-英文-base-学术数据集mnli",
- "describe": "该模型使用bert-base-uncased在multi_nli数据集(将英文数据集重新翻译得到中文数据集)上面进行了训练得到。",
- "hot": 525,
- "pic": "example.jpg",
- "uuid": "nlp-bert-zero-shot-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "零样本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-bert-entity-embedding-chinese-base",
- "label": "Bert实体向量-中文-通用领域-base",
- "describe": "",
- "hot": 514,
- "pic": "example.png",
- "uuid": "nlp-bert-entity-embedding-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本向量",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-manual-face-liveness-flxc",
- "label": "静默人脸活体检测模型-炫彩",
- "describe": "静默人脸活体检测模型-炫彩",
- "hot": 512,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-liveness-flxc",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸活体检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-faq-question-answering-chinese-gov-base",
- "label": "StructBERT FAQ问答-中文-政务领域-base",
- "describe": "政务领域FAQ问答模型以StructBERT FAQ问答-中文-通用领域-base模型为基础,在政务领域数据上微调得到,适用于政务领域FAQ问答任务,包括但不局限于社保、公积金等场景;",
- "hot": 496,
- "pic": "example.jpg",
- "uuid": "nlp-structbert-faq-question-answering-chinese-gov-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "FAQ问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-ponet-extractive-summarization-doc-level-chinese-base",
- "label": "PoNet抽取式篇章摘要模型-中文-base-ICASSP2023-MUG-Track2",
- "describe": "该模型基于PoNet模型架构,在AliMeeting4MUG Corpus训练,进行抽取式篇章摘要任务。",
- "hot": 494,
- "pic": "example.jpg",
- "uuid": "nlp-ponet-extractive-summarization-doc-level-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "抽取式摘要",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-pt-16k-common-vocab1617-tensorflow1-online",
- "label": "UniASR语音识别-葡萄牙语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 492,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-pt-16k-common-vocab1617-tensorflow1-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-bert-entity-matching-chinese-base",
- "label": "Bert实体相关性-中文-通用领域-base",
- "describe": "输入带实体标记的句子A,以及一个候选句子列表,模型输出句子A中的实体与列表中每个候选句子的相关性得分(0-1,分数越高代表两者越相关),",
- "hot": 487,
- "pic": "example.png",
- "uuid": "nlp-bert-entity-matching-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语义相关性",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "ofa-text-classification-mnli-large-en",
- "label": "OFA自然语言推理模型-英文-数据集MNLI-large",
- "describe": "给定一个前提句和一个假设句,任务是预测前提是否包含假设(蕴含),与假设相矛盾(矛盾),或者两者都不包含(中性)。",
- "hot": 469,
- "pic": "example.jpg",
- "uuid": "ofa-text-classification-mnli-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-ir-face-recognition-ood-rts",
- "label": "人脸识别OOD模型",
- "describe": "",
- "hot": 467,
- "pic": "example.jpg",
- "uuid": "cv-ir-face-recognition-ood-rts",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-base-en",
- "label": "OFA预训练模型-英文-通用领域-base",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 464,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-base-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "nlp-nested-ner-named-entity-recognition-chinese-base-med",
- "label": "NestedNER命名实体识别-中文-医疗领域-base",
- "describe": "",
- "hot": 453,
- "pic": "example.jpeg",
- "uuid": "nlp-nested-ner-named-entity-recognition-chinese-base-med",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-vi-16k-common-vocab1001-pytorch-online",
- "label": "UniASR语音识别-越南语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 451,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-vi-16k-common-vocab1001-pytorch-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-user-satisfaction-estimation-chinese",
- "label": "HiTransUSE用户满意度估计-中文-电商-base",
- "describe": "支持对话级的用户满意度分析,输出(不满意,中立,满意)三种标签",
- "hot": 436,
- "pic": "example.png",
- "uuid": "nlp-user-satisfaction-estimation-chinese",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分类",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-video-inpainting",
- "label": "视频编辑-通用领域",
- "describe": "视频inpainting模型,输入视频和mask,可实现对指定帧范围内的video inpainting",
- "hot": 436,
- "pic": "example.jpg",
- "uuid": "cv-video-inpainting",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频修复",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-ponet-document-segmentation-topic-level-chinese-base",
- "label": "PoNet文本话题分割模型-中文-base-ICASSP2023-MUG-Track1",
- "describe": "该模型基于PoNet模型架构,在AliMeeting4MUG Corpus训练,对带段落的长文本进行中文话题分割。",
- "hot": 431,
- "pic": "example.png",
- "uuid": "nlp-ponet-document-segmentation-topic-level-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本分割",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-pt-16k-common-vocab1617-tensorflow1-offline",
- "label": "UniASR语音识别-葡萄牙语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 427,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-pt-16k-common-vocab1617-tensorflow1-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "ofa-visual-entailment-snli-ve-distilled-v2-en",
- "label": "OFA视觉语义蕴含-英文-通用领域-蒸馏33M",
- "describe": "视觉蕴含任务:给定一个图像作为前提,一个自然语言句子作为假设,要求模型正确分配图像和文本三个标签(蕴含、中性和矛盾)。",
- "hot": 392,
- "pic": "example.jpg",
- "uuid": "ofa-visual-entailment-snli-ve-distilled-v2-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉蕴含",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-object-detection-3d-depe",
- "label": "DEPE-3D目标检测-自动驾驶领域",
- "describe": "DEPE是基于Transformer的纯视觉3D目标检测模型,在PETRV2基础上,使用LiDAR点云监督depth训练并提升3DPE的准确度,推理阶段只依赖视觉信息,相比原PETRv2方法在nuScenes-val上有0.52%的提升。",
- "hot": 382,
- "pic": "example.jpg",
- "uuid": "cv-object-detection-3d-depe",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "3D目标检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-summarization-gigaword-large-en",
- "label": "OFA文本摘要-英文-数据集gigaword-large",
- "describe": "gigaword摘要任务:给定长文本,输出文本摘要。",
- "hot": 376,
- "pic": "example.png",
- "uuid": "ofa-summarization-gigaword-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本摘要",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-vi-16k-common-vocab1001-pytorch-offline",
- "label": "UniASR语音识别-越南语-通用-16k-离线",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 366,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-vi-16k-common-vocab1001-pytorch-offline",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "speech-uniasr-asr-2pass-fa-16k-common-vocab1257-pytorch-online",
- "label": "UniASR语音识别-波斯语-通用-16k-实时",
- "describe": "UniASR是离线流式一体化语音识别系统。UniASR同时具有高精度和低延时的特点,不仅能够实时输出语音识别结果,而且能够在说话句尾用高精度的解码结果修正输出,与此同时,UniASR采用动态延时训练的方式,替代了之前维护多套延时流式系统的做法。",
- "hot": 359,
- "pic": "example.jpg",
- "uuid": "speech-uniasr-asr-2pass-fa-16k-common-vocab1257-pytorch-online",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "语音识别",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "nlp-structbert-fill-mask-chinese-large",
- "label": "StructBERT完形填空模型-中文-large",
- "describe": "nlp_structbert_fill-mask_chinese-large是海量中文数据训练的自然语言理解预训练模型。",
- "hot": 354,
- "pic": "example.png",
- "uuid": "nlp-structbert-fill-mask-chinese-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "预训练",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-vitb-video-single-object-tracking-procontext",
- "label": "ProContEXT视频单目标跟踪-通用领域",
- "describe": "该模型是基于Transformer的单目标跟踪网络,在开源数据集GOT-10k, TrackingNet上均达到SOTA精度。",
- "hot": 350,
- "pic": "example.gif",
- "uuid": "cv-vitb-video-single-object-tracking-procontext",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频单目标跟踪",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-resnet50-live-category",
- "label": "直播商品类目识别模型-中文-电商领域",
- "describe": "本模型采用ResNet-50网络结构,对直播视频进行商品类目识别,输入视频片段,输出直播商品类目标签,目前已经覆盖了8K多类的细粒度的商品类别。",
- "hot": 341,
- "pic": "example.jpg",
- "uuid": "cv-resnet50-live-category",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "直播视频内容识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-bilibili-image-super-resolution",
- "label": "RealCUGAN图像超分辨率",
- "describe": "todo",
- "hot": 334,
- "pic": "example.jpg",
- "uuid": "cv-bilibili-image-super-resolution",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像超分辨",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-mgimn-faq-question-answering-chinese-base",
- "label": "MGIMN FAQ问答-中文-通用领域-base",
- "describe": "MGIMN FAQ问答模型以StructBERT预训练模型为底座,采用多维度交互式匹配模型网络结构,通过小样本meta-learning的方式在海量业务数据预训练(亿级)、微调(百万级)得到,相对于StructBERT模型效果更优,适用于FAQ问答任务和通用小样本分类任务;",
- "hot": 332,
- "pic": "example.png",
- "uuid": "nlp-mgimn-faq-question-answering-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "FAQ问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-xlmr-named-entity-recognition-viet-ecommerce-title",
- "label": "XLM-R命名实体识别-越南语-电商领域(商品标题)-base",
- "describe": "XLM-R命名实体识别-越南语-电商领域(商品标题)-base是基于20K电商领域商品标题数据训练得到的越南语命名实体识别模型,可根据用户输入的越南语商品标题文本产出命名实体识别结果。",
- "hot": 331,
- "pic": "example.png",
- "uuid": "nlp-xlmr-named-entity-recognition-viet-ecommerce-title",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "命名实体识别",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-mdm-motion-generation",
- "label": "运动生成-人体运动-英文",
- "describe": "根据文本描述,生成对应的人体运动数据",
- "hot": 321,
- "pic": "example.gif",
- "uuid": "cv-mdm-motion-generation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "运动生成",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "uni-fold-monomer",
- "label": "Uni-Fold-Monomer 开源的蛋白质单体结构预测模型",
- "describe": "开源的蛋白质单体结构预测模型,输入蛋白质单体的一级结构(1D序列),预测蛋白质的三级结构(3D位置),同时给出预测结果的置信度。本模型主要用于蛋白质单体的预测。",
- "hot": 311,
- "pic": "example.jpg",
- "uuid": "uni-fold-monomer",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "蛋白质结构生成",
- "field": "未知"
- },
- {
- "price": "1",
- "name": "cv-hdformer-body-3d-keypoints-video",
- "label": "HDFormer人体关键点-通用领域-3D",
- "describe": "输入一段单人视频,实现端到端的3D人体关键点检测,输出视频中每一帧的3D人体关键点坐标。",
- "hot": 311,
- "pic": "example.jpg",
- "uuid": "cv-hdformer-body-3d-keypoints-video",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人体3D关键点",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-csanmt-translation-ru2en-base",
- "label": "CSANMT连续语义增强机器翻译-俄英-通用领域-base",
- "describe": "基于连续语义增强的神经机器翻译模型以有限的训练样本为锚点,学习连续语义分布以建模全局的句子空间,并据此构建神经机器翻译引擎,有效提升数据的利用效率,显著改善模型的泛化能力和鲁棒性。",
- "hot": 301,
- "pic": "example.jpg",
- "uuid": "nlp-csanmt-translation-ru2en-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "翻译",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-manual-face-liveness-flir",
- "label": "人脸活体检测模型-IR",
- "describe": "",
- "hot": 300,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-liveness-flir",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸活体检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-visual-entailment-snli-ve-large-en",
- "label": "OFA图像语义蕴含-英文-通用领域-large",
- "describe": "图文蕴含任务:给定图片和文本a(对图片的陈述),文本b(可选),判断文本c是否成立。",
- "hot": 298,
- "pic": "example.jpg",
- "uuid": "ofa-visual-entailment-snli-ve-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视觉蕴含",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-manual-face-liveness-flrgb",
- "label": "人脸活体检测模型-RGB",
- "describe": "",
- "hot": 293,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-liveness-flrgb",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸活体检测",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-r2p1d-video-embedding",
- "label": "CMD视频表征模型",
- "describe": "一种自监督视频表征学习方案,通过在代理任务中显式地解耦场景与运动信息(context and motion decoupling),强制视频模型同时捕捉静态背景与动态行为两方面特征。",
- "hot": 285,
- "pic": "example.jpg",
- "uuid": "cv-r2p1d-video-embedding",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频表征",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-bert-fill-mask-chinese-base",
- "label": "BERT完形填空模型-中文-base",
- "describe": "nlp_bert_fill-mask_chinese-base 是wikipedia_zh/baike/news训练的自然语言理解预训练模型。",
- "hot": 273,
- "pic": "example.png",
- "uuid": "nlp-bert-fill-mask-chinese-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "预训练",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-veco-fill-mask-large",
- "label": "VECO完形填空模型-多语言-large",
- "describe": "nlp_veco_fill-mask-large是CommonCrawl Corpus训练的自然语言理解多语言预训练模型。",
- "hot": 272,
- "pic": "example.png",
- "uuid": "nlp-veco-fill-mask-large",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "预训练",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-cartoon-stable-diffusion-design",
- "label": "卡通系列文生图模型",
- "describe": "",
- "hot": 271,
- "pic": "example.jpg",
- "uuid": "cv-cartoon-stable-diffusion-design",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "cv-nerf-3d-reconstruction-accelerate-damo",
- "label": "NeRF快速三维重建模型",
- "describe": "输入一段环绕物体一周的视频,基于隐式神经辐射场(NeRF),对该物体进行快速3d重建渲染。",
- "hot": 264,
- "pic": "example.jpg",
- "uuid": "cv-nerf-3d-reconstruction-accelerate-damo",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "NeRF快速三维重建模型",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-unet-video-deinterlace",
- "label": "视频去场纹",
- "describe": "视频去场纹,相比较于yadif、DIN等SOTA方法,对于大运动和低画质场景下的场纹有较好的场纹去除能力",
- "hot": 245,
- "pic": "example.jpg",
- "uuid": "cv-unet-video-deinterlace",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频去场纹",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-image-caption-meme-large-zh",
- "label": "OFA表情包文本生成器",
- "describe": "",
- "hot": 230,
- "pic": "example.jpg",
- "uuid": "ofa-image-caption-meme-large-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-dro-resnet18-video-depth-estimation-indoor",
- "label": "循环神经优化器-视频流深度和相机轨迹估计",
- "describe": "我们提出zero-order的循环神经网络优化器(DRO), 不需要求解梯度, 直接利用神经网络来预测下次更新的方向和步长。将优化目标cost,放入到神经网络中,每次迭代都会参考之前尝试的历史信息,从而给出更加精准的预测。也就是说,如果错误的预测值,就会使得cost变大,正确的预测值会使得cost变小,在不断尝试中,神经网络学习到了如何使得cost变小。",
- "hot": 223,
- "pic": "example.jpg",
- "uuid": "cv-dro-resnet18-video-depth-estimation-indoor",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频深度估计",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-mgimn-faq-question-answering-chinese-finance-base",
- "label": "MGIMN FAQ问答-中文-金融领域-base",
- "describe": "金融领域MGIMN FAQ问答模型以MGIMN FAQ问答-中文-通用领域-base模型为基础,在金融领域数据上微调得到,适用于金融领域FAQ问答任务,包括但不局限于:银行、保险等场景;MGIMN系列模型相对于StructBERT FAQ系列模型效果更优;",
- "hot": 223,
- "pic": "example.png",
- "uuid": "nlp-mgimn-faq-question-answering-chinese-finance-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "FAQ问答",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "nlp-space-dialog-state-tracking",
- "label": "SPACE对话状态追踪-英文-base",
- "describe": "该模型是 SPACE 基于一个对话状态跟踪数据集 MultiWOZ2.2 微调后的下游模型,称作 space_dialog-state-tracking,可专门用来做旅游、餐馆等领域的对话状态跟踪任务。",
- "hot": 210,
- "pic": "example.jpg",
- "uuid": "nlp-space-dialog-state-tracking",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "任务型对话",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "speech-ecapa-tdnn-sv-en-voxceleb-16k",
- "label": "ECAPATDNN说话人确认-英文-VoxCeleb-16k",
- "describe": "ECAPA TDNN模型是说话人识别领域的常用模型之一,该模型在公开数据集VoxCeleb2上进行训练,VoxCeleb-O上的EER指标为0.862。该模型适用于16k英文测试数据,可以用于说话人确认、说话人日志等任务。",
- "hot": 195,
- "pic": "example.png",
- "uuid": "speech-ecapa-tdnn-sv-en-voxceleb-16k",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "说话人确认",
- "field": "听觉"
- },
- {
- "price": "1",
- "name": "cv-cartoon-stable-diffusion-illustration",
- "label": "卡通系列文生图模型-漫画风",
- "describe": "",
- "hot": 186,
- "pic": "example.jpg",
- "uuid": "cv-cartoon-stable-diffusion-illustration",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "ofa-image-caption-coco-6b-en",
- "label": "OFA图像描述-英文-通用领域-6B",
- "describe": "根据用户输入的任意图片,AI智能创作模型写出“一句话描述”,可用于图像标签和图像简介。",
- "hot": 176,
- "pic": "example.jpg",
- "uuid": "ofa-image-caption-coco-6b-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像描述",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-manual-face-recognition-frir",
- "label": "IR人脸识别模型FRIR",
- "describe": "IR人脸识别模型FRIR",
- "hot": 162,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-recognition-frir",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "mplug-image-text-retrieval-flickr30k-large-en",
- "label": "mPLUG图文检索模型-英文-large",
- "describe": "达摩MPLUG英文图文检索large模型",
- "hot": 162,
- "pic": "example.jpg",
- "uuid": "mplug-image-text-retrieval-flickr30k-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图文检索",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "mo-di-diffusion",
- "label": "迪士尼风格扩散生成模型",
- "describe": "在迪士尼动画电影截图上finetune的Stable Diffusion模型",
- "hot": 156,
- "pic": "example.jpg",
- "uuid": "mo-di-diffusion",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "nlp-translation-quality-estimation-multilingual",
- "label": "QEMind翻译质量评估-多语言-通用领域",
- "describe": "对翻译质量进行打分评估的模型,支持多个语向,获WMT 2021世界机器翻译大赛质量评估DA子任务冠军",
- "hot": 154,
- "pic": "example.jpg",
- "uuid": "nlp-translation-quality-estimation-multilingual",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "句子相似度",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "analog-diffusion",
- "label": "胶片质感扩散生成模型",
- "describe": "输出胶片摄影质感的图像;微调自Stable Diffusion",
- "hot": 150,
- "pic": "example.jpg",
- "uuid": "analog-diffusion",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "multilingual-glm-summarization-en",
- "label": "mGLM多语言大模型-生成式摘要-英文",
- "describe": "mGLM多语言大模型可从大段文本中提取关键信息,为你生成简短的英文摘要,支持多种语言输入",
- "hot": 146,
- "pic": "example.png",
- "uuid": "multilingual-glm-summarization-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本摘要",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-swinb-video-panoptic-segmentation-vipseg",
- "label": "视频全景分割-VideoKNet-SwinB",
- "describe": "基于Video-K-Net架构,SwinB作为backbone的视频全景分割模型。",
- "hot": 145,
- "pic": "example.jpg",
- "uuid": "cv-swinb-video-panoptic-segmentation-vipseg",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频全景分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-large-en",
- "label": "OFA预训练模型-英文-通用领域-large",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 141,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-large-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-quadtree-attention-image-matching-outdoor",
- "label": "",
- "describe": "",
- "hot": 141,
- "pic": "example.jpg",
- "uuid": "cv-quadtree-attention-image-matching-outdoor",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像匹配",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "codegeex-code-translation-13b",
- "label": "CodeGeeX-代码翻译-13B",
- "describe": "CodeGeeX是一个具有130亿参数的多编程语言代码生成预训练模型,在20多种编程语言的代码语料库(>8500亿Token)上经过历时两个月预训练得到。CodeGeeX采用华为MindSpore框架实现,在鹏城实验室的“鹏城云脑ll”平台上训练而成。",
- "hot": 141,
- "pic": "example.jpg",
- "uuid": "codegeex-code-translation-13b",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "uni-fold-multimer",
- "label": "Uni-Fold-Multimer 开源的蛋白质复合物结构预测模型",
- "describe": "一个开源的蛋白质复合物结构预测模型。",
- "hot": 140,
- "pic": "example.jpg",
- "uuid": "uni-fold-multimer",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "蛋白质结构生成",
- "field": "未知"
- },
- {
- "price": "1",
- "name": "cv-manual-face-recognition-frfm",
- "label": "口罩人脸识别模型FRFM-large",
- "describe": "口罩人脸识别模型FRFM-large",
- "hot": 131,
- "pic": "example.jpg",
- "uuid": "cv-manual-face-recognition-frfm",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "人脸识别",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-huge-en",
- "label": "OFA预训练模型-英文-通用领域-huge",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 130,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-huge-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "anything-v4-0",
- "label": "二次元风格生成扩散模型-anything-v4.0",
- "describe": "为日本动漫爱好者设计的latent diffusion模型。",
- "hot": 129,
- "pic": "example.jpg",
- "uuid": "anything-v4-0",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-large-zh",
- "label": "OFA预训练模型-中文-通用领域-large",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 117,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-large-zh",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-unifuse-panorama-depth-estimation",
- "label": "基于单向融合的全景图深度估计",
- "describe": "单目全景图的深度估计",
- "hot": 115,
- "pic": "example.jpg",
- "uuid": "cv-unifuse-panorama-depth-estimation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "全景图深度估计",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-cartoon-stable-diffusion-watercolor",
- "label": "卡通系列文生图模型-水彩风",
- "describe": "",
- "hot": 107,
- "pic": "example.jpg",
- "uuid": "cv-cartoon-stable-diffusion-watercolor",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "price": "1",
- "name": "ofa-pretrain-medium-en",
- "label": "OFA预训练模型-英文-通用领域-medium",
- "describe": "OFA的预训练ckpt,能够在完全不改变模型结构的情况下进行下游任务的finetune,是finetune的基础ckpt。",
- "hot": 106,
- "pic": "example.jpg",
- "uuid": "ofa-pretrain-medium-en",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "",
- "field": "多模态"
- },
- {
- "price": "1",
- "name": "cv-rdevos-video-object-segmentation",
- "label": "基于循环动态编码的视频目标分割",
- "describe": "给定一个视频帧序列,和视频第一帧中想要分割的不同物体的掩码(mask),模型会预测视频后续帧中对应物体的掩码(mask)",
- "hot": 104,
- "pic": "example.gif",
- "uuid": "cv-rdevos-video-object-segmentation",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "视频目标分割",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "nlp-ponet-fill-mask-english-base",
- "label": "PoNet预训练模型-英文-base",
- "describe": "nlp_ponet_fill-mask_english-base是用bookcorpus/wikitext训练的预训练PoNet模型。",
- "hot": 104,
- "pic": "example.jpg",
- "uuid": "nlp-ponet-fill-mask-english-base",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "预训练",
- "field": "自然语言"
- },
- {
- "price": "1",
- "name": "cv-adaint-image-color-enhance-models",
- "label": "Adaptive-Interval-3DLUT图像调色",
- "describe": "",
- "hot": 104,
- "pic": "example.jpg",
- "uuid": "cv-adaint-image-color-enhance-models",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "图像颜色增强",
- "field": "机器视觉"
- },
- {
- "price": "1",
- "name": "cv-cartoon-stable-diffusion-flat",
- "label": "卡通系列文生图模型-扁平风",
- "describe": "",
- "hot": 104,
- "pic": "example.jpg",
- "uuid": "cv-cartoon-stable-diffusion-flat",
- "status": "online",
- "type": "dateset,notebook,train,inference",
- "version": "v20221001",
- "scenes": "文本生成图片",
- "field": "大模型"
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/chatglm2-6b",
- "field": "大模型",
- "scenes": "",
- "frameworks": "pytorch",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "chatglm2-6b",
- "status": "online",
- "version": "v20221001",
- "uuid": "chatglm2-6b-v20221001",
- "label": "chatglm2-6b大模型",
- "describe": "ChatGLM2-6B 是开源中英双语对话模型 ChatGLM-6B 的第二代版本",
- "pic": "example.png",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "resource_gpu": "1"
- },
- "inference": {
- "resource_memory": "7G",
- "resource_cpu": "43",
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/chatglm2-6b-32k",
- "field": "大模型",
- "scenes": "",
- "frameworks": "pytorch",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "chatglm2-6b-32k",
- "status": "online",
- "version": "v20221001",
- "uuid": "chatglm2-6b-32k-v20221001",
- "label": "chatglm2-6b-32k大模型",
- "describe": "ChatGLM2-6B-32K在ChatGLM2-6B的基础上进一步强化了对于长文本的理解能力,能够更好的处理最多32K长度的上下文。具体地,我们基于位置插值(Positional Interpolation)的方法对位置编码进行了更新,并在对话阶段使用 32K 的上下文长度训练。在实际的使用中,如果您面临的上下文长度基本在 8K 以内,我们推荐使用ChatGLM2-6B;如果您需要处理超过 8K 的上下文长度,我们推荐使用ChatGLM2-6B-32K",
- "pic": "example.png",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "resource_gpu": "1"
- },
- "inference": {
- "resource_memory": "7G",
- "resource_cpu": "43",
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/qwen-7b-chat",
- "field": "大模型",
- "scenes": "",
- "frameworks": "pytorch",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "qwen-7b-chat",
- "status": "online",
- "version": "v20221001",
- "uuid": "qwen-7b-chat-v20221001",
- "label": "通义千问-7B-Chat",
- "describe": "通义千问-7B(Qwen-7B) 是阿里云研发的通义千问大模型系列的70亿参数规模的模型。Qwen-7B是基于Transformer的大语言模型, 在超大规模的预训练数据上进行训练得到。预训练数据类型多样,覆盖广泛,包括大量网络文本、专业书籍、代码等。同时,在Qwen-7B的基础上,我们使用对齐机制打造了基于大语言模型的AI助手Qwen-7B-Chat",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "resource_gpu": "1"
- },
- "inference": {
- "resource_memory": "7G",
- "resource_cpu": "43",
- "resource_gpu": "1"
- }
- },
- {
- "doc": "https://github.com/tencentmusic/cube-studio/tree/master/aihub/deep-learning/chat-llama2",
- "field": "大模型",
- "scenes": "",
- "frameworks": "pytorch",
- "type": "dateset,notebook,train,evaluate,inference,web",
- "name": "chat-llama2",
- "status": "online",
- "version": "v20221001",
- "uuid": "chat-llama2-v20221001",
- "label": "llama2 模型",
- "describe": "llama2 模型",
- "pic": "example.jpg",
- "hot": "1",
- "price": "1",
- "dataset": {},
- "notebook": {
- "jupyter": [],
- "appendix": []
- },
- "train": {
- "resource_gpu": "1"
- },
- "inference": {
- "resource_memory": "30G",
- "resource_cpu": "10",
- "resource_gpu": "1"
- }
- }
-]
diff --git a/myapp/init-automl.json b/myapp/init-automl.json
deleted file mode 100644
index bd603103..00000000
--- a/myapp/init-automl.json
+++ /dev/null
@@ -1,29 +0,0 @@
-[
- {
- "job_type": "Job",
- "project": "public",
- "name": "test",
- "namespace": "automl",
- "describe": "nni功能测试",
- "parallel_trial_count": 3,
- "max_trial_count": 12,
- "objective_type": "maximize",
- "objective_goal": 0.99,
- "objective_metric_name": "accuracy",
- "algorithm_name": "Random",
- "parameters": {
- "batch_size": {"_type":"choice", "_value": [16, 32, 64, 128]},
- "momentum":{"_type":"uniform","_value":[0, 1]}
- },
- "job_json": {
-
- },
- "job_worker_image": "ccr.ccs.tencentyun.com/cube-studio/nni:20230601",
- "working_dir": "/mnt/admin/pipeline/example/nni/",
- "job_worker_command": "python demo.py",
- "resource_memory": "1G",
- "resource_cpu": "1",
- "resource_gpu": "0"
- }
-]
-
diff --git a/myapp/init-dataset.csv b/myapp/init-dataset.csv
deleted file mode 100644
index f4d78819..00000000
--- a/myapp/init-dataset.csv
+++ /dev/null
@@ -1,9 +0,0 @@
-name,label,describe,source_type,source,industry,field,usage,research,storage_class,file_type,status,years,url,path,download_url,storage_size,entries_num,duration,price,status,icon,owner
-MNIST,手写数字数据集,"包含一组60,000个示例的训练集和一组10,000个示例的测试集。数字已经过尺寸标准化,以适合 20x20 像素框,同时保持其纵横比,并在固定尺寸的图像中居中",开源,github,图像处理,视觉,传统机器学习和深度学习入门,svm、分类,压缩,gz,正常,,http://yann.lecun.com/exdb/mnist/,,"http://yann.lecun.com/exdb/mnist/train-images-idx3-ubyte.gz
-http://yann.lecun.com/exdb/mnist/train-labels-idx1-ubyte.gz
-http://yann.lecun.com/exdb/mnist/t10k-images-idx3-ubyte.gz
-http://yann.lecun.com/exdb/mnist/t10k-labels-idx1-ubyte.gz",11M," 60,000 个示例的训练集和 10,000 个示例的测试集",,0,正常,/static/assets/images/dataset/mnist.png,admin
-Fashion-MNIST,时尚产品数据,"包含60,000个训练图像和10,000个测试图像。类似MNIST的时尚产品数据库。",开源,github,图像处理,视觉,传统机器学习和深度学习入门,图像分类,压缩,gz,正常,,https://github.com/zalandoresearch/fashion-mnist,,"http://fashion-mnist.s3-website.eu-central-1.amazonaws.com/train-images-idx3-ubyte.gz
-http://fashion-mnist.s3-website.eu-central-1.amazonaws.com/train-labels-idx1-ubyte.gz
-http://fashion-mnist.s3-website.eu-central-1.amazonaws.com/t10k-images-idx3-ubyte.gz
-http://fashion-mnist.s3-website.eu-central-1.amazonaws.com/t10k-labels-idx1-ubyte.gz",5M,"60,000个训练图像和10,000个测试图像",,0,正常,/static/assets/images/dataset/fashion-mnist.png,admin
diff --git a/myapp/init-etl-pipeline.json b/myapp/init-etl-pipeline.json
deleted file mode 100644
index 9be9611d..00000000
--- a/myapp/init-etl-pipeline.json
+++ /dev/null
@@ -1,1708 +0,0 @@
-[
- {
- "name": "dau",
- "describe": "dau计算",
- "config": {
- "alert_user": "admin"
- },
- "workflow": "airflow",
- "dag_json": {
- "cos导入hdfs-1686184253953": {
- "label": "数据导入",
- "location": [
- 304,
- 96
- ],
- "color": {
- "color": "rgba(0,170,200,1)",
- "bg": "rgba(0,170,200,0.02)"
- },
- "template": "cos导入hdfs",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "hdfsPath": {
- "type": "str",
- "item_type": "str",
- "label": "hdfs文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "hdfs://xx/xxx",
- "placeholder": "",
- "describe": "目标hdfs文件路径,不包括文件名,支持${YYYYMMDD}等的日期变量。如果没hdfs路径权限,先联系平台管理员。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "cosPath": {
- "type": "str",
- "item_type": "str",
- "label": "源cos文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/xx/${YYYYMMDD}.tar.gz",
- "placeholder": "",
- "describe": "源cos文件路径,需包括文件名,支持${YYYYMMDD}等的日期变量。如果有多个文件上传,先打成一个.tar.gz压缩包。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "ifNeedZip": {
- "type": "str",
- "item_type": "str",
- "label": "是否需要解压缩",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "是否需要解压缩 {0:不需要,1:需要}。解压方式为tar zcvf。解压后文件会放在目标文件夹。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "出库入库",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "hdfsPath": "hdfs://xx/xxx",
- "cosPath": "/xx/${YYYYMMDD}.tar.gz",
- "ifNeedZip": "1",
- "label": "数据导入"
- },
- "upstream": [],
- "task_id": 1
- },
- "hdfs入库至hive-1686184263002": {
- "label": "数据入库",
- "location": [
- 304,
- 224
- ],
- "color": {
- "color": "rgba(0,170,200,1)",
- "bg": "rgba(0,170,200,0.02)"
- },
- "template": "hdfs入库至hive",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "charSet": {
- "type": "str",
- "item_type": "str",
- "label": "源文件字符集",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "UTF-8",
- "placeholder": "",
- "describe": "源文件字符集",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "databaseName": {
- "type": "str",
- "item_type": "str",
- "label": "hive数据库名称",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "hive数据库名称",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "tableName": {
- "type": "str",
- "item_type": "str",
- "label": "hive表名",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "hive表名",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "delimiter": {
- "type": "str",
- "item_type": "str",
- "label": "源文件分隔符, 填ascii码",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "9",
- "placeholder": "",
- "describe": "默认TAB,ascii码:9",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "failedOnZeroWrited": {
- "type": "str",
- "item_type": "str",
- "label": "入库为空时任务处理",
- "require": 1,
- "choice": [
- "1",
- "0"
- ],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "无源文件或入库记录为0时,可以指定任务为成功(0)或失败(1),默认失败(1)",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "partitionType": {
- "type": "str",
- "item_type": "str",
- "label": "分区格式",
- "require": 1,
- "choice": [
- "P_${YYYYMM}",
- "P_${YYYYMMDD}",
- "P_${YYYYMMDDHH}",
- "NULL"
- ],
- "range": "",
- "default": "P_${YYYYMMDDHH}",
- "placeholder": "",
- "describe": "分区格式:P_${YYYYMM}、P_${YYYYMMDD}、P_${YYYYMMDDHH}、NULL",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "sourceFilePath": {
- "type": "str",
- "item_type": "str",
- "label": "数据文件hdfs路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "支持三种日期变量:${YYYYMM}、${YYYYMMDD}、${YYYYMMDDHH}。系统用任务实例的数据时间替换日期变量。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "sourceFileNames": {
- "type": "str",
- "item_type": "str",
- "label": "源文件名",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "*",
- "placeholder": "",
- "describe": "源文件名(支持通配符*和${YYYYMMDD});入库不做检查",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "sourceColumnNames": {
- "type": "str",
- "item_type": "str",
- "label": "源文件的栏位名称",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "源文件的栏位名称,以逗号分割(结尾不能是逗号),必须保证列数和文件内容一致(创建临时表所用表列名)。例如column1,column2,column3。注:不允许输入空格,源文件栏位名称只由大小写字符、数字和下划线组成",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "targetColumnNames": {
- "type": "str",
- "item_type": "str",
- "label": "字段映射关系,即hive表的列名",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "字段映射关系,即hive表的列名。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "loadMode": {
- "type": "str",
- "item_type": "str",
- "label": "数据入库模式",
- "require": 1,
- "choice": [
- "TRUNCATE",
- "APPEND"
- ],
- "range": "",
- "default": "TRUNCATE",
- "placeholder": "",
- "describe": "数据入库模式,TRUNCATE或APPEND;",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "出库入库",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "charSet": "UTF-8",
- "databaseName": "",
- "tableName": "",
- "delimiter": "9",
- "failedOnZeroWrited": "1",
- "partitionType": "P_${YYYYMMDDHH}",
- "sourceFilePath": "",
- "sourceFileNames": "*",
- "sourceColumnNames": "",
- "targetColumnNames": "",
- "loadMode": "TRUNCATE",
- "label": "数据入库"
- },
- "upstream": [
- "cos导入hdfs-1686184253953"
- ],
- "task_id": 2
- },
- "SQL-1686184276800": {
- "label": "局部特征计算",
- "location": [
- -16,
- 352
- ],
- "color": {
- "color": "rgba(0,200,153,1)",
- "bg": "rgba(0,200,153,0.02)"
- },
- "template": "SQL",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "filterSQL": {
- "type": "text",
- "item_type": "sql",
- "label": "计算加工逻辑",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "\n --库名,替换下面的demo_database\n use demo_database;\n\n --建表语句,替换下面的demo_table,修改字段。一定要加“if not exists”,这样使只在第一次运行时建表\n CREATE TABLE if not exists demo_table(\n qimei36 STRING COMMENT '唯一设备ID',\n userid_id STRING COMMENT '用户id(各app的用户id)',\n device_id STRING COMMENT '设备id(各app的device_id)',\n ftime INT COMMENT '数据分区时间 格式:yyyymmdd'\n )\n PARTITION BY LIST( ftime ) --定义分区字段,替换掉ftime。\n (\n PARTITION p_20220323 VALUES IN ( 20220323 ), --初始分区,分区名替换p_20220323,分区值替换20220323\n PARTITION default\n )\n STORED AS ORCFILE COMPRESS;\n\n -- 分区,根据时间参数新建分区。\n alter table demo_table drop partition (p_${YYYYMMDD});\n alter table demo_table add partition p_${YYYYMMDD} values in (${YYYYMMDD});\n\n -- 写入,用你的sql逻辑替换。\n insert table demo_table\n select * from other_db::other_table partition(p_${YYYYMMDD}) t;\n ",
- "placeholder": "",
- "describe": "从hive导出数据的sql,比如 select a,b,c FROM table where imp_date='${YYYYMMDD}' ;sql末尾不要用分号结尾",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "special_para": {
- "type": "str",
- "item_type": "str",
- "label": "hive特殊参数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "set hive.exec.parallel = true;set hive.execute.engine=spark;set hive.multi.join.use.hive=false;set hive.spark.failed.retry=false;",
- "placeholder": "",
- "describe": "hive特殊参数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "数据计算",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "filterSQL": "\n --库名,替换下面的demo_database\n use demo_database;\n\n --建表语句,替换下面的demo_table,修改字段。一定要加“if not exists”,这样使只在第一次运行时建表\n CREATE TABLE if not exists demo_table(\n qimei36 STRING COMMENT '唯一设备ID',\n userid_id STRING COMMENT '用户id(各app的用户id)',\n device_id STRING COMMENT '设备id(各app的device_id)',\n ftime INT COMMENT '数据分区时间 格式:yyyymmdd'\n )\n PARTITION BY LIST( ftime ) --定义分区字段,替换掉ftime。\n (\n PARTITION p_20220323 VALUES IN ( 20220323 ), --初始分区,分区名替换p_20220323,分区值替换20220323\n PARTITION default\n )\n STORED AS ORCFILE COMPRESS;\n\n -- 分区,根据时间参数新建分区。\n alter table demo_table drop partition (p_${YYYYMMDD});\n alter table demo_table add partition p_${YYYYMMDD} values in (${YYYYMMDD});\n\n -- 写入,用你的sql逻辑替换。\n insert table demo_table\n select * from other_db::other_table partition(p_${YYYYMMDD}) t;\n ",
- "special_para": "set hive.exec.parallel = true;set hive.execute.engine=spark;set hive.multi.join.use.hive=false;set hive.spark.failed.retry=false;",
- "label": "局部特征计算"
- },
- "upstream": [
- "hdfs入库至hive-1686184263002"
- ],
- "task_id": 3
- },
- "SparkScala-1686184279367": {
- "label": "局部特征计算",
- "location": [
- 304,
- 352
- ],
- "color": {
- "color": "rgba(0,200,153,1)",
- "bg": "rgba(0,200,153,0.02)"
- },
- "template": "SparkScala",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "jar_path": {
- "type": "text",
- "item_type": "str",
- "label": "jar包路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "填写jar包在notebook里的路径,示例/mnt/admin/pipeline_test.py",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "className": {
- "type": "str",
- "item_type": "str",
- "label": "jar包中主类的名字",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "jar包中主类的名字",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "files": {
- "type": "text",
- "item_type": "str",
- "label": "资源文件列表(--files or --archieves)",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "暂未支持",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "programSpecificParams": {
- "type": "text",
- "item_type": "str",
- "label": "传递给程序的参数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "传递给程序的参数,空格分隔,不要换行",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "options": {
- "type": "text",
- "item_type": "str",
- "label": "spark扩展参数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "\n 选项(spark-submit的--conf参数)。不带分号,使用换行分隔(例如):\n spark.driver.maxResultSize=15G\n spark.driver.cores=4\n spark支持一系列--conf扩展属性,此处可以直接填写。例如:spark.yarn.am.waitTime=100s。\n 提交任务时后台会将参数带上提交。换行分隔!!\n ",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "dynamicAllocation": {
- "type": "choice",
- "item_type": "str",
- "label": "是否动态资源分配",
- "require": 1,
- "choice": [
- "1",
- "0"
- ],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "是否动态资源分配,是:1;否:0",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "driver_memory": {
- "type": "str",
- "item_type": "str",
- "label": "driver内存大小",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2g",
- "placeholder": "",
- "describe": "driver内存大小",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "num_executors": {
- "type": "str",
- "item_type": "str",
- "label": "executor数量",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "4",
- "placeholder": "",
- "describe": "executor数量",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "executor_memory": {
- "type": "str",
- "item_type": "str",
- "label": "executor内存大小",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2g",
- "placeholder": "",
- "describe": "executor内存大小",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "executor_cores": {
- "type": "str",
- "item_type": "str",
- "label": "executor核心数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2",
- "placeholder": "",
- "describe": "executor核心数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "task.main.timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时时间,单位分钟",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "480",
- "placeholder": "",
- "describe": "超时时间,单位分钟:480 (代表8小时)",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "task.check.timeout": {
- "type": "str",
- "item_type": "str",
- "label": "check超时时间,单位分钟",
- "require": 1,
- "choice": [
- "5",
- "10"
- ],
- "range": "",
- "default": "5",
- "placeholder": "",
- "describe": "check超时时间,单位分钟",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "数据计算",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "jar_path": "",
- "className": "",
- "files": "",
- "programSpecificParams": "",
- "options": "",
- "dynamicAllocation": "1",
- "driver_memory": "2g",
- "num_executors": "4",
- "executor_memory": "2g",
- "executor_cores": "2",
- "task.main.timeout": "480",
- "task.check.timeout": "5",
- "label": "局部特征计算"
- },
- "upstream": [
- "hdfs入库至hive-1686184263002"
- ],
- "task_id": 4
- },
- "pyspark-1686184281148": {
- "label": "局部特征计算",
- "location": [
- 608,
- 352
- ],
- "color": {
- "color": "rgba(0,200,153,1)",
- "bg": "rgba(0,200,153,0.02)"
- },
- "template": "pyspark",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "py_script_path": {
- "type": "str",
- "item_type": "str",
- "label": "pyspark脚本路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "填写pyspark脚本在notebook里的路径,示例/mnt/admin/pipeline_test.py",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "files": {
- "type": "str",
- "item_type": "str",
- "label": "资源文件列表",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "暂未支持",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "pyFiles": {
- "type": "str",
- "item_type": "str",
- "label": "执行脚本依赖文件列表(spark-submit中的--py-files)",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "暂未支持",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "programSpecificParams": {
- "type": "text",
- "item_type": "str",
- "label": "传递给程序的参数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "传递给程序的参数,空格分隔,不要换行",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "options": {
- "type": "text",
- "item_type": "str",
- "label": "spark扩展参数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "选项(spark-submit的--conf参数)。例如:spark.yarn.am.waitTime=100s。换行分隔!!",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "dynamicAllocation": {
- "type": "str",
- "item_type": "str",
- "label": "是否动态资源分配",
- "require": 1,
- "choice": [
- 1,
- 0
- ],
- "range": "",
- "default": 1,
- "placeholder": "",
- "describe": "是否动态资源分配,是:1;否:0",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "driver_memory": {
- "type": "str",
- "item_type": "str",
- "label": "driver内存大小",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2g",
- "placeholder": "",
- "describe": "driver内存大小",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "num_executors": {
- "type": "str",
- "item_type": "str",
- "label": "executor数量",
- "require": 1,
- "choice": [],
- "range": "",
- "default": 4,
- "placeholder": "",
- "describe": "executor数量",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "executor_memory": {
- "type": "str",
- "item_type": "str",
- "label": "executor内存大小",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2g",
- "placeholder": "",
- "describe": "executor内存大小",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "executor_cores": {
- "type": "str",
- "item_type": "str",
- "label": "executor核心数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": 2,
- "placeholder": "",
- "describe": "executor核心数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "task.main.timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时时间,单位分钟",
- "require": 1,
- "choice": [],
- "range": "",
- "default": 480,
- "placeholder": "",
- "describe": "超时时间,单位分钟:480 (代表8小时)",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "task.check.timeout": {
- "type": "int",
- "item_type": "int",
- "label": "check超时时间,单位分钟",
- "require": 1,
- "choice": [
- "5",
- "10"
- ],
- "range": "",
- "default": "5",
- "placeholder": "",
- "describe": "check超时时间,单位分钟",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "数据计算",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "py_script_path": "",
- "files": "",
- "pyFiles": "",
- "programSpecificParams": "",
- "options": "",
- "dynamicAllocation": 1,
- "driver_memory": "2g",
- "num_executors": 4,
- "executor_memory": "2g",
- "executor_cores": 2,
- "task.main.timeout": 480,
- "task.check.timeout": "5",
- "label": "局部特征计算"
- },
- "upstream": [
- "hdfs入库至hive-1686184263002"
- ],
- "task_id": 5
- },
- "hive出库至hdfs-1686184293917": {
- "label": "结果计算",
- "location": [
- 304,
- 496
- ],
- "color": {
- "color": "rgba(0,170,200,1)",
- "bg": "rgba(0,170,200,0.02)"
- },
- "template": "hive出库至hdfs",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "databaseName": {
- "type": "str",
- "item_type": "str",
- "label": "hive表所在的database",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "hive表所在的database",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "destCheckFileName": {
- "type": "str",
- "item_type": "str",
- "label": "对账文件名",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "对账文件名",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "destCheckFilePath": {
- "type": "str",
- "item_type": "str",
- "label": "对账文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "对账文件路径",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "destFileDelimiter": {
- "type": "str",
- "item_type": "str",
- "label": "出库文件分隔符",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "9",
- "placeholder": "",
- "describe": "出库文件分隔符,填ascii字符对应的数字。默认TAB:9",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "destFilePath": {
- "type": "str",
- "item_type": "str",
- "label": "出库文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "出库文件路径",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "filterSQL": {
- "type": "text",
- "item_type": "sql",
- "label": "源SQL",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "select t1,t2,t3 from your_table where imp_date=${YYYYMMDD}",
- "placeholder": "",
- "describe": "源SQL",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "出库入库",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "databaseName": "",
- "destCheckFileName": "",
- "destCheckFilePath": "",
- "destFileDelimiter": "9",
- "destFilePath": "",
- "filterSQL": "select t1,t2,t3 from your_table where imp_date=${YYYYMMDD}",
- "label": "结果计算"
- },
- "upstream": [
- "SQL-1686184276800",
- "pyspark-1686184281148",
- "SparkScala-1686184279367"
- ],
- "task_id": 6
- },
- "hdfs导入cos-1686184296749": {
- "label": "数据导出",
- "location": [
- 304,
- 608
- ],
- "color": {
- "color": "rgba(0,170,200,1)",
- "bg": "rgba(0,170,200,0.02)"
- },
- "template": "hdfs导入cos",
- "templte_common_ui_config": {
- "任务元数据": {
- "crontab": {
- "type": "str",
- "item_type": "str",
- "label": "调度周期",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "1 1 * * *",
- "placeholder": "",
- "describe": "周期任务的时间设定 * * * * * 一次性任务可不填写
表示为 minute hour day month week",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "selfDepend": {
- "type": "str",
- "item_type": "str",
- "label": "自依赖判断",
- "require": 1,
- "choice": [
- "自依赖",
- "单实例运行",
- "多实例运行"
- ],
- "range": "",
- "default": "单实例运行",
- "placeholder": "",
- "describe": "一个任务的多次调度实例之间是否要进行前后依赖",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "ResourceGroup": {
- "type": "str",
- "item_type": "str",
- "label": "队列",
- "require": 1,
- "choice": [
- "default",
- "queue1",
- "queue2"
- ],
- "range": "",
- "default": "default",
- "placeholder": "",
- "describe": "队列",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- },
- "监控配置": {
- "alert_user": {
- "type": "str",
- "item_type": "str",
- "label": "报警用户",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "admin,",
- "placeholder": "",
- "describe": "报警用户,逗号分隔",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "timeout": {
- "type": "str",
- "item_type": "str",
- "label": "超时中断",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "task运行时长限制,为0表示不限制(单位s)",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- },
- "retry": {
- "type": "str",
- "item_type": "str",
- "label": "重试次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "重试次数",
- "editable": 1,
- "addable": 0,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "templte_ui_config": {
- "参数": {
- "hdfsPath": {
- "type": "str",
- "item_type": "str",
- "label": "hdfs文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "hdfs://xx/xxx",
- "placeholder": "",
- "describe": "源hdfs文件路径,包括文件名,支持通配符*,支持${YYYYMMDD}等的日期变量。如果没hdfs路径权限,联系平台管理员。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "cosPath": {
- "type": "str",
- "item_type": "str",
- "label": "目标cos文件路径",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/xx/xx/${YYYYMMDD}.tar.gz",
- "placeholder": "",
- "describe": "目标cos文件路径,需包括文件名,支持${YYYYMMDD}等的日期变量,如果有多个文件上传,会在自动在cos文件名后面添加一个随机串。",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "ifNeedZip": {
- "type": "str",
- "item_type": "str",
- "label": "是否需要压缩",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "是否需要压缩 {0:不需要,1:需要}。压缩会压缩成单个文件。压缩方式为.tar.gz",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- },
- "template-group": "出库入库",
- "task-config": {
- "crontab": "1 1 * * *",
- "selfDepend": "单实例运行",
- "ResourceGroup": "default",
- "alert_user": "admin,",
- "timeout": "0",
- "retry": "0",
- "hdfsPath": "hdfs://xx/xxx",
- "cosPath": "/xx/xx/${YYYYMMDD}.tar.gz",
- "ifNeedZip": "1",
- "label": "数据导出"
- },
- "upstream": [
- "hive出库至hdfs-1686184293917"
- ],
- "task_id": 7
- }
- }
- }
-]
diff --git a/myapp/init-inference.json b/myapp/init-inference.json
deleted file mode 100644
index 68eb44d5..00000000
--- a/myapp/init-inference.json
+++ /dev/null
@@ -1,93 +0,0 @@
-{
- "tf-mnist": {
- "project_name": "public",
- "service_name": "mnist-202208011",
- "model_name": "mnist",
- "service_describe": "tf 图像分类",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/tfserving:2.3.4",
- "model_version": "v2022.08.01.1",
- "model_path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/tf-mnist.tar.gz",
- "service_type": "tfserving",
- "env": "TF_CPP_VMODULE=http_server=1\nTZ=Asia/Shanghai",
- "ports": "8501",
- "metrics": "8501:/metrics",
- "command": "wget https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/tf-mnist.tar.gz && tar -zxvf tf-mnist.tar.gz && mkdir -p /models/mnist/202207281/ && cp -r tf-mnist/* /models/mnist/202207281/ && /usr/bin/tf_serving_entrypoint.sh --model_config_file=/config/models.config --monitoring_config_file=/config/monitoring.config --platform_config_file=/config/platform.config",
- "health": "8501:/v1/models/mnist/versions/202207281/metadata",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt,kubeflow-archives(pvc):/archives",
- "resource_memory": "2G",
- "resource_cpu": "2",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/images/serving/tfserving/example"
- }
- },
- "pytorch-resnet50": {
- "project_name": "public",
- "service_name": "resnet50-202208012",
- "model_name": "resnet50",
- "service_describe": "pytorch 图像分类",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/torchserve:0.5.3-cpu",
- "model_version": "v2022.08.01.2",
- "model_path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50.mar",
- "service_type": "torch-server",
- "env": "",
- "ports": "8080,8081",
- "metrics": "8082:/metrics",
- "workdir": "/models",
- "command": "wget https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50.mar && mkdir -p /models && cp /config/* /models/ && cp resnet50.mar /models/ && torchserve --start --model-store /models --models resnet50=resnet50.mar --foreground --ts-config=/config/config.properties",
- "health": "8080:/ping",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt,kubeflow-archives(pvc):/archives",
- "resource_memory": "5G",
- "resource_cpu": "5",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/images/serving/torchserver/example"
- }
- },
- "torchscript-resnet50": {
- "project_name": "public",
- "service_name": "resnet50-202208013",
- "model_name": "resnet50",
- "service_describe": "torchscript 图像分类",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/tritonserver:22.07-py3",
- "model_version": "v2022.08.01.3",
- "model_path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50-torchscript.pt",
- "service_type": "triton-server",
- "env": "",
- "ports": "8000,8002",
- "metrics": "",
- "workdir": "",
- "command": "wget https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50-torchscript.pt && mkdir -p /models/resnet50/202208013/ && cp /config/* /models/resnet50/ && cp -r resnet50-torchscript.pt /models/resnet50/202208013/model.pt && tritonserver --model-repository=/models --strict-model-config=true --log-verbose=1",
- "health": "8000:/v2/health/ready",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt,kubeflow-archives(pvc):/archives",
- "resource_memory": "5G",
- "resource_cpu": "5",
- "inference_config": "\n---config.pbtxt\n\nname: \"resnet50\"\nplatform: \"pytorch_libtorch\"\nmax_batch_size: 0\ninput \n[\n {\n name: \"INPUT__0\"\n data_type: TYPE_FP32\n format: FORMAT_NCHW\n dims: [ 3, 224, 224 ]\n reshape: {\n shape: [ 1, 3, 224, 224 ]\n }\n }\n]\n \noutput \n[\n {\n name: \"OUTPUT__0\"\n data_type: TYPE_FP32\n dims: [ 1000 ]\n reshape: {\n shape: [ 1, 1000 ]\n }\n }\n]\n \n\nparameters: { key: \"DISABLE_OPTIMIZED_EXECUTION\" value: { string_value:\"true\" } }\nparameters: { key: \"INFERENCE_MODE\" value: { string_value: \"false\" } }\n",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/images/serving/triton-server/example"
- }
- },
- "onnx-resnet50": {
- "project_name": "public",
- "service_name": "resnet50-202208014",
- "model_name": "resnet50",
- "service_describe": "onnx 图像分类",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/tritonserver:22.07-py3",
- "model_version": "v2022.08.01.4",
- "model_path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50.onnx",
- "service_type": "triton-server",
- "env": "",
- "ports": "8000,8002",
- "metrics": "",
- "workdir": "",
- "command": "wget https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50.onnx && mkdir -p /models/resnet50/202208014/ && cp /config/* /models/resnet50/ && cp -r resnet50.onnx /models/resnet50/202208014/model.onnx && tritonserver --model-repository=/models --strict-model-config=true --log-verbose=1",
- "health": "8000:/v2/health/ready",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt,kubeflow-archives(pvc):/archives",
- "resource_memory": "5G",
- "resource_cpu": "5",
- "inference_config": "---config.pbtxt\n\nname: \"resnet50\"\nplatform: \"onnxruntime_onnx\"\nbackend: \"onnxruntime\"\nmax_batch_size : 0\n\ninput [\n {\n name: \"input_name\"\n data_type: TYPE_FP32\n format: FORMAT_NCHW\n dims: [ 3, 224, 224 ]\n reshape { shape: [ 1, 3, 224, 224 ] }\n }\n]\noutput [\n {\n name: \"output_name\"\n data_type: TYPE_FP32\n dims: [ 1000 ]\n reshape { shape: [ 1, 1000 ] }\n }\n]\n\nparameters { key: \"intra_op_thread_count\" value: { string_value: \"10\" } }\nparameters { key: \"execution_mode\" value: { string_value: \"1\" } }\nparameters { key: \"inter_op_thread_count\" value: { string_value: \"10\" } }\n",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/images/serving/triton-server/example"
- }
- }
-}
-
-
diff --git a/myapp/init-job-template.json b/myapp/init-job-template.json
deleted file mode 100644
index 266ce324..00000000
--- a/myapp/init-job-template.json
+++ /dev/null
@@ -1,3141 +0,0 @@
-{
- "自定义镜像": {
- "project_name": "基础命令",
- "image_name": "ubuntu:18.04",
- "gitpath": "",
- "image_describe": "开源ubuntu:18.04基础镜像",
- "job_template_name": "自定义镜像",
- "job_template_describe": "使用用户自定义镜像作为运行镜像",
- "job_template_expand": {
- "index": 1
- },
- "job_template_args": {
- "参数": {
- "images": {
- "type": "str",
- "item_type": "str",
- "label": "要调试的镜像",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "ccr.ccs.tencentyun.com/cube-studio/ubuntu-gpu:cuda11.8.0-cudnn8-python3.9",
- "placeholder": "",
- "describe": "要调试的镜像,基础镜像参考",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "workdir": {
- "type": "str",
- "item_type": "str",
- "label": "启动目录",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/mnt/xx",
- "placeholder": "",
- "describe": "启动目录",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "command": {
- "type": "str",
- "item_type": "str",
- "label": "启动命令",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "sh start.sh",
- "placeholder": "",
- "describe": "启动命令",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "logical": {
- "project_name": "基础命令",
- "image_name": "python:3.9",
- "gitpath": "",
- "image_describe": "python3.9",
- "job_template_name": "logical",
- "job_template_describe": "任务流量的逻辑节点(企业版)",
- "job_template_expand": {
- "index": 2
- },
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=1\nTASK_RESOURCE_MEMORY=1G\nTASK_RESOURCE_GPU=0",
- "job_template_args": {}
- },
- "python": {
- "project_name": "基础命令",
- "image_name": "python:3.9",
- "gitpath": "",
- "image_describe": "python3.9基础镜像",
- "job_template_name": "python",
- "job_template_describe": "python代码执行(企业版)",
- "job_template_expand": {
- "index": 3
- },
- "job_template_args": {}
- },
- "drop-duplicates": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "去除重复样本",
- "job_template_name": "drop-duplicates",
- "job_template_describe": "去除重复样本(企业版)",
- "job_template_command": "python3 launcher.py --process_type drop_duplicates",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "drop-missing": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "删除缺失率过高的值",
- "job_template_name": "drop-missing",
- "job_template_describe": "删除缺失率过高的值(企业版)",
- "job_template_command": "python3 launcher.py --process_type drop_high_missing",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "drop-stablize": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "去除值过于单一的变量",
- "job_template_name": "drop-stablize",
- "job_template_describe": "去除值过于单一的变量(企业版)",
- "job_template_command": "python3 launcher.py --process_type drop_stablize",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "fill-missing": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "填充缺失值",
- "job_template_name": "fill-missing",
- "job_template_describe": "填充缺失值(企业版)",
- "job_template_command": "python3 launcher.py --process_type fill_missing",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 4,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "one-hot": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "枚举类one hot展开",
- "job_template_name": "one-hot",
- "job_template_describe": "枚举类one hot展开(企业版)",
- "job_template_command": "python lunacher.py --process_type one_hot",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 5,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "calculate-metric": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "获取变量的统计量",
- "job_template_name": "calculate-metric",
- "job_template_describe": "获取变量的统计量(企业版)",
- "job_template_command": "python3 launcher.py --process_type calculate_metric",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 11,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "outlier-detection": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "异常值检测",
- "job_template_name": "outlier-detection",
- "job_template_describe": "异常值检测(企业版)",
- "job_template_command": "python launcher.py --process_type outlier_detection",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 13,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "pca": {
- "project_name": "数据预处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process",
- "image_describe": "pca降维",
- "job_template_name": "pca",
- "job_template_describe": "pca降维(企业版)",
- "job_template_command": "python launcher.py --process_type pca",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 14,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/data-process"
- },
- "job_template_args": {}
- },
- "calculate-correlation": {
- "project_name": "特征工程",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/feature-process",
- "image_describe": "计算列的相关性计算",
- "job_template_name": "calculate-correlation",
- "job_template_describe": "计算列的相关性计算(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/feature-process"
- },
- "job_template_args": {}
- },
-
- "datax": {
- "project_name": "数据导入导出",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/datax:20230601",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/datax",
- "image_describe": "datax异构数据源同步",
- "job_template_name": "datax",
- "job_template_describe": "datax异构数据源同步",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/datax"
- },
- "job_template_args": {
- "参数": {
- "-f": {
- "type": "str",
- "item_type": "str",
- "label": "job.json文件地址,书写格式参考",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/usr/local/datax/job/job.json",
- "placeholder": "",
- "describe": "job.json文件地址,书写格式参考",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "feature": {
- "project_name": "数据导入导出",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/feature:latest",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/feature",
- "image_describe": "特征导入",
- "job_template_name": "feature",
- "job_template_describe": "特征导入(todo)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/feature"
- },
- "job_template_args": {
- "参数": {
- "--features": {
- "type": "str",
- "item_type": "str",
- "label": "导入特征",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "导入特征,逗号分割要导入的特征",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "dataset": {
- "project_name": "数据导入导出",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/dataset:20230501",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/dataset",
- "image_describe": "数据集导入",
- "job_template_name": "dataset",
- "job_template_describe": "数据集导入",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/dataset"
- },
- "job_template_args": {
- "参数": {
- "--src_type": {
- "type": "str",
- "item_type": "str",
- "label": "数据集的来源",
- "require": 1,
- "choice": [
- "当前平台",
- "huggingface"
- ],
- "range": "",
- "default": "当前平台",
- "placeholder": "",
- "describe": "数据集的来源",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--name": {
- "type": "str",
- "item_type": "str",
- "label": "数据集的名称",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "数据集的名称",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--version": {
- "type": "str",
- "item_type": "str",
- "label": "数据集的版本",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "latest",
- "placeholder": "",
- "describe": "数据集的版本",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--partition": {
- "type": "str",
- "item_type": "str",
- "label": "数据集的分区,或者子数据集",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "数据集的分区,或者子数据集",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--save_dir": {
- "type": "str",
- "item_type": "str",
- "label": "数据集的保存地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "数据集的保存地址",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "model-download": {
- "project_name": "数据导入导出",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/model_download:20221001",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_download",
- "image_describe": "模型导入",
- "job_template_name": "model-download",
- "job_template_describe": "模型导入",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 4,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_download"
- },
- "job_template_args": {
- "参数": {
- "--from": {
- "type": "str",
- "item_type": "str",
- "label": "模型来源地",
- "require": 1,
- "choice": [
- "模型管理",
- "推理服务",
- "huggingface"
- ],
- "range": "",
- "default": "模型管理",
- "placeholder": "",
- "describe": "模型来源地",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_name": {
- "type": "str",
- "item_type": "str",
- "label": "模型名(a-z0-9-字符组成,最长54个字符)",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型名",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--sub_model_name": {
- "type": "str",
- "item_type": "str",
- "label": "子模型名(a-z0-9-字符组成,最长54个字符)",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "子模型名,对于包含多个子模型的用户填写",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_version": {
- "type": "str",
- "item_type": "str",
- "label": "模型版本号",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "v2022.10.01.1",
- "placeholder": "",
- "describe": "模型版本号",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_status": {
- "type": "str",
- "item_type": "str",
- "label": "模型状态",
- "require": 1,
- "choice": [
- "online",
- "offline",
- "test"
- ],
- "range": "",
- "default": "online",
- "placeholder": "",
- "describe": "模型状态,模型来自推理服务时有效",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--save_path": {
- "type": "str",
- "item_type": "str",
- "label": "下载目的目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/xx/download/model/",
- "placeholder": "",
- "describe": "下载目的目录",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "hadoop": {
- "project_name": "数据处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hadoop",
- "image_describe": "hadoop大数据组件客户端",
- "job_template_name": "hadoop",
- "job_template_describe": "hadoop大数据组件,hdfs,sqoop,spark(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_expand": {
- "index": 0,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hadoop",
- "HostNetwork": true
- },
- "job_template_env": "",
- "job_template_args": {}
- },
- "volcanojob": {
- "project_name": "数据处理",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/volcano:20211001",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/volcano",
- "image_describe": "有序分布式任务",
- "job_template_name": "volcanojob",
- "job_template_describe": "有序分布式任务",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/volcano"
- },
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_args": {
- "参数": {
- "--image": {
- "type": "str",
- "item_type": "str",
- "label": "",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "ccr.ccs.tencentyun.com/cube-studio/ubuntu-gpu:cuda11.8.0-cudnn8-python3.9",
- "placeholder": "",
- "describe": "worker镜像,直接运行你代码的环境镜像基础镜像",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--working_dir": {
- "type": "str",
- "item_type": "str",
- "label": "启动目录",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/mnt/xx",
- "placeholder": "",
- "describe": "启动目录",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--command": {
- "type": "str",
- "item_type": "str",
- "label": "启动命令",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "echo aa",
- "placeholder": "",
- "describe": "启动命令",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--num_worker": {
- "type": "str",
- "item_type": "str",
- "label": "占用机器个数",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "3",
- "placeholder": "",
- "describe": "占用机器个数",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "ray": {
- "project_name": "数据处理",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/ray:gpu-20210601",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/ray",
- "image_describe": "ray分布式任务",
- "job_template_name": "ray",
- "job_template_describe": "python多机分布式任务",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/ray"
- },
- "job_template_args": {
- "参数": {
- "-n": {
- "type": "str",
- "item_type": "",
- "label": "分布式任务worker的数量",
- "require": 1,
- "choice": [
- ],
- "range": "$min,$max",
- "default": "3",
- "placeholder": "",
- "describe": "分布式任务worker的数量",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "-i": {
- "type": "str",
- "item_type": "str",
- "label": "每个worker的初始化脚本文件地址,用来安装环境",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "每个worker的初始化脚本文件地址,用来安装环境",
- "describe": "每个worker的初始化脚本文件地址,用来安装环境",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "-f": {
- "type": "str",
- "item_type": "str",
- "label": "python启动命令,例如 python3 /mnt/xx/xx.py",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "python /mnt/admin/demo.py",
- "placeholder": "",
- "describe": "python启动命令,例如 python3 /mnt/xx/xx.py",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "sparkjob": {
- "project_name": "数据处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/spark",
- "image_describe": "spark serverless分布式任务",
- "job_template_name": "sparkjob",
- "job_template_describe": "spark serverless分布式任务(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/spark"
- },
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_args": {}
- },
- "ray-sklearn": {
- "project_name": "机器学习框架",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/ray_sklearn",
- "image_describe": "sklearn基于ray的分布式",
- "job_template_name": "ray-sklearn",
- "job_template_describe": "sklearn基于ray的分布式(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/ray_sklearn"
- },
- "job_template_args": {}
- },
- "adaboost": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/adaboost",
- "image_describe": "adaboost算法",
- "job_template_name": "adaboost",
- "job_template_describe": "adaboost算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/adaboost"
- },
- "job_template_args": {}
- },
- "xgb": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/xgb",
- "image_describe": "xgb算法单机",
- "job_template_name": "xgb",
- "job_template_describe": "xgb算法单机(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/xgb"
- },
- "job_template_args": {}
- },
- "bayesian": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/bayesian",
- "image_describe": "bayesian算法",
- "job_template_name": "bayesian",
- "job_template_describe": "bayesian算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/bayesian"
- },
- "job_template_args": {}
- },
- "crf": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/crf",
- "image_describe": "crf算法",
- "job_template_name": "crf",
- "job_template_describe": "crf算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/crf"
- },
- "job_template_args": {
- }
- },
- "catboost": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/catboost",
- "image_describe": "catboost算法",
- "job_template_name": "catboost",
- "job_template_describe": "catboost算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/catboost"
- },
- "job_template_args": {
- }
- },
- "fm": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/fm",
- "image_describe": "fm算法",
- "job_template_name": "fm",
- "job_template_describe": "fm算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/fm"
- },
- "job_template_args": {
- }
- },
- "hmm": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hmm",
- "image_describe": "hmm算法",
- "job_template_name": "hmm",
- "job_template_describe": "hmm算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hmm"
- },
- "job_template_args": {
- }
- },
- "lda": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/lda",
- "image_describe": "lda算法",
- "job_template_name": "lda",
- "job_template_describe": "lda算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/lda"
- },
- "job_template_args": {
- }
- },
- "mcmc": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mcmc",
- "image_describe": "mcmc算法",
- "job_template_name": "mcmc",
- "job_template_describe": "mcmc算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mcmc"
- },
- "job_template_args": {
- }
- },
- "svd": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/svd",
- "image_describe": "svd算法",
- "job_template_name": "svd",
- "job_template_describe": "svd算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/svd"
- },
- "job_template_args": {
- }
- },
- "svm": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/svm",
- "image_describe": "svm算法",
- "job_template_name": "svm",
- "job_template_describe": "svm算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/svm"
- },
- "job_template_args": {
- }
- },
- "decision-tree": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/decision-tree",
- "image_describe": "decision-tree算法",
- "job_template_name": "decision-tree",
- "job_template_describe": "decision-tree算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/decision-tree"
- },
- "job_template_args": {}
- },
- "gbdt": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/gbdt",
- "image_describe": "gbdt算法",
- "job_template_name": "gbdt",
- "job_template_describe": "gbdt算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/gbdt"
- },
- "job_template_args": {}
- },
- "hyperparam-search": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hyperparameter-search-ray",
- "image_describe": "超参搜索算法",
- "job_template_name": "hyperparam-search",
- "job_template_describe": "超参搜索算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hyperparameter-search-ray"
- },
- "job_template_args": {}
- },
- "kmean": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/kmean",
- "image_describe": "kmean算法",
- "job_template_name": "kmean",
- "job_template_describe": "kmean算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/kmean"
- },
- "job_template_args": {}
- },
- "knn": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/knn",
- "image_describe": "knn算法",
- "job_template_name": "knn",
- "job_template_describe": "knn算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/knn"
- },
- "job_template_args": {}
- },
- "lightgbm": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/lightgbm",
- "image_describe": "lightgbm算法",
- "job_template_name": "lightgbm",
- "job_template_describe": "lightgbm算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/lightgbm"
- },
- "job_template_args": {}
- },
- "lr": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/lr",
- "image_describe": "lr算法",
- "job_template_name": "lr",
- "job_template_describe": "lr算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/lr"
- },
- "job_template_args": {}
- },
- "random-forest": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/random-forest",
- "image_describe": "random-forest算法",
- "job_template_name": "random-forest",
- "job_template_describe": "random-forest算法(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/random-forest"
- },
- "job_template_args": {}
- },
- "apriori": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/apriori",
- "image_describe": "apriori算法",
- "job_template_name": "apriori",
- "job_template_describe": "apriori算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/apriori"
- },
- "job_template_args": {
- }
- },
- "mst": {
- "project_name": "机器学习算法",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mst",
- "image_describe": "mst算法",
- "job_template_name": "mst",
- "job_template_describe": "mst算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mst"
- },
- "job_template_args": {
- }
- },
- "tfjob": {
- "project_name": "深度学习",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/tf_distributed_train_k8s:20230801",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/tf_distributed_train_k8s",
- "image_describe": "tf 分布式训练",
- "job_template_name": "tfjob",
- "job_template_describe": "tf 分布式训练",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=4\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/tf_distributed_train_k8s"
- },
- "job_template_args": {
- "参数": {
- "--image": {
- "type": "str",
- "item_type": "str",
- "label": "worker镜像,直接运行你代码的环境镜像",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "kubeflow/tf-mnist-with-summaries:latest",
- "placeholder": "",
- "describe": "worker镜像,直接运行你代码的环境镜像 基础镜像",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--working_dir": {
- "type": "str",
- "item_type": "str",
- "label": "命令的启动目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/admin/pipeline/example/tf/",
- "placeholder": "",
- "describe": "命令的启动目录",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--command": {
- "type": "str",
- "item_type": "str",
- "label": "启动命令,例如 python3 xxx.py",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "python /var/tf_mnist/mnist_with_summaries.py",
- "placeholder": "启动命令,例如 python3 xxx.py",
- "describe": "启动命令,例如 python3 xxx.py",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--num_worker": {
- "type": "str",
- "item_type": "str",
- "label": "分布式训练worker的数目",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2",
- "placeholder": "分布式训练worker的数目",
- "describe": "分布式训练worker的数目",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "pytorchjob": {
- "project_name": "深度学习",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/pytorch:20201010",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/pytorch",
- "image_describe": "pytorch分布式训练",
- "job_template_name": "pytorchjob",
- "job_template_old_names": [
- "pytorchjob-train"
- ],
- "job_template_describe": "pytorch 分布式训练",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/pytorch"
- },
- "job_template_args": {
- "参数": {
- "--image": {
- "type": "str",
- "item_type": "str",
- "label": "worker镜像,直接运行你代码的环境镜像",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "ccr.ccs.tencentyun.com/cube-studio/ubuntu-gpu:cuda11.8.0-cudnn8-python3.9",
- "placeholder": "",
- "describe": "worker镜像,直接运行你代码的环境镜像 基础镜像",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--working_dir": {
- "type": "str",
- "item_type": "str",
- "label": "命令的启动目录",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/mnt/admin/pipeline/example/pytorch",
- "placeholder": "",
- "describe": "命令的启动目录",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--command": {
- "type": "str",
- "item_type": "str",
- "label": "启动命令,例如 python3 xxx.py",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "bash start.sh",
- "placeholder": "启动命令,例如 python3 xxx.py",
- "describe": "启动命令,例如 python3 xxx.py",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--num_worker": {
- "type": "str",
- "item_type": "str",
- "label": "分布式训练worker的数目",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "3",
- "placeholder": "分布式训练worker的数目",
- "describe": "分布式训练worker的数目",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "paddlejob": {
- "project_name": "深度学习",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/paddle",
- "image_describe": "paddle分布式训练",
- "job_template_name": "paddlejob",
- "job_template_describe": "paddle 分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 4,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/paddle"
- },
- "job_template_args": {}
- },
- "mindspore": {
- "project_name": "深度学习",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mindspore",
- "image_describe": "mindspore 分布式训练",
- "job_template_name": "mindspore",
- "job_template_describe": "mindspore 分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 5,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mindspore"
- },
- "job_template_args": {}
- },
- "mxnet": {
- "project_name": "深度学习",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mxnet",
- "image_describe": "mxnet分布式训练",
- "job_template_name": "mxnet",
- "job_template_describe": "mxnet 分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 6,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mxnet"
- },
- "job_template_args": {}
- },
- "kaldi": {
- "project_name": "深度学习",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/kaldi",
- "image_describe": "kaldi音频分布式",
- "job_template_name": "kaldi",
- "job_template_version": "Alpha",
- "job_template_old_names": [
- "kaldi"
- ],
- "job_template_describe": "kaldi音频分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=4\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 7,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/kaldi"
- },
- "job_template_args": {}
- },
- "horovod": {
- "project_name": "分布式加速",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/horovod:20230801",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/horovod",
- "image_describe": "horovod 分布式训练",
- "job_template_name": "horovod",
- "job_template_describe": "horovod 分布式训练",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/horovod"
- },
- "job_template_args": {
- "参数": {
- "--work_images": {
- "type": "str",
- "item_type": "str",
- "label": "worker的运行镜像,直接运行你代码的环境镜像 基础镜像",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "horovod/horovod:0.20.0-tf2.3.0-torch1.6.0-mxnet1.5.0-py3.7-cpu",
- "placeholder": "",
- "describe": "worker的运行镜像",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--working_dir": {
- "type": "str",
- "item_type": "str",
- "label": "命令的启动目录",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/mnt/admin/pipeline/example/horovod/",
- "placeholder": "",
- "describe": "命令的启动目录",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--command": {
- "type": "str",
- "item_type": "str",
- "label": "训练启动命令",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "python /examples/tensorflow2_mnist.py",
- "placeholder": "",
- "describe": "训练启动命令",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--num_worker": {
- "type": "str",
- "item_type": "str",
- "label": "分布式worker的数目",
- "require": 1,
- "choice": [],
- "range": "",
- "default": 2,
- "placeholder": "",
- "describe": "分布式worker的数目",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "deepspeed": {
- "project_name": "分布式加速",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/deepspeed",
- "image_describe": "deepspeed 分布式训练",
- "job_template_name": "deepspeed",
- "job_template_describe": "deepspeed 分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=2\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/deepspeed"
- },
- "job_template_args": {}
- },
- "hpc": {
- "project_name": "分布式加速",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hpc",
- "image_describe": "hpc高性能计算",
- "job_template_name": "hpc",
- "job_template_describe": "hpc高性能计算(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/hpc"
- },
- "job_template_args": {}
- },
- "colossalai": {
- "project_name": "分布式加速",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/colossalai",
- "image_describe": "colossalai 分布式训练",
- "job_template_name": "colossalai",
- "job_template_describe": "colossalai 分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/colossalai"
- },
- "job_template_args": {}
- },
- "mpi": {
- "project_name": "分布式加速",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mpi",
- "image_describe": "mpi 分布式训练",
- "job_template_name": "mpi",
- "job_template_describe": "mpi 分布式训练(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mpi"
- },
- "job_template_args": {}
- },
- "oneflow": {
- "project_name": "分布式加速",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/oneflow",
- "image_describe": "oneflow 分布式训练",
- "job_template_name": "oneflow",
- "job_template_describe": "oneflow 分布式训练(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/oneflow"
- },
- "job_template_args": {}
- },
- "din": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/din:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/din",
- "image_describe": "din算法",
- "job_template_name": "din",
- "job_template_describe": "din算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/din"
- },
- "job_template_args": {
- }
- },
- "comirec": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/comirec:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/comirec",
- "image_describe": "comirec算法",
- "job_template_name": "comirec",
- "job_template_describe": "comirec算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/comirec"
- },
- "job_template_args": {
- }
- },
- "mind": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/mind:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mind",
- "image_describe": "mind 算法",
- "job_template_name": "mind",
- "job_template_describe": "mind 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mind"
- },
- "job_template_args": {
- }
- },
- "mmoe": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/mmoe:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mmoe",
- "image_describe": "mmoe 算法",
- "job_template_name": "mmoe",
- "job_template_describe": "mmoe 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/mmoe"
- },
- "job_template_args": {
- }
- },
- "ple": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/ple:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/ple",
- "image_describe": "ple 算法",
- "job_template_name": "ple",
- "job_template_describe": "ple 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/ple"
- },
- "job_template_args": {
- }
- },
- "esmm": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/esmm:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/esmm",
- "image_describe": "esmm 算法",
- "job_template_name": "esmm",
- "job_template_describe": "esmm 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/esmm"
- },
- "job_template_args": {
- }
- },
- "dtower": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/dtower:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/dtower",
- "image_describe": "dtower 算法",
- "job_template_name": "dtower",
- "job_template_describe": "dtower 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/dtower"
- },
- "job_template_args": {
- }
- },
- "deepfm": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/deepfm:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/deepfm",
- "image_describe": "deepfm 算法",
- "job_template_name": "deepfm",
- "job_template_describe": "deepfm 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/deepfm"
- },
- "job_template_args": {
- }
- },
- "youtube": {
- "project_name": "推荐类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/youtube:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/youtube",
- "image_describe": "youtube 算法",
- "job_template_name": "youtube",
- "job_template_describe": "youtube 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/youtube"
- },
- "job_template_args": {
- }
- },
- "model-evaluation": {
- "project_name": "模型处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_evaluation",
- "image_describe": "模型评估",
- "job_template_name": "model-evaluation",
- "job_template_describe": "模型评估(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_evaluation"
- },
- "job_template_args": {}
- },
- "model-convert": {
- "project_name": "模型处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_convert",
- "image_describe": "模型转换",
- "job_template_name": "model-convert",
- "job_template_describe": "模型转换(企业版)",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_convert"
- },
- "job_template_args": {}
- },
- "model-compression": {
- "project_name": "模型处理",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_compression",
- "image_describe": "模型压缩",
- "job_template_name": "model-compression",
- "job_template_describe": "模型压缩(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_compression"
- },
- "job_template_args": {}
- },
- "model-register": {
- "project_name": "模型服务化",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/model_register:20230501",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_register",
- "image_describe": "注册模型",
- "job_template_name": "model-register",
- "job_template_describe": "注册模型",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_register"
- },
- "job_template_args": {
- "参数": {
- "--project_name": {
- "type": "str",
- "item_type": "str",
- "label": "部署项目名",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "public",
- "placeholder": "",
- "describe": "部署项目名",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_name": {
- "type": "str",
- "item_type": "str",
- "label": "模型名",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型名(a-z0-9-字符组成,最长54个字符)",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_version": {
- "type": "str",
- "item_type": "str",
- "label": "模型版本号",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "{{ datetime.datetime.now().strftime('v%Y.%m.%d.1') }}",
- "placeholder": "",
- "describe": "模型版本号",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_path": {
- "type": "str",
- "item_type": "str",
- "label": "模型地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型地址",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_metric": {
- "type": "str",
- "item_type": "str",
- "label": "模型指标",
- "require": 0,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型指标",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--describe": {
- "type": "str",
- "item_type": "str",
- "label": "模型描述",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型描述",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--framework": {
- "type": "str",
- "item_type": "str",
- "label": "模型框架",
- "require": 1,
- "choice": [
- "lr",
- "xgb",
- "tf",
- "pytorch",
- "onnx",
- "tensorrt",
- "aihub"
- ],
- "range": "",
- "default": "tf",
- "placeholder": "",
- "describe": "模型框架",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--inference_framework": {
- "type": "str",
- "item_type": "str",
- "label": "推理框架",
- "require": 1,
- "choice": [
- "sklearn",
- "tfserving",
- "torch-server",
- "onnxruntime",
- "triton-server",
- "aihub"
- ],
- "range": "",
- "default": "tfserving",
- "placeholder": "",
- "describe": "推理框架",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "model-offline-predict": {
- "project_name": "模型服务化",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/volcano:offline-predict-20220101",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_offline_predict",
- "image_describe": "分布式离线推理",
- "job_template_name": "model-offline-predict",
- "job_template_describe": "分布式离线推理",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "NO_RESOURCE_CHECK=true\nTASK_RESOURCE_CPU=4\nTASK_RESOURCE_MEMORY=4G\nTASK_RESOURCE_GPU=0",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/model_offline_predict"
- },
- "job_template_args": {
- "参数": {
- "--image": {
- "type": "str",
- "item_type": "str",
- "label": "",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "ccr.ccs.tencentyun.com/cube-studio/ubuntu-gpu:cuda11.8.0-cudnn8-python3.9",
- "placeholder": "",
- "describe": "worker镜像,直接运行你代码的环境镜像基础镜像",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--working_dir": {
- "type": "str",
- "item_type": "str",
- "label": "启动目录",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/mnt/xx",
- "placeholder": "",
- "describe": "启动目录",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--command": {
- "type": "str",
- "item_type": "str",
- "label": "环境安装和任务启动命令",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "/mnt/xx/../start.sh",
- "placeholder": "",
- "describe": "环境安装和任务启动命令",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--num_worker": {
- "type": "str",
- "item_type": "str",
- "label": "占用机器个数",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "3",
- "placeholder": "",
- "describe": "占用机器个数",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "deploy-service": {
- "project_name": "模型服务化",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/deploy-service:20211001",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/deploy-service",
- "image_describe": "模型部署推理服务",
- "job_template_name": "deploy-service",
- "job_template_describe": "模型部署推理服务",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "kubeflow-pipeline",
- "job_template_env": "",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/deploy-service"
- },
- "job_template_args": {
- "模型信息": {
- "--project_name": {
- "type": "str",
- "item_type": "str",
- "label": "项目组名称",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "public",
- "placeholder": "",
- "describe": "项目组名称",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--label": {
- "type": "str",
- "item_type": "str",
- "label": "中文描述描述",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "demo推理服务",
- "placeholder": "",
- "describe": "推理服务描述",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--model_name": {
- "type": "str",
- "item_type": "str",
- "label": "模型名",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型名",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--model_version": {
- "type": "str",
- "item_type": "str",
- "label": "模型版本号",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "v2022.10.01.1",
- "placeholder": "",
- "describe": "模型版本号",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--model_path": {
- "type": "str",
- "item_type": "str",
- "label": "模型地址",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "模型地址",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- },
- "部署信息": {
- "--service_type": {
- "type": "str",
- "item_type": "str",
- "label": "推理服务类型",
- "require": 1,
- "choice": [
- "serving",
- "ml-server",
- "tfserving",
- "torch-server",
- "onnxruntime",
- "triton-server"
- ],
- "range": "",
- "default": "service",
- "placeholder": "",
- "describe": "推理服务类型",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--images": {
- "type": "str",
- "item_type": "str",
- "label": "推理服务镜像",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "推理服务镜像",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--working_dir": {
- "type": "str",
- "item_type": "str",
- "label": "推理容器工作目录",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "推理容器工作目录,个人工作目录/mnt/$username",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--command": {
- "type": "str",
- "item_type": "str",
- "label": "推理容器启动命令",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "推理容器启动命令",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--args": {
- "type": "str",
- "item_type": "str",
- "label": "推理容器启动参数",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "推理容器启动参数",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--env": {
- "type": "text",
- "item_type": "str",
- "label": "推理容器环境变量",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "推理容器环境变量",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--host": {
- "type": "str",
- "item_type": "str",
- "label": "部署域名",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "部署域名,留空自动生成",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--ports": {
- "type": "str",
- "item_type": "str",
- "label": "推理容器暴露端口",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "80",
- "placeholder": "",
- "describe": "推理容器暴露端口",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--replicas": {
- "type": "str",
- "item_type": "str",
- "label": "pod副本数",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "pod副本数",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--resource_memory": {
- "type": "str",
- "item_type": "str",
- "label": "每个pod占用内存",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "2G",
- "placeholder": "",
- "describe": "每个pod占用内存",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--resource_cpu": {
- "type": "str",
- "item_type": "str",
- "label": "每个pod占用cpu",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "2",
- "placeholder": "",
- "describe": "每个pod占用cpu",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--resource_gpu": {
- "type": "str",
- "item_type": "str",
- "label": "每个pod占用gpu",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "0",
- "placeholder": "",
- "describe": "每个pod占用gpu",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--volume_mount": {
- "type": "str",
- "item_type": "str",
- "label": "挂载",
- "require": 0,
- "choice": [
- ],
- "range": "",
- "default": "kubeflow-user-workspace(pvc):/mnt",
- "placeholder": "",
- "describe": "容器的挂载,支持pvc/hostpath/configmap三种形式,格式示例:$pvc_name1(pvc):/$container_path1,$hostpath1(hostpath):/$container_path2,注意pvc会自动挂载对应目录下的个人rtx子目录",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "media-download": {
- "project_name": "多媒体类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/video-audio:20210601",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/video-audio",
- "image_describe": "分布式媒体文件处理",
- "job_template_name": "media-download",
- "job_template_describe": "分布式下载媒体文件",
- "job_template_command": "python start_download.py",
- "job_template_volume": "2G(memory):/dev/shm",
- "job_template_account": "kubeflow-pipeline",
- "job_template_expand": {
- "index": 1,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/video-audio"
- },
- "job_template_args": {
- "参数": {
- "--num_worker": {
- "type": "str",
- "item_type": "str",
- "label": "分布式任务的worker数目",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "3",
- "placeholder": "分布式任务的worker数目",
- "describe": "分布式任务的worker数目",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--download_type": {
- "type": "enum",
- "item_type": "str",
- "label": "下载类型",
- "require": 1,
- "choice": [
- "url"
- ],
- "range": "",
- "default": "url",
- "placeholder": "",
- "describe": "下载类型",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--input_file": {
- "type": "str",
- "item_type": "str",
- "label": "下载信息文件地址",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "下载信息文件地址
url类型,每行格式:$url $local_path",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "video-img": {
- "project_name": "多媒体类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/video-audio:20210601",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/video-audio",
- "image_describe": "分布式媒体文件处理",
- "job_template_name": "video-img",
- "job_template_describe": "分布式视频提取图片",
- "job_template_command": "python start_video_img.py",
- "job_template_volume": "2G(memory):/dev/shm",
- "job_template_account": "kubeflow-pipeline",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/video-audio"
- },
- "job_template_args": {
- "参数": {
- "--num_workers": {
- "type": "str",
- "item_type": "str",
- "label": "",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "3",
- "placeholder": "",
- "describe": "worker数量",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--input_file": {
- "type": "str",
- "item_type": "str",
- "label": "",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "配置文件地址,每行格式:
$local_video_path $des_img_dir $frame_rate",
- "describe": "配置文件地址,每行格式:
$local_video_path $des_img_dir $frame_rate",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "video-audio": {
- "project_name": "多媒体类模板",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/video-audio:20210601",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/video-audio",
- "image_describe": "分布式媒体文件处理",
- "job_template_name": "video-audio",
- "job_template_describe": "分布式视频提取音频",
- "job_template_command": "python start_video_audio.py",
- "job_template_volume": "2G(memory):/dev/shm",
- "job_template_account": "kubeflow-pipeline",
- "job_template_expand": {
- "index": 3,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/video-audio"
- },
- "job_template_args": {
- "参数": {
- "--num_workers": {
- "type": "str",
- "item_type": "str",
- "label": "worker数量",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "3",
- "placeholder": "",
- "describe": "worker数量",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- },
- "--input_file": {
- "type": "str",
- "item_type": "str",
- "label": "",
- "require": 1,
- "choice": [
- ],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "配置文件地址,每行格式:
$local_video_path $des_audio_path",
- "editable": 1,
- "condition": "",
- "sub_args": {
- }
- }
- }
- }
- },
- "object-detection-on-darknet": {
- "project_name": "机器视觉",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/object_detection_on_darknet:v1",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/object_detection_on_darknet",
- "image_describe": "yolo目标识别",
- "job_template_name": "object-detection-on-darknet",
- "job_template_describe": "yolo目标识别",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 4,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/object_detection_on_darknet"
- },
- "job_template_args": {
- "参数": {
- "--train_cfg": {
- "type": "text",
- "item_type": "str",
- "label": "模型参数配置、训练配置",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "[net]\n# Testing\n# batch=1\n# subdivisions=1\n# Training\nbatch=64\nsubdivisions=16\nwidth=608\nheight=608\nchannels=3\nmomentum=0.9\ndecay=0.0005\nangle=0\nsaturation = 1.5\nexposure = 1.5\nhue=.1\n\nlearning_rate=0.001\nburn_in=1000\nmax_batches = 50150\npolicy=steps\nsteps=400000,450000\nscales=.1,.1\n\n[convolutional]\nbatch_normalize=1\nfilters=32\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=32\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n######################\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=1024\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=1024\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=1024\nactivation=leaky\n\n[convolutional]\nsize=1\nstride=1\npad=1\nfilters=255\nactivation=linear\n\n\n[yolo]\nmask = 6,7,8\nanchors = 10,13, 16,30, 33,23, 30,61, 62,45, 59,119, 116,90, 156,198, 373,326\nclasses=80\nnum=9\njitter=.3\nignore_thresh = .7\ntruth_thresh = 1\nrandom=1\n\n\n[route]\nlayers = -4\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[upsample]\nstride=2\n\n[route]\nlayers = -1, 61\n\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=512\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=512\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=512\nactivation=leaky\n\n[convolutional]\nsize=1\nstride=1\npad=1\nfilters=255\nactivation=linear\n\n\n[yolo]\nmask = 3,4,5\nanchors = 10,13, 16,30, 33,23, 30,61, 62,45, 59,119, 116,90, 156,198, 373,326\nclasses=80\nnum=9\njitter=.3\nignore_thresh = .7\ntruth_thresh = 1\nrandom=1\n\n\n\n[route]\nlayers = -4\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[upsample]\nstride=2\n\n[route]\nlayers = -1, 36\n\n\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=256\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=256\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=256\nactivation=leaky\n\n[convolutional]\nsize=1\nstride=1\npad=1\nfilters=255\nactivation=linear\n\n\n[yolo]\nmask = 0,1,2\nanchors = 10,13, 16,30, 33,23, 30,61, 62,45, 59,119, 116,90, 156,198, 373,326\nclasses=80\nnum=9\njitter=.3\nignore_thresh = .7\ntruth_thresh = 1\nrandom=1\n\n",
- "placeholder": "",
- "describe": "模型参数配置、训练配置",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--data_cfg": {
- "type": "text",
- "item_type": "str",
- "label": "训练数据配置",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "classes= 80\ntrain = /root/darknet/coco_data/coco/trainvalno5k.txt\n#valid = coco_testdev\nvalid = /root/darknet/coco_data/coco/5k.txt\nnames = /root/darknet/data/coco.names\nbackup = /root/darknet/backup\neval=coco\n\n",
- "placeholder": "",
- "describe": "训练数据配置",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--weights": {
- "type": "str",
- "item_type": "str",
- "label": "预训练模型权重文件",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "",
- "placeholder": "",
- "describe": "预训练模型权重文件",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "stable-diffusion": {
- "project_name": "大模型",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/stable-diffusion",
- "image_describe": "stable-diffusion 算法",
- "job_template_name": "stable-diffusion",
- "job_template_describe": "stable-diffusion 算法(企业版)",
- "job_template_version": "Alpha",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/stable-diffusion"
- },
- "job_template_args": {
- }
- },
- "chatglm2": {
- "project_name": "大模型",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/chatglm2",
- "image_describe": "chatglm2",
- "job_template_name": "chatglm2",
- "job_template_describe": "chatglm2 lora微调(企业版)",
- "job_template_version": "Release",
- "job_template_workdir": "/root/chatglm",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/chatglm2"
- },
- "job_template_args": {
- "训练相关参数": {
- "--train_path": {
- "type": "str",
- "item_type": "str",
- "label": "训练数据集",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/admin/pipeline/example/chatglm2/spo_0.json",
- "placeholder": "",
- "describe": "训练数据集",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--model_name_or_path": {
- "type": "str",
- "item_type": "str",
- "label": "模型目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/admin/pipeline/example/chatglm2/chatglm2-6b/",
- "placeholder": "",
- "describe": "模型目录",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--per_device_train_batch_size": {
- "type": "str",
- "item_type": "str",
- "label": "batch_size",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "batch_size",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--max_len": {
- "type": "str",
- "item_type": "str",
- "label": "最大长度",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1560",
- "placeholder": "",
- "describe": "最大长度",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--max_src_len": {
- "type": "str",
- "item_type": "str",
- "label": "最大长度",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1024",
- "placeholder": "",
- "describe": "最大长度",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--learning_rate": {
- "type": "str",
- "item_type": "str",
- "label": "学习率",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1e-4",
- "placeholder": "",
- "describe": "学习率",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--weight_decay": {
- "type": "str",
- "item_type": "str",
- "label": "权重衰减",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0.1",
- "placeholder": "",
- "describe": "权重衰减",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--num_train_epochs": {
- "type": "str",
- "item_type": "str",
- "label": "训练轮数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2",
- "placeholder": "",
- "describe": "训练轮数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--gradient_accumulation_steps": {
- "type": "str",
- "item_type": "str",
- "label": "梯度累积次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "4",
- "placeholder": "",
- "describe": "梯度累积次数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--warmup_ratio": {
- "type": "str",
- "item_type": "str",
- "label": "预热比例",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0.1",
- "placeholder": "",
- "describe": "预热比例",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--lora_dim": {
- "type": "str",
- "item_type": "str",
- "label": "lora 维度",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "16",
- "placeholder": "",
- "describe": "lora 维度",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--lora_alpha": {
- "type": "str",
- "item_type": "str",
- "label": "lora alpha",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "64",
- "placeholder": "",
- "describe": "lora alpha",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--lora_dropout": {
- "type": "str",
- "item_type": "str",
- "label": "lora dropout",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "0.1",
- "placeholder": "",
- "describe": "lora dropout",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--lora_module_name": {
- "type": "str",
- "item_type": "str",
- "label": "lora module name",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "query_key_value,dense_h_to_4h,dense_4h_to_h,dense",
- "placeholder": "",
- "describe": "lora module name",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--seed": {
- "type": "str",
- "item_type": "str",
- "label": "随机种子",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1234",
- "placeholder": "",
- "describe": "随机种子",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--ds_file": {
- "type": "str",
- "item_type": "str",
- "label": "deepspeed配置文件",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "ds_zero2_no_offload.json",
- "placeholder": "",
- "describe": "deepspeed配置文件",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--show_loss_step": {
- "type": "str",
- "item_type": "str",
- "label": "打印 loss 速率",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "10",
- "placeholder": "",
- "describe": "打印 loss 速率",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--output_dir": {
- "type": "str",
- "item_type": "str",
- "label": "保存地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/admin/pipeline/example/chatglm2",
- "placeholder": "",
- "describe": "保存地址",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- },
- "llama2": {
- "project_name": "大模型",
- "image_name": "python:3.9",
- "gitpath": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/llama2",
- "image_describe": "llama2",
- "job_template_name": "llama2",
- "job_template_describe": "llama2 lora微调(企业版)",
- "job_template_version": "Release",
- "job_template_workdir": "/root/Llama2-Chinese/train/sft",
- "job_template_command": "",
- "job_template_volume": "",
- "job_template_account": "",
- "job_template_env": "",
- "job_template_expand": {
- "index": 2,
- "help_url": "https://github.com/tencentmusic/cube-studio/tree/master/job-template/job/llama2"
- },
- "job_template_args": {
- "训练相关参数": {
- "--model_name_or_path": {
- "type": "str",
- "item_type": "str",
- "label": "模型目录",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/data/home/pengluan/llama2/Llama2-Chinese-7b-Chat/",
- "placeholder": "",
- "describe": "模型目录",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--train_files": {
- "type": "str",
- "item_type": "str",
- "label": "训练数据集",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/data/home/pengluan/llama2/train_sft.csv",
- "placeholder": "",
- "describe": "训练数据集",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--validation_files": {
- "type": "str",
- "item_type": "str",
- "label": "验证数据集",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/data/home/pengluan/llama2/dev_sft.csv",
- "placeholder": "",
- "describe": "验证数据集",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--per_device_train_batch_size": {
- "type": "str",
- "item_type": "str",
- "label": "batch_size",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "batch_size",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--per_device_eval_batch_size": {
- "type": "str",
- "item_type": "str",
- "label": "batch_size",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1",
- "placeholder": "",
- "describe": "batch_size",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--max_eval_samples": {
- "type": "str",
- "item_type": "str",
- "label": "最大评估样本数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "800",
- "placeholder": "",
- "describe": "最大评估样本数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--learning_rate": {
- "type": "str",
- "item_type": "str",
- "label": "学习率",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1e-4",
- "placeholder": "",
- "describe": "学习率",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--gradient_accumulation_steps": {
- "type": "str",
- "item_type": "str",
- "label": "梯度累积次数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "16",
- "placeholder": "",
- "describe": "梯度累积次数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--num_train_epochs": {
- "type": "str",
- "item_type": "str",
- "label": "训练轮数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2",
- "placeholder": "",
- "describe": "训练轮数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--warmup_steps": {
- "type": "str",
- "item_type": "str",
- "label": "预热steps数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "400",
- "placeholder": "",
- "describe": "预热steps数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--load_in_bits": {
- "type": "str",
- "item_type": "str",
- "label": "load_in_bits",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "4",
- "placeholder": "",
- "describe": "load_in_bits",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--lora_r": {
- "type": "str",
- "item_type": "str",
- "label": "lora_r",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "8",
- "placeholder": "",
- "describe": "lora_r",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--lora_alpha": {
- "type": "str",
- "item_type": "str",
- "label": "lora alpha",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "32",
- "placeholder": "",
- "describe": "lora alpha",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--target_modules": {
- "type": "str",
- "item_type": "str",
- "label": "lora module name",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "q_proj,k_proj,v_proj,o_proj,down_proj,gate_proj,up_proj",
- "placeholder": "",
- "describe": "lora module name",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--logging_steps": {
- "type": "str",
- "item_type": "str",
- "label": "打印 log 速率",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "10",
- "placeholder": "",
- "describe": "打印 log 速率",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--preprocessing_num_workers": {
- "type": "str",
- "item_type": "str",
- "label": "数据预处理时使用的进程数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "10",
- "placeholder": "",
- "describe": "数据预处理时使用的进程数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--save_steps": {
- "type": "str",
- "item_type": "str",
- "label": "保存step数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "200",
- "placeholder": "",
- "describe": "保存step数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--save_total_limit": {
- "type": "str",
- "item_type": "str",
- "label": "保存总限制数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "2000",
- "placeholder": "",
- "describe": "保存总限制数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--eval_steps": {
- "type": "str",
- "item_type": "str",
- "label": "评估 step数",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "200",
- "placeholder": "",
- "describe": "评估 step数",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--seed": {
- "type": "str",
- "item_type": "str",
- "label": "随机种子",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "1234",
- "placeholder": "",
- "describe": "随机种子",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--block_size": {
- "type": "str",
- "item_type": "str",
- "label": "block_size",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "4096",
- "placeholder": "",
- "describe": "block_size",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--deepspeed": {
- "type": "str",
- "item_type": "str",
- "label": "deepspeed配置文件",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "ds_config_zero2.json",
- "placeholder": "",
- "describe": "deepspeed配置文件",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- },
- "--output_dir": {
- "type": "str",
- "item_type": "str",
- "label": "保存地址",
- "require": 1,
- "choice": [],
- "range": "",
- "default": "/mnt/admin/pipeline/example/llama2",
- "placeholder": "",
- "describe": "保存地址",
- "editable": 1,
- "condition": "",
- "sub_args": {}
- }
- }
- }
- }
-}
diff --git a/myapp/init-pipeline.json b/myapp/init-pipeline.json
deleted file mode 100644
index d74b9beb..00000000
--- a/myapp/init-pipeline.json
+++ /dev/null
@@ -1,77 +0,0 @@
-{
- "darknet-yolov3": {
- "pipeline": {
- "name": "imageAI",
- "describe": "图像预测+物体检测+视频跟踪",
- "project": "public",
- "parameter": {
- "demo": "true",
- "img": "/static/assets/images/pipeline/yolo.jpg"
- },
- "dag_json": {
- "download-data": {
- "upstream": []
- },
- "yolov3-object-recognition": {
- "upstream": [
- "download-data"
- ]
- },
- "deploy-darknet-web-service": {
- "upstream": [
- "yolov3-object-recognition"
- ]
- }
- }
- },
- "tasks": [
- {
- "job_templete": "自定义镜像",
- "name": "download-data",
- "label": "下载处理标注数据",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt/",
- "args": {
- "images": "ccr.ccs.tencentyun.com/cube-studio/ubuntu-gpu:cuda11.8.0-cudnn8-python3.9",
- "workdir": "/mnt/admin/",
- "command": "wget https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/pipeline/coco_data_sample.zip && unzip -o coco_data_sample.zip && cd coco_data_sample && bash reset_file.sh"
- }
- },
- {
- "job_templete": "object-detection-on-darknet",
- "name": "yolov3-object-recognition",
- "label": "目标识别训练",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt/",
- "resource_memory": "5G",
- "resource_cpu": "5",
- "args": {
- "--train_cfg": "[net]\n# Testing\n# batch=1\n# subdivisions=1\n# Training\nbatch=64\nsubdivisions=16\nwidth=608\nheight=608\nchannels=3\nmomentum=0.9\ndecay=0.0005\nangle=0\nsaturation = 1.5\nexposure = 1.5\nhue=.1\n\nlearning_rate=0.001\nburn_in=1000\nmax_batches = 50023\npolicy=steps\nsteps=400000,450000\nscales=.1,.1\n\n[convolutional]\nbatch_normalize=1\nfilters=32\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=32\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=64\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n# Downsample\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=2\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=1024\nsize=3\nstride=1\npad=1\nactivation=leaky\n\n[shortcut]\nfrom=-3\nactivation=linear\n\n######################\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=1024\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=1024\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=512\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=1024\nactivation=leaky\n\n[convolutional]\nsize=1\nstride=1\npad=1\nfilters=255\nactivation=linear\n\n\n[yolo]\nmask = 6,7,8\nanchors = 10,13, 16,30, 33,23, 30,61, 62,45, 59,119, 116,90, 156,198, 373,326\nclasses=80\nnum=9\njitter=.3\nignore_thresh = .7\ntruth_thresh = 1\nrandom=1\n\n\n[route]\nlayers = -4\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[upsample]\nstride=2\n\n[route]\nlayers = -1, 61\n\n\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=512\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=512\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=256\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=512\nactivation=leaky\n\n[convolutional]\nsize=1\nstride=1\npad=1\nfilters=255\nactivation=linear\n\n\n[yolo]\nmask = 3,4,5\nanchors = 10,13, 16,30, 33,23, 30,61, 62,45, 59,119, 116,90, 156,198, 373,326\nclasses=80\nnum=9\njitter=.3\nignore_thresh = .7\ntruth_thresh = 1\nrandom=1\n\n\n\n[route]\nlayers = -4\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[upsample]\nstride=2\n\n[route]\nlayers = -1, 36\n\n\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=256\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=256\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nfilters=128\nsize=1\nstride=1\npad=1\nactivation=leaky\n\n[convolutional]\nbatch_normalize=1\nsize=3\nstride=1\npad=1\nfilters=256\nactivation=leaky\n\n[convolutional]\nsize=1\nstride=1\npad=1\nfilters=255\nactivation=linear\n\n\n[yolo]\nmask = 0,1,2\nanchors = 10,13, 16,30, 33,23, 30,61, 62,45, 59,119, 116,90, 156,198, 373,326\nclasses=80\nnum=9\njitter=.3\nignore_thresh = .7\ntruth_thresh = 1\nrandom=1\n\n",
- "--data_cfg": "classes= 80\ntrain = /mnt/admin/coco_data_sample/train.txt\nvalid = /mnt/admin/coco_data_sample/valid.txt\nnames = /mnt/admin/coco_data_sample/coco.names\nbackup = /mnt/admin/coco_data_sample/yolo\neval=coco\n\n",
- "--weights": "/mnt/admin/coco_data_sample/yolov3.weights"
- }
- },
- {
- "job_templete": "deploy-service",
- "name": "deploy-darknet-web-service",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt/",
- "label": "部署模型web服务",
- "args": {
- "--label": "目标识别推理服务",
- "--model_name": "yolov3",
- "--model_version": "v2022.10.01.1",
- "--model_path": "",
- "--service_type": "serving",
- "--images": "ccr.ccs.tencentyun.com/cube-studio/target-detection",
- "--working_dir": "",
- "--command": "",
- "--args": "",
- "--env": "YOLO_DATA_PATH=/mnt/admin/coco_data_sample/yolo/coco.data\nYOLO_CFG_PATH=/mnt/admin/coco_data_sample/yolo/yolov3.cfg\nYOLO_WEIGHTS_PATH=/mnt/admin/coco_data_sample/yolo/yolov3.weights",
- "--ports": "8080",
- "--replicas": "1",
- "--resource_memory": "5G",
- "--resource_cpu": "5",
- "--resource_gpu": "0"
- }
- }
- ]
- }
-}
\ No newline at end of file
diff --git a/myapp/init-service.json b/myapp/init-service.json
deleted file mode 100644
index 79ea6da1..00000000
--- a/myapp/init-service.json
+++ /dev/null
@@ -1,75 +0,0 @@
-{
- "mysql-ui": {
- "project_name": "public",
- "service_name": "mysql-ui",
- "service_describe": "可视化编辑mysql数据库",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/phpmyadmin",
- "command": "",
- "env": "PMA_HOST=mysql-service.infra\nPMA_PORT=3306\nPMA_USER=root\nPMA_PASSWORD=admin",
- "ports": "80",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/blob/master/docs/example/service.md"
- }
- },
- "redis-ui": {
- "project_name": "public",
- "service_name": "redis-ui",
- "service_describe": "可视化编辑redis数据库",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/patrikx3:latest",
- "command": "",
- "env": "REDIS_NAME=default\nREDIS_HOST=redis-master.infra\nREDIS_PORT=6379\nREDIS_PASSWORD=admin",
- "ports": "7843",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/blob/master/docs/example/service.md"
- }
- },
- "mongo-express": {
- "project_name": "public",
- "service_name": "mongo-express",
- "service_describe": "可视化编辑mongo数据库",
- "image_name": "mongo-express:0.54.0",
- "command": "",
- "env": "ME_CONFIG_MONGODB_SERVER=xx.xx.xx.xx\nME_CONFIG_MONGODB_PORT=xx\nME_CONFIG_MONGODB_ENABLE_ADMIN=true\nME_CONFIG_MONGODB_ADMINUSERNAME=xx\nME_CONFIG_MONGODB_ADMINPASSWORD=xx\nME_CONFIG_MONGODB_AUTH_DATABASE=xx\nME_CONFIG_MONGODB_AUTH_USERNAME=xx\nME_CONFIG_MONGODB_AUTH_PASSWORD=xx\nVCAP_APP_HOST=0.0.0.0\nVCAP_APP_PORT=8081\nME_CONFIG_OPTIONS_EDITORTHEME=ambiance",
- "ports": "8081",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/blob/master/docs/example/service.md"
- }
- },
- "neo4j": {
- "project_name": "public",
- "service_name": "neo4j",
- "service_describe": "可视化编辑图数据库neo4j",
- "image_name": "ccr.ccs.tencentyun.com/cube-studio/neo4j:4.4",
- "command": "",
- "env": "NEO4J_AUTH=neo4j/admin",
- "ports": "7474,7687",
- "volume_mount": "kubeflow-user-workspace(pvc):/mnt,/data/k8s/kubeflow/pipeline/workspace/admin/neo4j(hostpath):/var/lib/neo4j/data",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/blob/master/docs/example/service.md"
- }
- },
- "jaeger": {
- "project_name": "public",
- "service_name": "jaeger",
- "service_describe": "jaeger链路追踪",
- "image_name": "jaegertracing/all-in-one:1.29",
- "command": "",
- "env": "",
- "ports": "16686,5775",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/blob/master/docs/example/service.md"
- }
- },
- "postgresql-ui": {
- "project_name": "public",
- "service_name": "postgresql-ui",
- "service_describe": "可视化编辑postgresql数据库",
- "image_name": "dpage/pgadmin4",
- "command": "",
- "env": "PGADMIN_DEFAULT_EMAIL=admin@tencent.com\nPGADMIN_DEFAULT_PASSWORD=root",
- "ports": "80",
- "expand": {
- "help_url": "https://github.com/tencentmusic/cube-studio/blob/master/docs/example/service.md"
- }
- }
-}
\ No newline at end of file
diff --git a/myapp/init-train-model.json b/myapp/init-train-model.json
deleted file mode 100644
index ed3d3ed0..00000000
--- a/myapp/init-train-model.json
+++ /dev/null
@@ -1,38 +0,0 @@
-{
- "tf-mnist": {
- "project_name": "public",
- "name": "mnist",
- "version": "v2022.08.01.1",
- "describe": "tf mnist 图像分类 tfserving推理",
- "path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/tf-mnist.tar.gz",
- "framework": "tf",
- "api_type": "tfserving"
- },
- "pytorch-resnet50": {
- "project_name": "public",
- "name": "resnet50",
- "version": "v2022.08.01.2",
- "describe": "pytorch resnet50 图像分类 torch-server推理",
- "path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50.mar",
- "framework": "pytorch",
- "api_type": "torch-server"
- },
- "torchscript-resnet50": {
- "project_name": "public",
- "name": "resnet50",
- "version": "v2022.08.01.3",
- "describe": "torchscript resnet50 图像分类 triton推理",
- "path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50-torchscript.pt",
- "framework": "pytorch",
- "api_type": "triton-server"
- },
- "onnx-resnet50": {
- "project_name": "public",
- "name": "resnet50",
- "version": "v2022.08.01.4",
- "describe": "onnx resnet50 图像分类 triton推理",
- "path": "https://docker-76009.sz.gfp.tencent-cloud.com/github/cube-studio/inference/resnet50.onnx",
- "framework": "onnx",
- "api_type": "triton-server"
- }
-}
\ No newline at end of file
diff --git a/myapp/jinja_context.py b/myapp/jinja_context.py
deleted file mode 100644
index bd888ab8..00000000
--- a/myapp/jinja_context.py
+++ /dev/null
@@ -1,223 +0,0 @@
-"""Defines the templating context for SQL Lab"""
-from datetime import datetime, timedelta
-import inspect
-import json
-import random
-import time
-from typing import Any, List, Optional, Tuple
-import uuid
-
-from dateutil.relativedelta import relativedelta
-from flask import g, request
-from jinja2.sandbox import SandboxedEnvironment
-
-from myapp import app
-
-# template context
-conf = app.config
-BASE_CONTEXT = {
- "datetime": datetime,
- "random": random,
- "relativedelta": relativedelta,
- "time": time,
- "timedelta": timedelta,
- "uuid": uuid,
-}
-BASE_CONTEXT.update(conf.get("JINJA_CONTEXT_ADDONS", {}))
-
-
-def url_param(param: str, default: Optional[str] = None) -> Optional[Any]:
- """Read a url or post parameter and use it in your SQL Lab query
-
- When in SQL Lab, it's possible to add arbitrary URL "query string"
- parameters, and use those in your SQL code. For instance you can
- alter your url and add `?foo=bar`, as in
- `{domain}/myapp/sqllab?foo=bar`. Then if your query is something like
- SELECT * FROM foo = '{{ url_param('foo') }}', it will be parsed at
- runtime and replaced by the value in the URL.
-
- As you create a visualization form this SQL Lab query, you can pass
- parameters in the explore view as well as from the dashboard, and
- it should carry through to your queries.
-
- :param param: the parameter to lookup
- :param default: the value to return in the absence of the parameter
- """
- if request.args.get(param):
- return request.args.get(param, default)
- # Supporting POST as well as get
- form_data = request.form.get("form_data")
- if isinstance(form_data, str):
- form_data = json.loads(form_data)
- url_params = form_data.get("url_params") or {}
- return url_params.get(param, default)
- return default
-
-
-def current_user_id() -> Optional[int]:
- """The id of the user who is currently logged in"""
- if hasattr(g, "user") and g.user:
- return g.user.id
- return None
-
-
-def current_username() -> Optional[str]:
- """The username of the user who is currently logged in"""
- if g.user:
- return g.user.username
- return None
-
-
-def filter_values(column: str, default: Optional[str] = None) -> List[str]:
- """ Gets a values for a particular filter as a list
-
- This is useful if:
- - you want to use a filter box to filter a query where the name of filter box
- column doesn't match the one in the select statement
- - you want to have the ability for filter inside the main query for speed
- purposes
-
- This searches for "filters" and "extra_filters" in ``form_data`` for a match
-
- Usage example::
-
- SELECT action, count(*) as times
- FROM logs
- WHERE action in ( {{ "'" + "','".join(filter_values('action_type')) + "'" }} )
- GROUP BY action
-
- :param column: column/filter name to lookup
- :param default: default value to return if there's no matching columns
- :return: returns a list of filter values
- """
- form_data = json.loads(request.form.get("form_data", "{}"))
- return_val = []
- for filter_type in ["filters", "extra_filters"]:
- if filter_type not in form_data:
- continue
-
- for f in form_data[filter_type]:
- if f["col"] == column:
- if isinstance(f["val"], list):
- for v in f["val"]:
- return_val.append(v)
- else:
- return_val.append(f["val"])
-
- if return_val:
- return return_val
-
- if default:
- return [default]
- else:
- return []
-
-
-class CacheKeyWrapper:
- """ Dummy class that exposes a method used to store additional values used in
- calculation of query object cache keys"""
-
- def __init__(self, extra_cache_keys: Optional[List[Any]] = None):
- self.extra_cache_keys = extra_cache_keys
-
- def cache_key_wrapper(self, key: Any) -> Any:
- """ Adds values to a list that is added to the query object used for calculating
- a cache key.
-
- This is needed if the following applies:
- - Caching is enabled
- - The query is dynamically generated using a jinja template
- - A username or similar is used as a filter in the query
-
- Example when using a SQL query as a data source ::
-
- SELECT action, count(*) as times
- FROM logs
- WHERE logged_in_user = '{{ cache_key_wrapper(current_username()) }}'
- GROUP BY action
-
- This will ensure that the query results that were cached by `user_1` will
- **not** be seen by `user_2`, as the `cache_key` for the query will be
- different. ``cache_key_wrapper`` can be used similarly for regular table data
- sources by adding a `Custom SQL` filter.
-
- :param key: Any value that should be considered when calculating the cache key
- :return: the original value ``key`` passed to the function
- """
- if self.extra_cache_keys is not None:
- self.extra_cache_keys.append(key)
- return key
-
-
-class BaseTemplateProcessor:
- """Base class for database-specific jinja context
-
- There's this bit of magic in ``process_template`` that instantiates only
- the database context for the active database as a ``models.Database``
- object binds it to the context object, so that object methods
- have access to
- that context. This way, {{ hive.latest_partition('mytable') }} just
- knows about the database it is operating in.
-
- This means that object methods are only available for the active database
- and are given access to the ``models.Database`` object and schema
- name. For globally available methods use ``@classmethod``.
- """
-
- engine: Optional[str] = None
-
- def __init__(
- self,
- database=None,
- query=None,
- table=None,
- extra_cache_keys: Optional[List[Any]] = None,
- **kwargs
- ):
- self.database = database
- self.query = query
- self.schema = None
- if query and query.schema:
- self.schema = query.schema
- elif table:
- self.schema = table.schema
- self.context = {
- "url_param": url_param,
- "current_user_id": current_user_id,
- "current_username": current_username,
- "cache_key_wrapper": CacheKeyWrapper(extra_cache_keys).cache_key_wrapper,
- "filter_values": filter_values,
- "form_data": {},
- }
- self.context.update(kwargs)
- self.context.update(BASE_CONTEXT)
- if self.engine:
- self.context[self.engine] = self
- self.env = SandboxedEnvironment()
-
- def process_template(self, sql: str, **kwargs) -> str:
- """Processes a sql template
-
- >>> sql = "SELECT '{{ datetime(2017, 1, 1).isoformat() }}'"
- >>> process_template(sql)
- "SELECT '2017-01-01T00:00:00'"
- """
- template = self.env.from_string(sql)
- kwargs.update(self.context)
- return template.render(kwargs)
-
-
-
-template_processors = {}
-keys = tuple(globals().keys())
-for k in keys:
- o = globals()[k]
- if o and inspect.isclass(o) and issubclass(o, BaseTemplateProcessor):
- template_processors[o.engine] = o
-
-
-def get_template_processor(database, table=None, query=None, **kwargs):
- TP = template_processors.get(database.backend, BaseTemplateProcessor)
- return TP(database=database, table=table, query=query, **kwargs)
-
-