灵活高效的容器化计算集群平台 支持多种框架与加速器
Backend.AI是一个基于容器的计算集群平台,支持多种计算和机器学习框架及编程语言。平台提供CUDA GPU、ROCm GPU、TPU和IPU等异构加速器支持,可按需分配和隔离计算资源,适合多租户环境。通过REST、GraphQL和WebSocket API暴露功能,为用户提供灵活高效的计算环境。此平台集成了先进的资源调度功能,可实现按需或批量分配计算资源。Backend.AI采用容器技术实现资源隔离,确保多租户环境的安全性和效率。其开放的API架构便于与现有系统集成,为科研、教育和企业用户提供了强大而灵活的计算解决方案。
Backend.AI is a streamlined, container-based computing cluster platform that hosts popular computing/ML frameworks and diverse programming languages, with pluggable heterogeneous accelerator support including CUDA GPU, ROCm GPU, TPU, IPU and other NPUs.
It allocates and isolates the underlying computing resources for multi-tenant computation sessions on-demand or in batches with customizable job schedulers with its own orchestrator. All its functions are exposed as REST/GraphQL/WebSocket APIs.
This repository contains all open-source server-side components and the client SDK for Python as a reference implementation of API clients.
src/ai/backend/
: Source codes
manager/
: Managermanager/api
: Manager API handlersagent/
: Agentagent/docker/
: Agent's Docker backendagent/k8s/
: Agent's Kubernetes backendkernel/
: Agent's kernel runner counterpartrunner/
: Agent's in-kernel prebuilt binarieshelpers/
: Agent's in-kernel helper packagecommon/
: Shared utilitiesclient/
: Client SDKcli/
: Unified CLI for all componentsstorage/
: Storage proxystorage/api
: Storage proxy's manager-facing and client-facing APIsweb/
: Web UI server
static/
: Backend.AI WebUI release artifactsplugin/
: Plugin subsystemtest/
: Integration test suitetestutils/
: Shared utilities used by unit testsmeta/
: Legacy meta packagedocs/
: Unified documentationtests/
manager/
, agent/
, ...: Per-component unit testsconfigs/
manager/
, agent/
, ...: Per-component sample configurationsdocker/
: Dockerfiles for auxiliary containersfixtures/
manager/
, ...: Per-component fixtures for development setup and testsplugins/
: A directory to place plugins such as accelerators, monitors, etc.scripts/
: Scripts to assist development workflows
install-dev.sh
: The single-node development setup script from the working copystubs/
: Type annotation stub packages written by ustools/
: A directory to host Pants-related toolingdist/
: A directory to put build artifacts (.whl files) and Pants-exported virtualenvschanges/
: News fragments for towncrierpants.toml
: The Pants configurationpyproject.toml
: Tooling configuration (towncrier, pytest, mypy)BUILD
: The root build config file**/BUILD
: Per-directory build config filesBUILD_ROOT
: An indicator to mark the build root directory for Pantsrequirements.txt
: The unified requirements file*.lock
, tools/*.lock
: The dependency lock filesdocker-compose.*.yml
: Per-version recommended halfstack container configsREADME.md
: This fileMIGRATION.md
: The migration guide for updating between major releasesVERSION
: The unified version declarationServer-side components are licensed under LGPLv3 to promote non-proprietary open innovation in the open-source community while other shared libraries and client SDKs are distributed under the MIT license.
There is no obligation to open your service/system codes if you just run the server-side components as-is (e.g., just run as daemons or import the components without modification in your codes). Please contact us (contact-at-lablup-com) for commercial consulting and more licensing details/options about individual use-cases.
Run scripts/install-dev.sh
after cloning this repository.
This script checks availability of all required dependencies such as Docker and bootstrap a development
setup. Note that it requires sudo
and a modern Python installed in the host system based on Linux
(Debian/RHEL-likes) or macOS.
Please consult our documentation for community-supported materials. Contact the sales team (contact@lablup.com) for professional paid support and deployment options.
Backend.AI provides websocket tunneling into individual computation sessions (containers), so that users can use their browsers and client CLI to access in-container applications directly in a secure way.
Backend.AI provides an abstraction layer on top of existing network-based storages (e.g., NFS/SMB), called vfolders (virtual folders). Each vfolder works like a cloud storage that can be mounted into any computation sessions and shared between users and user groups with differentiated privileges.
It routes external API requests from front-end services to individual agents. It also monitors and scales the cluster of multiple agents (a few tens to hundreds).
src/ai/backend/manager
backendai_scheduler_v10
backendai_hook_v20
backendai_webapp_v20
backendai_monitor_stats_v10
backendai_monitor_error_v10
It manages individual server instances and launches/destroys Docker containers where REPL daemons (kernels) run. Each agent on a new EC2 instance self-registers itself to the instance registry via heartbeats.
src/ai/backend/agent
backendai_accelerator_v21
backendai_monitor_stats_v10
backendai_monitor_error_v10
It provides a unified abstraction over multiple different network storage devices with vendor-specific enhancements such as real-time performance metrics and filesystem operation acceleration APIs.
src/ai/backend/storage
It hosts the SPA (single-page application) packaged from our web UI codebase for end-users and basic administration tasks.
src/ai/backend/web
Synchronizing the static Backend.AI WebUI version:
$ scripts/download-webui-release.sh <target version to download>
Computing environment recipes (Dockerfile) to build the container images to execute on top of the Backend.AI platform.
A programmable sandbox implemented using ptrace-based system call filtering written in Rust.
A set of libc overrides for resource control and web-based interactive stdin (paired with agents).
We offer client SDKs in popular programming languages. These SDKs are freely available with MIT License to ease integration with both commercial and non-commercial software products and services.
pip install backend.ai-client
npm install backend.ai-client
composer require lablup/backend.ai-client
backendai_accelerator_v21
ai.backend.accelerator.cuda
: CUDA accelerator pluginai.backend.accelerator.cuda
(mock): CUDA mockup plugin
ai.backend.accelerator.rocm
: ROCm accelerator pluginbackendai_monitor_stats_v10
ai.backend.monitor.stats
backendai_monitor_error_v10
ai.backend.monitor.error
These components still exist but are no longer actively maintained.
The front-end support libraries to handle multi-media outputs (e.g., SVG plots, animated vector graphics)
lablup
) is installed inside kernel containers.We now recommend using in-kernel applications such as Jupyter Lab, Visual Studio Code Server, or native SSH connection to kernels via our client SDK or desktop apps.
Backend.AI Core Version | Python Version | Pantsbuild version |
---|---|---|
24.03.x / 24.09.x | 3.12.x | 2.21.x |
23.03.x / 23.09.x | 3.11.x | 2.19.x |
22.03.x / 22.09.x | 3.10.x | |
21.03.x / 21.09.x | 3.8.x |
Refer to LICENSE file.
字节跳动发布的AI编程神器IDE
Trae是一种自适应的集成开发环境(IDE),通过自动化和多元协作改变开发流程。利用Trae,团队能够更快速、精确地编写和部署代码,从而提高编程效率和项目交付速度。Trae具备上下文感知和代码自动完成功能,是提升开发效率的理想工具。
全能AI智能助手,随时解答生活与工作的多样问题
问小白,由元石科技研发的AI智能助手,快速准确地解答各种生活和工作问题,包括但不限于搜索、规划和社交互动,帮助用户在日常生活中提高效率,轻松管理个人事务。
实时语音翻译/同声传译工具
Transly是一个多场景的AI大语言模型驱动的同声传译、专业翻译助手,它拥有超精准的音频识别翻译能力,几乎零延迟的使用体验和支持多国语言可以让你带它走遍全球,无论你是留学生、商务人士、韩剧美剧爱好者,还是出国游玩、多国会议、跨国追星等等,都可以满足你所有需要同传的场景需求,线上线下通用,扫除语言障碍,让全世界的语言交流不再有国界。
一键生成PPT和Word,让学习生活更轻松
讯飞智文是一个利用 AI 技术的项目,能够帮助用户生成 PPT 以及各类文档。无论是商业领域的市场分析报告、年度目标制定,还是学生群体的职业生涯规划、实习避坑指南,亦或是活动策划、旅游攻略等内容,它都能提供支持,帮助用户精准表达,轻松呈现各种信息。
深度推理能力全新升级,全面对标OpenAI o1
科大讯飞的星火大模型,支持语言理解、知识问答和文本创作等多功能,适用于多种文件和业务场景,提升办公和日常生活的效率。讯飞星火是一个提供丰富智能服务的平台,涵盖科技资讯、图像创作、写作辅助、编程解答、科研文献解读等功能,能为不同需求的用户提供便捷高效的帮助,助力用户轻松获取信息、解决问题,满足多样化使用场景。
一种基于大语言模型的高效单流解耦语音令牌文本到语音合成模型
Spark-TTS 是一个基于 PyTorch 的开源文本到语音合成项目,由多个知名机构联合参与。该项目提供了高效的 LLM(大语言模型)驱动的语音合成方案,支持语音克隆和语音创建功能,可通过命令行界面(CLI)和 Web UI 两种方式使用。用户可以根据需求调整语音的性别、音高、速度等参数,生成高质量的语音。该项目适用于多种场景,如有声读物制作、智能语音助手开发等。
AI助力,做PPT更简单!
咔片是一款轻量化在线演示设计工具,借助 AI 技术,实现从内容生成到智能设计的一站式 PPT 制作服务。支持多种文档格式导入生成 PPT,提供海量模板、智能美化、素材替换等功能,适用于销售、教师、学生等各类人群,能高效制作出高品质 PPT,满足不同场景演示需求。
选题、配图、成文,一站式创作,让内容运营更高效
讯飞绘文,一个AI集成平台,支持写作、选题、配图、排版和发布。高效生成适用于各类媒体的定制内容,加速品牌传播,提升内容营销效果。
专业的AI公文写作平台,公文写作神器
AI 材料星,专业的 AI 公文写作辅助平台,为体制内工作人员提供高效的公文写作解决方案。拥有海量公文文库、9 大核心 AI 功能,支持 30 + 文稿类型生成,助力快速完成领导讲话、工作总结、述职报告等材料,提升办公效率,是体制打工人的得力写作神器。
OpenAI Agents SDK,助力开发者便捷使用 OpenAI 相关功能。
openai-agents-python 是 OpenAI 推出的一款强大 Python SDK,它为开发者提供了与 OpenAI 模型交互的高效工具,支持工具调用、结果处理、追踪等功能,涵盖多种应用场景,如研究助手、财务研究等,能显著提升开发效率,让开发者更轻松地利用 OpenAI 的技术优势。
最新AI工具、AI资讯
独家AI资源、AI项目落地
微信扫一扫关注公众号