@inproceedings{jiaqigu2021L2ight,
title = {L2ight: Enabling On-Chip Learning for Optical Neural Networks via Efficient in-situ Subspace Optimization},
author = {Jiaqi Gu and Hanqing Zhu and Chenghao Feng and Zixuan Jiang and Ray T. Chen and David Z. Pan},
booktitle = {Conference on Neural Information Processing Systems (NeurIPS)},
year = {2021}
}
<h3><p align="center">Fast, Scalable, Easy Customization, Support Hardware-Aware Cross-Layer Co-Design</p></h3>
<p align="center">
<a href="https://github.com/JeremieMelo/pytorch-onn/blob/release/LICENSEE">
<img alt="MIT License" src="https://img.shields.io/apm/l/atomic-design-ui.svg?">
</a>
</p>
<br />
Integrated neuromorphic photonics simulation framework based on PyTorch. It supports coherent and incoherent optical neural networks (ONNs) training/inference on GPUs. It can scale up to million-parameter ONNs with efficient implementation.
Researchers on neuromorphic photonics, optical AI system design, photonic integrated circuit optimization, ONN training/inference.
CUDA-backed fast GPU support, optimized highly-parallel tensorized processing, versatile APIs for device/circuit/architecture/algorithm co-optimization
git clone https://github.com/JeremieMelo/pytorch-onn.git
cd pytorch-onn python3 setup.py install --user clean
or
./setup.sh
Construct optical NN models as simple as constructing a normal pytorch model.
import torch.nn as nn import torch.nn.functional as F import torchonn as onn from torchonn.models import ONNBaseModel class ONNModel(ONNBaseModel): def __init__(self, device=torch.device("cuda:0)): super().__init__(device=device) self.conv = onn.layers.MZIBlockConv2d( in_channels=1, out_channels=8, kernel_size=3, stride=1, padding=1, dilation=1, bias=True, miniblock=4, mode="usv", decompose_alg="clements", photodetect=True, device=device, ) self.pool = nn.AdaptiveAvgPool2d(5) self.linear = onn.layers.MZIBlockLinear( in_features=8*5*5, out_features=10, bias=True, miniblock=4, mode="usv", decompose_alg="clements", photodetect=True, device=device, ) self.conv.reset_parameters() self.linear.reset_parameters() def forward(self, x): x = torch.relu(self.conv(x)) x = self.pool(x) x = x.flatten(1) x = self.linear(x) return x
weight, usv, phase modes and their conversion.python3 unitest/test_op.py
, and check the runtime comparison.fft, hadamard, zero_bias, and trainable modes.| File | Description |
|---|---|
| torchonn/ | Library source files with model, layer, and device definition |
| torchonn/op | Basic operators and CUDA-accelerated operators |
| torchonn/layers | Optical device-implemented layers |
| torchonn/models | Base ONN model templete |
| torchonn/devices | Optical device parameters and configurations |
| examples/ | ONN model building and training examples |
| examples/configs | YAML-based configuration files |
| examples/core | ONN model definition and training utility |
| example/train.py | training script |
The examples/ folder contains more examples to train the ONN
models.
An example optical convolutional neural network MZI_CLASS_CNN is defined in examples/core/models/mzi_cnn.py.
Training facilities, e.g., optimizer, critetion, lr_scheduler, models are built in examples/core/builder.py.
The training and validation logic is defined in examples/train.py.
All training hyperparameters are hierarchically defined in the yaml configuration file examples/configs/mnist/mzi_onn/train.yml (The final config is the union of all default.yml from higher-level directories and this specific train.yml ).
By running the following commands,
# train the example MZI-based CNN model with 2 64-channel Conv layers and 1 Linear layer # training will happend in usv mode to optimize U, Sigma, and V* # projected gradient descent will be applied to guarantee the orthogonality of U and V* # the final step will convert unitary matrices into MZI phases and evaluate in the phase mode cd examples python3 train.py configs/mnist/mzi_cnn/train.yml # [followed by any command-line arguments that override the values in config file, e.g., --optimizer.lr=0.001]
Detailed documentations coming soon.
Jiaqi Gu (jqgu@utexas.edu)
Neural operator-enabled fast photonic device simulation: See NeurOLight, NeurIPS 2022.
Automatic photonic tensor core design: See ADEPT, DAC 2022.
Endurance-enhanced photonic in-memory computing: See ELight, ASP-DAC 2022.
Scalable ONN on-chip learning: See L2ight, NeurIPS 2021.
Memory-efficient ONN architecture: See Memory-Efficient-ONN, ICCV 2021.
SqueezeLight: Scalable ONNs with Multi-Operand Ring Resonators: See SqueezeLight, DATE 2021.


全球首个AI音乐社区
音述AI是全球首个AI音乐社区,致力让每个人都能用音乐表达自我。音述AI提供零门槛AI创作工具,独创GETI法则帮助用户精准定义音乐风格,AI润色功能支持自动优化作品质感。音述AI支持交流讨论、二次创作与价值变现。针对中文用户的语言习惯与文化背景进行专门优化,支持国风融合、C-pop等本土音乐标签,让技术更好地承载人文表达。


一站式搞定所有学习需求
不再被海量信息淹没,开始真正理解知识。Lynote 可摘要 YouTube 视频、PDF、文章等内容。即时创建笔记,检测 AI 内容并下载资料,将您的学习效率提升 10 倍。


为AI短剧协作而生
专为AI短剧协作而生的AniShort正式发布,深度重构AI短剧全流程生产模式,整合创意策划、制作执行、实时协作、在线审片、资产复用等全链路功能,独创无限画布、双轨并行工业化工作流与Ani智能体助手,集成多款主流AI大模型,破解素材零散、版本混乱、沟通低效等行业痛点,助力3人团队效率提升800%,打造标准化、可追溯的AI短剧量产体系,是AI短剧团队协同创作、提升制作效率的核心工具。


能听懂你表达的视频模型
Seedance two是基于seedance2.0的中国大模型,支持图像、视频、音频、文本四种模态输入,表达方式更丰富,生成也更可控。


国内直接访问,限时3折
输入简单文字,生成想要的图片,纳米香蕉中文站基于 Google 模型的 AI 图片生成网站,支持文字生图、图生图。官网价格限时3折活动


职场AI,就用扣子
AI办公助手,复杂任务高效处理。办公效率低?扣子空间AI助手支持播客生成、PPT制作、网页开发及报告写作,覆盖科研、商业、舆情等领域的专家Agent 7x24小时响应,生活工作无缝切换,提升50%效率!


多风格AI绘画神器
堆友平台由阿里巴巴设计团队创建,作为一款AI驱动的设计工具,专为设计师提供一站式增长服务。功能覆盖海量3D素材、AI绘画、实时渲染以及专业抠图,显著提升设计品质和效率。平台不仅提供工具,还是一个促进创意交流和个人发展的空间,界面友好,适合所有级别的设计师和创意工作者。


零代码AI应用开发平台
零代码AI应用开发平台,用户只需一句话简单描述需求,AI能自动生成小程序、APP或H5网页应用,无需编写代码。


免费创建高清无水印Sora视频
Vora是一个免费创建高清无水印Sora视频的AI工具


最适合小白的AI自 动化工作流平台
无需编码,轻松生成可复用、可变现的AI自动化工作流