| flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.2/flash_attn-2.8.2+cu12torch2.7cxx11abiFALSE-cp310-cp310-linux_x86_64.whl | |
| hf-xet==1.1.5 | |
| torch==2.7.0 | |
| torchvision>=0.22.0 | |
| torchao | |
| opencv-python>=4.9.0.80 | |
| diffusers | |
| transformers>=4.49.0 | |
| tokenizers>=0.20.3 | |
| accelerate>=1.1.1 | |
| tqdm | |
| imageio | |
| easydict | |
| ftfy | |
| dashscope | |
| imageio-ffmpeg | |
| numpy | |
| wandb | |
| omegaconf | |
| einops | |
| av | |
| safetensors | |
| opencv-python | |
| git+https://github.com/openai/CLIP.git | |
| open_clip_torch | |
| starlette | |
| pycocotools | |
| lmdb | |
| matplotlib | |
| sentencepiece | |
| pydantic | |
| scikit-image | |
| huggingface_hub[cli] | |
| dominate | |
| nvidia-pyindex | |
| nvidia-tensorrt | |
| pycuda | |
| onnx | |
| onnxruntime | |
| onnxscript | |
| onnxconverter_common | |
| flask | |
| flask-socketio | |
| aiohttp | |
| Pillow | |