| """setup.py for axolotl""" |
|
|
| import platform |
| import re |
| from importlib.metadata import PackageNotFoundError, version |
|
|
| from setuptools import find_packages, setup |
|
|
|
|
| def parse_requirements(): |
| _install_requires = [] |
| _dependency_links = [] |
| with open("./requirements.txt", encoding="utf-8") as requirements_file: |
| lines = [r.strip() for r in requirements_file.readlines()] |
| for line in lines: |
| is_extras = ( |
| "flash-attn" in line |
| or "flash-attention" in line |
| or "deepspeed" in line |
| or "mamba-ssm" in line |
| or "lion-pytorch" in line |
| ) |
| if line.startswith("--extra-index-url"): |
| |
| _, url = line.split() |
| _dependency_links.append(url) |
| elif not is_extras and line and line[0] != "#": |
| |
| _install_requires.append(line) |
|
|
| try: |
| if "Darwin" in platform.system(): |
| _install_requires.pop(_install_requires.index("xformers==0.0.23.post1")) |
| else: |
| torch_version = version("torch") |
| _install_requires.append(f"torch=={torch_version}") |
|
|
| version_match = re.match(r"^(\d+)\.(\d+)(?:\.(\d+))?", torch_version) |
| if version_match: |
| major, minor, patch = version_match.groups() |
| major, minor = int(major), int(minor) |
| patch = ( |
| int(patch) if patch is not None else 0 |
| ) |
| else: |
| raise ValueError("Invalid version format") |
|
|
| if (major, minor) >= (2, 3): |
| _install_requires.pop(_install_requires.index("xformers==0.0.23.post1")) |
| _install_requires.append("xformers>=0.0.26.post1") |
| elif (major, minor) >= (2, 2): |
| _install_requires.pop(_install_requires.index("xformers==0.0.23.post1")) |
| _install_requires.append("xformers>=0.0.25.post1") |
| except PackageNotFoundError: |
| pass |
|
|
| return _install_requires, _dependency_links |
|
|
|
|
| install_requires, dependency_links = parse_requirements() |
|
|
|
|
| setup( |
| name="axolotl", |
| version="0.4.0", |
| description="LLM Trainer", |
| long_description="Axolotl is a tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures.", |
| package_dir={"": "src"}, |
| packages=find_packages(), |
| install_requires=install_requires, |
| dependency_links=dependency_links, |
| extras_require={ |
| "flash-attn": [ |
| "flash-attn==2.5.8", |
| ], |
| "fused-dense-lib": [ |
| "fused-dense-lib @ git+https://github.com/Dao-AILab/flash-attention@v2.5.8#subdirectory=csrc/fused_dense_lib", |
| ], |
| "deepspeed": [ |
| "deepspeed==0.14.2", |
| "deepspeed-kernels", |
| ], |
| "mamba-ssm": [ |
| "mamba-ssm==1.2.0.post1", |
| ], |
| "auto-gptq": [ |
| "auto-gptq==0.5.1", |
| ], |
| "mlflow": [ |
| "mlflow", |
| ], |
| "lion-pytorch": [ |
| "lion-pytorch==0.1.2", |
| ], |
| "galore": [ |
| "galore_torch", |
| ], |
| }, |
| ) |
|
|