| from setuptools import setup |
|
|
| with open("README.md", "r", encoding="utf-8") as readme_file: |
| readme = readme_file.read() |
|
|
| requirements = [ |
| "numpy", |
| "scipy", |
| "matplotlib", |
| "torch", |
| "torchvision", |
| "opencv-python", |
| "CLIP @ git+https://github.com/openai/CLIP.git" |
| ] |
|
|
| setup( |
| name='clipseg', |
| packages=['clipseg'], |
| package_dir={'clipseg': 'models'}, |
| package_data={'clipseg': [ |
| "../weights/*.pth", |
| ]}, |
| version='0.0.1', |
| url='https://github.com/timojl/clipseg', |
| python_requires='>=3.9', |
| install_requires=requirements, |
| description='This repository contains the code used in the paper "Image Segmentation Using Text and Image Prompts".', |
| long_description=readme, |
| long_description_content_type="text/markdown", |
| ) |
|
|