-
Notifications
You must be signed in to change notification settings - Fork 33
/
setup.py
46 lines (39 loc) · 1.51 KB
/
setup.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
import os
from setuptools import setup, find_packages
# Utility function to read the README file.
# Used for the long_description. It's nice, because now 1) we have a top level
# README file and 2) it's easier to type in the README file than to put a raw
# string in below ...
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
with open('requirements.txt') as f:
required = f.read().splitlines()
setup(
name = "selenium-crawler",
version = "0.1.0",
packages = find_packages(),
# Project uses reStructuredText, so ensure that the docutils get
# installed or upgraded on the target machine
install_requires=required,
package_data = {
# If any package contains *.txt or *.rst files, include them:
'': ['*.txt', '*.rst'],
# And include any *.msg files found in the 'hello' package, too:
'hello': ['*.msg'],
},
# metadata for upload to PyPI
author = "Cory Walker",
author_email = "[email protected]",
description = ("Sometimes sites make crawling hard. Selenium-crawler uses "
"Selenium automation to fix that."),
license = "LICENSE.txt",
keywords = "selenium crawling crawl automate ads landing",
url = "https://github.com/cmwslw/selenium-crawler",
long_description=read('README.md'),
test_suite = "seleniumcrawler.tests.test_all",
classifiers=[
"Development Status :: 3 - Alpha",
"Topic :: Utilities",
"License :: OSI Approved :: MIT License",
],
)