/
meta.yaml
121 lines (111 loc) · 3.59 KB
/
meta.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
#
# See the guidelines for updating this recipe for new releases:
# https://github.com/scrapy/scrapy/wiki/Scrapy-release-procedure#conda-packages
#
{% set name = "Scrapy" %}
{% set version = "2.11.2" %}
{% set sha256 = "dfbd565384fc3fffeba121f5a3a2d0899ac1f756d41432ca0879933fbfb3401d" %}
{% set blake2b256 = "f21f5524416a64c030fbe18caeba079e7176836b281bf9eb50b79efdf8015063" %}
package:
name: {{ name|lower }}
version: {{ version }}
source:
fn: {{ name }}-{{ version }}.tar.gz
url: https://files.pythonhosted.org/packages/{{ blake2b256[:2] }}/{{ blake2b256[2:4] }}/{{ blake2b256[4:] }}/{{ name|lower }}-{{ version }}.tar.gz
sha256: {{ sha256 }}
build:
script: {{ PYTHON }} -m pip install . --no-deps -vv
number: 0
skip: true # [py2k]
entry_points:
- scrapy = scrapy.cmdline:execute
# The requirements below shall match the requirements from the target package
# version (check setup.py for changes).
requirements:
build:
- python # [build_platform != target_platform]
- cross-python_{{ target_platform }} # [build_platform != target_platform]
host:
- python
- pip
- setuptools
run:
- python
- pywin32 # [win]
# Same order as setup.py:
- twisted >=18.9.0
- cryptography >=36.0.0
- cssselect >=0.9.1
- itemloaders >=1.0.1
- parsel >=1.5.0
- pyOpenSSL >=21.0.0
- queuelib >=1.4.2
- service_identity >=18.1.0
- w3lib >=1.17.0
- zope.interface >=5.1.0
- protego >=0.1.15
- itemadapter >=0.1.0
- setuptools
- packaging
- tldextract
- lxml >=4.4.1
- defusedxml >=0.7.1
- pydispatcher >=2.0.5
test:
requires:
- attrs
- botocore
- brotli
- bpython # [not win]
- dataclasses
- ipython
- google-cloud-storage
- pip
- jmespath
- pillow
- pytest
- testfixtures
- priority >=1.1.0, <2.0
source_files:
- conftest.py
- scrapy
- tests
commands:
- scrapy version -v
- pip install setuptools-scm # required by indirect dependency pytest-forked
# test requirements unavailable on conda:
- pip install sybil
- mv scrapy _scrapy
# https://github.com/scrapy/scrapy/issues/2653#issuecomment-598610517
- rm tests/test_utils_asyncio.py # [win and py==38]
- rm -f tests/test_command_parse.py tests/test_commands.py tests/test_crawler.py tests/test_downloader_handlers.py tests/test_downloader_handlers_http2.py # [osx]
# Test failures since Scrapy 2.4:
- rm -f tests/test_command_check.py tests/test_command_parse.py tests/test_commands.py tests/test_crawler.py tests/test_feedexport.py
# test_convert_image fails because expected vs actual color does not match
# by 1 bit (#0079FF vs #0080FF)
- rm -f tests/test_pipeline_images.py
# MediaPipelineDeprecatedMethodsTestCase leaves files that interfere with its tests
- rm -f tests/test_pipeline_media.py
# The removed tests/test_pipeline_media.py file is a dependency
- rm -f tests/test_pipeline_files.py
# Started failing in 2.11.1
- rm -f tests/test_command_shell.py
# Tests need to be fixed in upstream.
- pytest _scrapy tests
about:
home: https://scrapy.org/
license: BSD-3-Clause-Clear
license_file: LICENSE
summary: A high-level Python Screen Scraping framework
description: |
Scrapy is an open source and collaborative framework for extracting the
data you need from websites in a fast, simple, yet extensible way.
doc_url: https://docs.scrapy.org
dev_url: https://github.com/scrapy/scrapy
extra:
recipe-maintainers:
- rmax
- kmike
- Gallaecio
- wRAR
- dangra