New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
unable to deploy with scrapyd-deploy #128
Comments
I'm not sure exactly. I was able to recreate your issue but once I had resolved it I couldn't recreate it again. |
Could you give me a step by step on how I would go about doing that ? I have next to zero experience with python/scrapy and did everything by following tutorials.
|
You probably didn't kill slyd correctly. |
i couldn't deploy because i didn't have the required argument "project name" |
Is it all sorted now then? |
no. I didn't understand what i needed to do and was waiting for some guidance. What should i do / check ? |
Change directory to the project Then it should be available on scrapyd (scrapyd-deploy -L local) |
Still not working... |
@MihaiCraciun also could you please paste your slyd/data/projects/new_project/scrapy.cfg? it can give more information |
cat scrapy.cfg
|
(venv)192-168-0-197:new_project Mihai$ scrapyd-deploy local -p new_project |
that doesn't looks right. could you also paste the slyd/data/projects/new_project/setup.py? |
The original file looked like this:
But :
Again, i deleted build, eggs and so on, and placed back the original setup.py (firs in this comment). Did a build and got : |
i can access http://localhost:6800/ just fine, all the commands i'm running are inside my virtual environment, in the project folder (new_project).. :( |
did you accidentally remove the slyd/data/projects/new_project/spider/settings.py? could you paste that too? |
i had the same issue and reported on #100 i never got it to work but thought it was me that was doing something wrong since i never used scrapy/scrapyd before |
here it is (new_project/spiders/settings.py)
|
I can remove the whole project and do it again .. just one question : can i have both portia and scrapyd running at the same time ? I tried to startup portia in another terminal window while having scrapyd running and got an error that there is another twistd application running. |
@MihaiCraciun yes, you can |
@MihaiCraciun Before you delete the whole project would you mind just deleting the |
I tried that earlier and it failed. I'm just now rebuilding the project and giving it another go. |
good : extraction is working i portia. Now.. can anybody tell me step by step what i should do .. so that i don't mess this up again ? |
you can find the steps on https://github.com/scrapinghub/portia/blob/master/README.md#deploying-a-project, if you find something unclear/obscure i'm glad to update it. |
im gonna re-install scrapyd and see if i can get it to work this time |
That's what got me in this mess the first time :)) To deploy on localhost:
And then schedule your spider with: curl http://localhost:6800/schedule.json -d project=your_project_name -d spider=your_spider_name |
IT'S FREAKIN' IMPOSSIBLE !!!!
Please ... HELP ! I'm out of ideas... |
Did you try it through curl If that doesn't work would you mind uploading your zip file so I can see it there's anything weird about it. |
There's nothing strange about the project.
Then could you post your output from the deploy here |
Traceback (most recent call last): |
I forgot a %s in the script. I've updated it above. would you mind trying again |
here it is :
|
I'm working in a virtual environment.. i don't know why it's trying to go in system var (or at least that's what i understand) |
I've no idea why it's doing that either, I'm not really familiar with how scrapyd works. |
that's my project folder from which i'm supposed to run scrapyd-deploy |
Looks like you are running them from the same folder then. |
HOLY C%@p it worked !!! What now ? :) |
I'll have to add it to the docs as a gotcha. |
What should I read next on actually using my spider ? I want to run my spider and save the data scraped to a database |
Schedule your spider through the API and you can monitor it through the web interface |
scrapinghub#128 Update docs to warn users not to run scrapyd in their project direc...
Hello,
Could you please help me figure out what I'm doing wrong ? Here are the steps:
i followed the portia install manual - all ok
i created a new project, entered an url, tagged an item - all ok
clicked "continue browsing", browsed through site, items were being extracted as expected - all ok
Next i wanted to deploy my spider:
1st try : i tried to run, as the docs specified, scrapyd-deploy your_scrapyd_target -p project_name - got error - scrapyd wasn't installed
fix: pip install scrapyd
2nd try : i launched scrapyd server (also missing from the docs), accessed http://localhost:6800/ -all ok
after a brief reading of scrapyd docs i found out i had to edit the file scrapy.cfg from my project : slyd/data/projects/new_project/scrapy.cfg
added the following :
[deploy:local]
url = http://localhost:6800/
went back to the console, checked all is ok :
$:> scrapyd-deploy -l
local http://localhost:6800/
$:> scrapyd-deploy -L local
default
seemed ok so i gave it another try :
$scrapyd-deploy local -p default
Packing version 1418722113
Deploying to project "default" in http://localhost:6800/addversion.json
Server response (200):
{"status": "error", "message": "IOError: [Errno 21] Is a directory: '/Users/Mihai/Work/www/4ideas/MarketWatcher/portia_tryout/portia/slyd/data/projects/new_project'"}
What am I missing ?
The text was updated successfully, but these errors were encountered: