Have similar problem. Here is my log:
Error: Deploy failed: b'{"status": "error", "message": "Internal error"}' ---> d66bd6bc900f Removing intermediate container 5161ba153fd3 Step 11/12 : RUN if [ -d "/app/addons_eggs" ]; then rm -f /app/*.dash-addon.egg; fi ---> Running in efbc0db2efa2 ---> 5c7e190770fe Removing intermediate container efbc0db2efa2 Step 12/12 : ENV PATH /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/b in ---> Running in 26838a17677e ---> 108961c0b30e Removing intermediate container 26838a17677e Successfully built 108961c0b30e Step 1/3 : FROM alpine:3.5 ---> 6c6084ed97e5 Step 2/3 : ADD kumo-entrypoint /kumo-entrypoint ---> Using cache ---> bdb8d4874ea6 Step 3/3 : RUN chmod +x /kumo-entrypoint ---> Using cache ---> 702ef707423a Successfully built 702ef707423a Entrypoint container is created successfully >>> Checking python dependencies No broken requirements found. >>> Getting spiders list: >>> Trying to get spiders from shub-image-info command WARNING: There're some errors on shub-image-info call: Exceeded container timeout 60s {"message": "shub-image-info exit code: -1", "details": null, "error": "image_in fo_error"} {"status": "error", "message": "Internal error"}
Does someone know what is wrong ?
Hello Simen,
Can see that there are successful deploys in your account. Would be great if you can mention as how you resolved the issue and help others with similar errors.
Regards,
Thriveni Patil
Unfortunately I don't know what the problem was. I created new empty scrapy project and deployed it. Then I was copying my main project to an empty one piece by piece and it worked.
In my main project I have some redundant scripts and spiders. Maybe they cause this problem. My advice would be to delete all of the unnecessary scripts.
Regards,
Artur
I think the error is
Exceeded container timeout 60s
I find the solution from here
https://github.com/scrapinghub/shub/issues/273
It can deploy project to scrapinghub now, but i still get another problem in my case.
When i try the second solution, my process will no inactive.
if __name__ == '__main__': process = CrawlerProcess() process.crawl(MySpider1) process.crawl(MySpider2) process.start() # the script will block here until all crawling jobs are finished
Hope any one try the solution work can share what the should type.
Simen Riiser
I get this when trying to deploy my spider:
Any ideas?