Start a new topic

Cant deploy spider

I get this when trying to deploy my spider:

  

Login succeeded
.
.
.

Successfully built f2bb6e18d4db
Entrypoint container is created successfully
>>> Checking python dependencies
No broken requirements found.
>>> Getting spiders list:
>>> Trying to get spiders from shub-image-info command
WARNING: There're some errors on shub-image-info call:
{"message": "", "error": "internal_error"}

  Any ideas?


Have similar problem. Here is my log:


  

Error: Deploy failed: b'{"status": "error", "message": "Internal error"}'
 ---> d66bd6bc900f
Removing intermediate container 5161ba153fd3
Step 11/12 : RUN if [ -d "/app/addons_eggs" ]; then rm -f /app/*.dash-addon.egg;
 fi
 ---> Running in efbc0db2efa2
 ---> 5c7e190770fe
Removing intermediate container efbc0db2efa2
Step 12/12 : ENV PATH /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/b
in
 ---> Running in 26838a17677e
 ---> 108961c0b30e
Removing intermediate container 26838a17677e
Successfully built 108961c0b30e
Step 1/3 : FROM alpine:3.5
 ---> 6c6084ed97e5
Step 2/3 : ADD kumo-entrypoint /kumo-entrypoint
 ---> Using cache
 ---> bdb8d4874ea6
Step 3/3 : RUN chmod +x /kumo-entrypoint
 ---> Using cache
 ---> 702ef707423a
Successfully built 702ef707423a
Entrypoint container is created successfully
>>> Checking python dependencies
No broken requirements found.
>>> Getting spiders list:
>>> Trying to get spiders from shub-image-info command
WARNING: There're some errors on shub-image-info call:
Exceeded container timeout 60s
{"message": "shub-image-info exit code: -1", "details": null, "error": "image_in
fo_error"}
{"status": "error", "message": "Internal error"}

 Does someone know what is wrong ? 

Hello Simen,


Can see that there are successful deploys in your account. Would be great if you can mention as how you resolved the issue and help others with similar errors.


Regards,

Thriveni Patil

Unfortunately I don't know what the problem was. I created new empty scrapy project and deployed it. Then I was copying my main project to an empty one piece by piece and it worked.


In my main project I have some redundant scripts and spiders. Maybe they cause this problem. My advice would be to delete all of the unnecessary scripts.


Regards,

Artur

 I think the error is

Exceeded container timeout 60s


I find the solution from here

https://github.com/scrapinghub/shub/issues/273


It can deploy project to scrapinghub now, but i still get another problem in my case.


When i try the second solution, my process will no inactive.


if __name__ == '__main__':
    process = CrawlerProcess()
    process.crawl(MySpider1)
    process.crawl(MySpider2)
    process.start() # the script will block here until all crawling jobs are finished

Hope any one try the solution work can share what the should type.


Login to post a comment