I'm building a spider using Portia and would like to display items on my website. I've a few ideas in mind on how to do that and would appreciate any help in deciding witch is the best way to go:
Method 1:
Set up a cron job that fetches data from Scrapy Cloud API
store the data in a local db
fetch data from local db to be displayed for users.
Method 2: Display data for users directly using Scrapy Cloud API.I'm not sure this is feasible for a number of reasons:
Scrapy Cloud API requests might be different for each job.
Scrapy Cloud API may not support features as us filtering and searching.
Method 3: Setup a dataset and request data from this dataset. Is there any way to access datasets via an API?
Thanks in advance.
0 Votes
tom posted
over 7 years ago
AdminBest Answer
Another good option is on your site to have a rest api set in place that allows your spider to send the data direct to that rest api that then can be processed to show on your web site.
Another good option is to dump your data from the spider to a s3 bucket that then could be picked up by a cron job on your site.
0 Votes
2 Comments
Sorted by
tomposted
over 7 years ago
AdminAnswer
Another good option is on your site to have a rest api set in place that allows your spider to send the data direct to that rest api that then can be processed to show on your web site.
Another good option is to dump your data from the spider to a s3 bucket that then could be picked up by a cron job on your site.
0 Votes
vazposted
over 7 years ago
Hi Sano,
I wouldn't recommend gave that decision to our side.
I suggest to try both methods and then choose that better fits for your project. If you want you can share with us which one you choose and why.
That could be extremely useful for other users with similar inquiries.
Hello,
I'm building a spider using Portia and would like to display items on my website. I've a few ideas in mind on how to do that and would appreciate any help in deciding witch is the best way to go:
0 Votes
tom posted over 7 years ago Admin Best Answer
Another good option is on your site to have a rest api set in place that allows your spider to send the data direct to that rest api that then can be processed to show on your web site.
Another good option is to dump your data from the spider to a s3 bucket that then could be picked up by a cron job on your site.
0 Votes
2 Comments
tom posted over 7 years ago Admin Answer
Another good option is on your site to have a rest api set in place that allows your spider to send the data direct to that rest api that then can be processed to show on your web site.
Another good option is to dump your data from the spider to a s3 bucket that then could be picked up by a cron job on your site.
0 Votes
vaz posted over 7 years ago
Hi Sano,
I wouldn't recommend gave that decision to our side.
I suggest to try both methods and then choose that better fits for your project. If you want you can share with us which one you choose and why.
That could be extremely useful for other users with similar inquiries.
Thanks for making this Community better.
Best regards,
Pablo
0 Votes
Login to post a comment