Upwork Scraper Without Stale Job Posts avatar
Upwork Scraper Without Stale Job Posts

Pricing

$9.99/month + usage

Go to Store
Upwork Scraper Without Stale Job Posts

Upwork Scraper Without Stale Job Posts

Developed by

Artur

Artur

Maintained by Community

Comes without any stale data like many other scrapers do. Low cost and efficient. Can filter out duplicates and posts older than 24h

5.0 (1)

Pricing

$9.99/month + usage

16

Total users

172

Monthly users

51

Runs succeeded

96%

Issue response

2.9 days

Last modified

a month ago

CB

Is this still working?

Closed

caring_bear opened this issue
2 months ago

I was using it for a dew days, and from nowhere, it stoped working - I get a 403 error, and it's not showing results anymore.

Is there someone who can take a look?

Thx

CB

caring_bear

2 months ago

More exactly it's been 5 days that I don't get the results.

arlusm avatar

Artur (arlusm)

2 months ago

It's an issue with the apify proxy, most likely it has been blacklisted, so you'll need to try with a different country code proxy. Recently the french ones seem to have gotten blocked as well. I tried with DE for the moment and it worked.

arlusm avatar

Artur (arlusm)

2 months ago

I think i'll add in a possibility to specify various proxy countries and the option to retry with new proxy countries automatically if one fails

arlusm avatar

Artur (arlusm)

2 months ago

Also you might want to use the key "&per_page=50" at the end of your URL, so that you get 50 requests at once, and won't have to check other pages and thereby reduce the risk of getting flagged by cloudflare/upwork

CB

caring_bear

2 months ago

Thx Artur! Is there any other way to use some kind of residential proxies or something similar to make sure I don't get into this problem again?

arlusm avatar

Artur (arlusm)

2 months ago

The scraper does use residential proxies. If you're using automation software you could check for failure and then try querying again with another country code.

CB

caring_bear

2 months ago

Thanks for the update. I tried changing the country, but after only a few executions, I started getting 403 errors again: https://monosnap.com/file/QXcYGl7c5jdkY3BVBFalqi9uJlitwy

It looks like Upwork is detecting the scraping pattern and blocking the IPs fairly quickly. Does this mean they are actively identifying and banning the proxies? If so, what can be done to avoid this?

Also, is there a list of available countries I can use for proxy rotation? That would help in managing the retries more effectively.

Thanks!

CB

caring_bear

2 months ago

I also implemented what you recommended, so I'm changing the country and separating each url to be sent with another country. There's still no consistency: https://monosnap.com/file/hwCKe7UNYGNDczNtAvB7yT5YdhnD1J

JO

jorge.lewis

2 months ago

Heya Artur, when could we expect this to be fixed? Cheers!

arlusm avatar

Artur (arlusm)

2 months ago

Looking into it right now, a quick fix is to try with another proxy. Sadly apify doesn't provide a list of what countries are available. In the docs they've only provided the two letter country code list: https://laendercode.net/en/2-letter-list.html

JO

jorge.lewis

2 months ago

legend. Good luck! "AI" worked btw.

CB

caring_bear

2 months ago

Hey Artur,

Any news on this? Any progress?

Best,

arlusm avatar

Artur (arlusm)

2 months ago

Hey.

I added an User Cookies field. You should match it with the proxy country code. So if you're located in Sweden for example, you should add SE for the proxy country.

To get the cookie:

-Open Upwork & Navigate to Job Search Open Brave/Chrome/Chromium and go to: https://www.upwork.com/nx/search/jobs/?nbs=1&q=stripe&page=4 Make sure you're logged in to Upwork. -Open Developer Tools Right-click anywhere on the page and select "Inspect" OR Press F12 or Ctrl + Shift + I (Windows/Linux) Press Cmd + Option + I (Mac)

  • Capture the Network Request In Developer Tools, go to the "Network" tab. In the search bar, filter by "jobs" Click on the request that matches: jobs/?nbs=1&q=stripe&page=4 -Copy Cookies Click on the "Headers" tab. Scroll down until you see "Cookie" under Request Headers. Select and copy everything after Cookie:.
  • Paste Cookies in Your Scraper Use the copied cookies for the input
ET

endless_text

2 months ago

Just tried with the cookie, most of the searches are still blocked, but is successfully retrieving some of the pages! Thanks for keeping up on this, Artur.

arlusm avatar

Artur (arlusm)

a month ago

Was working on a solution today and even though I didn't get the solution totally working, requests seemed to go through. I suspect upwork made it easier to scrape once again. I'll close this issue for the moment. If you're still experiencing any issues, create a new issue or reply to other open ones. Thanks.