You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi there, I installed SerpBear on Synology per https://mariushosting.com/how-to-install-serpbear-on-your-synology-nas/. Deployed via Portainer with the specified Docker compose. I used the specified Synology settings, kept port 3000, but the one thing I did differently was to setup the synology reverse proxy using a custom domain and a Let'sEncrypt certificate. I signed up and tried both ScrapingRobot and ScrapingAnt APIs according to SerpBear documentation. Entered APIs as outlined in documentation. Updated settings, Restarted stacks in Portainer. Login to SerpBear without issue. All seems to be working fine. I can enter keywords and can click to update serps but no data is fetched. I know some of the keywords we rank in top 10 (and a few higher) but all show no data and just "> 100"
Checked docker logs on Synology and get this (since this is public, in this past of the log I used dummy keywords and changed the domain to mydomain.com):
[0] syscall: 'getaddrinfo',
[0] code: 'EAI_AGAIN',
[0] errno: -3001,
[0] at GetAddrInfoReqWrap.onlookup [as oncomplete] (node:dns:107:26) {
[0] cause: Error: getaddrinfo EAI_AGAIN api.scrapingant.com
[0] at async refreshAndUpdateKeywords (/app/.next/server/chunks/816.js:110:23) {
[0] at async refreshKeywords (/app/.next/server/chunks/816.js:232:28)
[0] at async Promise.all (index 0)
[0] at async scrapeKeywordFromGoogle (/app/.next/server/chunks/816.js:378:87)
[0] at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
[0] at Object.fetch (node:internal/deps/undici/undici:11413:11)
[0] [ERROR_MESSAGE]: TypeError: fetch failed
[0] [ERROR] Scraping Keyword : keyword #2 . Error: undefined
[0] GET /api/keywords?domain=mydomain-com
[0] POST /api/keywords
[0] GET /api/keywords?domain=mydomain-com
[0] [SUCCESS] Updating the Keyword: Keyword 3
[0] }
[0] path: '/app/data/failed_queue.json'
[0] syscall: 'open',
[0] code: 'ENOENT',
[0] errno: -2,
[0] [Error: ENOENT: no such file or directory, open '/app/data/failed_queue.json'] {
[0] time taken: 5168.428608059883ms
[0] ALL DONE!!!
[0] }
[0] }
[0] hostname: 'api.scrapingant.com'
[0] syscall: 'getaddrinfo',
[0] code: 'EAI_AGAIN',
[0] errno: -3001,
[0] at GetAddrInfoReqWrap.onlookup [as oncomplete] (node:dns:107:26) {
[0] cause: Error: getaddrinfo EAI_AGAIN api.scrapingant.com
[0] at async refreshAndUpdateKeywords (/app/.next/server/chunks/816.js:110:23) {
[0] at async refreshKeywords (/app/.next/server/chunks/816.js:232:28)
[0] at async Promise.all (index 0)
[0] at async scrapeKeywordFromGoogle (/app/.next/server/chunks/816.js:378:87)
[0] at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
[0] at Object.fetch (node:internal/deps/undici/undici:11413:11)
[0] [ERROR_MESSAGE]: TypeError: fetch failed
[0] [ERROR] Scraping Keyword : Test Keyword . Error: undefined
[0] GET /api/keywords?domain=mydomain-com
[0] POST /api/keywords
[0] GET /api/keywords?domain=mydomain-com
[0] GET /api/settings
[0] domains: 1
[0] GET /api/domains?withstats=true
[0] POST /api/domains
[0] GET /api/settings
[0] domains: 0
[0] GET /api/domains?withstats=true
Did some searching and it appears that the code "EAI_AGAIN" may be linked to a DNS fail? I can't understand what that would be. Any other ideas on why it appears I can't connect with EITHER scraping API site?
I checked the data files and they are being written -- but there's nothing in them -- I assume because it's not connecting to pull data.
Thoughts are welcome.
The text was updated successfully, but these errors were encountered:
Hi there, I installed SerpBear on Synology per https://mariushosting.com/how-to-install-serpbear-on-your-synology-nas/. Deployed via Portainer with the specified Docker compose. I used the specified Synology settings, kept port 3000, but the one thing I did differently was to setup the synology reverse proxy using a custom domain and a Let'sEncrypt certificate. I signed up and tried both ScrapingRobot and ScrapingAnt APIs according to SerpBear documentation. Entered APIs as outlined in documentation. Updated settings, Restarted stacks in Portainer. Login to SerpBear without issue. All seems to be working fine. I can enter keywords and can click to update serps but no data is fetched. I know some of the keywords we rank in top 10 (and a few higher) but all show no data and just "> 100"
Checked docker logs on Synology and get this (since this is public, in this past of the log I used dummy keywords and changed the domain to mydomain.com):
[0] syscall: 'getaddrinfo',
[0] code: 'EAI_AGAIN',
[0] errno: -3001,
[0] at GetAddrInfoReqWrap.onlookup [as oncomplete] (node:dns:107:26) {
[0] cause: Error: getaddrinfo EAI_AGAIN api.scrapingant.com
[0] at async refreshAndUpdateKeywords (/app/.next/server/chunks/816.js:110:23) {
[0] at async refreshKeywords (/app/.next/server/chunks/816.js:232:28)
[0] at async Promise.all (index 0)
[0] at async scrapeKeywordFromGoogle (/app/.next/server/chunks/816.js:378:87)
[0] at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
[0] at Object.fetch (node:internal/deps/undici/undici:11413:11)
[0] [ERROR_MESSAGE]: TypeError: fetch failed
[0] [ERROR] Scraping Keyword : keyword #2 . Error: undefined
[0] GET /api/keywords?domain=mydomain-com
[0] POST /api/keywords
[0] GET /api/keywords?domain=mydomain-com
[0] [SUCCESS] Updating the Keyword: Keyword 3
[0] }
[0] path: '/app/data/failed_queue.json'
[0] syscall: 'open',
[0] code: 'ENOENT',
[0] errno: -2,
[0] [Error: ENOENT: no such file or directory, open '/app/data/failed_queue.json'] {
[0] time taken: 5168.428608059883ms
[0] ALL DONE!!!
[0] }
[0] }
[0] hostname: 'api.scrapingant.com'
[0] syscall: 'getaddrinfo',
[0] code: 'EAI_AGAIN',
[0] errno: -3001,
[0] at GetAddrInfoReqWrap.onlookup [as oncomplete] (node:dns:107:26) {
[0] cause: Error: getaddrinfo EAI_AGAIN api.scrapingant.com
[0] at async refreshAndUpdateKeywords (/app/.next/server/chunks/816.js:110:23) {
[0] at async refreshKeywords (/app/.next/server/chunks/816.js:232:28)
[0] at async Promise.all (index 0)
[0] at async scrapeKeywordFromGoogle (/app/.next/server/chunks/816.js:378:87)
[0] at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
[0] at Object.fetch (node:internal/deps/undici/undici:11413:11)
[0] [ERROR_MESSAGE]: TypeError: fetch failed
[0] [ERROR] Scraping Keyword : Test Keyword . Error: undefined
[0] GET /api/keywords?domain=mydomain-com
[0] POST /api/keywords
[0] GET /api/keywords?domain=mydomain-com
[0] GET /api/settings
[0] domains: 1
[0] GET /api/domains?withstats=true
[0] POST /api/domains
[0] GET /api/settings
[0] domains: 0
[0] GET /api/domains?withstats=true
Did some searching and it appears that the code "EAI_AGAIN" may be linked to a DNS fail? I can't understand what that would be. Any other ideas on why it appears I can't connect with EITHER scraping API site?
I checked the data files and they are being written -- but there's nothing in them -- I assume because it's not connecting to pull data.
Thoughts are welcome.
The text was updated successfully, but these errors were encountered: