At short - I am getting an error regarding name lookup timed out and don't understand where to search for the problem.
The long story - I've been trying to install MediaWiki Visual Editor, and launched a Heroku based node.js parsoid server for that. I was running the experimental wiki both on localhost and in OpenShift, and it all worked great. However, when I tried to launch it in my main wiki, the editor failed in completing the "round-trip" (wiki-text > html > wiki-text) with the following results:
When re-editing, the editor "forgot" last time edits
Error message - After repeating this action three or four times it resulted in the following error message: parsoid-server-http-request-error: "name lookup timed out".
My main assumption was that it had to do with some inability to cURL the other sever, and therefor tried to make cURL request around the trip but it all worked out with no problem. could it be a firewall I am not aware of?
Second assumption - something to do with server side to client side cache.
Any clue where to search for will be very helpful.
This is what solved this:
shortened the heroku app url
added time (3000) to $wgHTTPConnectTimeout and $wgHTTPTimeout.
The "forgetting" problem mentioned was due to client side cache.
Related
i need help with SonarQube.
I’m trying to use the latest verions (8.9) on CentOS 7, i did all but i’m not able to see the main page because it gets stuck in the loading page after login.
There aren’t error in log file, the unique “error” that i found, if i inspect the source, is:
Application failed to start! SyntaxError: Unexpected token x in JSON at position 556.
Any suggest??
Error:
I solved the problem after time and time. It was so strange by the way after an accuratly check of the network i find the problem.
It was the firewall that blocked some request after login page, i hope this can help those in this same situation..
This is a network issue. The firewall is blocking the sonarqube request
I know it sounds a lot like other issues here in Stackoverflow, bear with me, it's not (not that I could tell)
I have a scraping app (using Puppeteer) that I use to scrape an Amazon public page.
It works great, I've debugged it by setting the headless: false and I see it works, and it gives me back the expected result.
The same app fails on Heroku, but the problem is not with launching or using Puppeteer (I have several indications), but probably because I'm being identified as a robot.
The error returned is:
waiting for selector `#link_continue input` failed: timeout 30000ms exceeded
Important to say that the error is a generic Puppeteer error that indicates that the selector I'm waiting for just doesn't appear on-page.
I know it should as it's a selector on the first page I navigate to, and it works locally (as mentioned before) - the selector always exists if the page loads.
I had the exactly same error when I've tried to run the scraping on my local machine before setting a User-Agent header. But at that time I could use the headless:false so I saw in my eyes that I'm being rejected due to illegal operations on their page (robots-like operations) so I was redirected to an error page that didn't contain this selector on it.
For this reason, I suspect it recognizes me as a robot, but I don't know how to debug it, it drives me crazy.
Now, if you'd like to reproduce the problem:
You need to wait for the mentioned selector on this site:
https://sellercentral.amazon.com/hz/fba/profitabilitycalculator/index
and then deploy it to Heroku and try to run it maybe 2-3 times
** Two questions: **
How can I proceed from here, I'm 99.9% sure it's the same issue I had previously, but I can't verify... any suggestions?
Given that this is actually the problem, can anyone suggest an easy-to-use/deploy hosting that also allow easy VPN configuration? I think Heroku doesn't give you to do that unless you have an enterprise account
Thanks
I would like to point out that Amazon is very good at blocking IPs. It is very likely that they already blacklisted IPs of cloud services like Heroku, Azure, etc... Previously I have observed services like Cloudflare, Akamai etc... blacklisting these known IPs.
In this scenario Rotating proxies could help you to avoid getting blocked.
I am randomly getting a login prompt asking me to log in to localhost:3000 while working on my project (in localhost). Sometimes it does not show up the whole day and next morning I get 10 prompts in a row (canceling through pressing escape), and then it starts working properly again.
These are the errors I get in the console when it has been canceled:
I suspect one of the following, but really I have no idea:
Webpack is not configured correctly
There is some resource that requires authentication (but I can't figure out which), I am running everything on localhost
There is something wrong with the node.js server thingy...
Any tips on what might be going on are very welcome.
Edit: My backend is mocked in localhost.
Edit 2: I found this: Receiving login prompt using integrated windows authentication and it is possible that the thing that is causing the error is some faulty configuration in my IIS. But it is difficult to verify since it only shows up randomly.
Webpack has nothing to do with your API calls.
401 errors are due to auth credentials failing.
It means that your backend ask you to update your token(s) used for auth in API calls.
I think this topic a lot of ppl post it already. I also look for lot of thread in stack overflow and some ibm page. They do give some comment of how to solve this problem. But i still facing the same problem, it still give this error.
Case start :
Actually one of my customer, they have facing this error during running on webpages and also notes ~
error 500 http web server: command not handled exception
Action taken but not solve:
The user id signing/creating the XPages allowed to run XPages in the
server document?
Do project clean, build automatically on designer
Do testing on creating new xpages page with simple lable " hello world" , this one can run successfully only.
Try to reload the http server by
tell http quit
load http
After do this the problem still same give the same error. May i know got any other way to solve it?
"Unresolved compilation errors: java.lang.String" points to a failed install as here. A class in java.lang or java.util should always be resolved. Ensure all services and any processes for Notes / Domino are stopped before the relevant install, jar files can get locked once used. Xpages can't load, Error 500; java.util cannot be resolved
Switch on "Display XPage runtime error page" in application's XSP Properties then you'll see a detailed error message which probably will lead you to the issue:
I frequently get this error when there is a mismatch/problem with access rights or the user login has expired and the app/database doesn't redirect to the login page. I suggest you double check the security basics first. (Might also explain why the app works locally, but not on the server).
So am getting a 409 Conflict with Failed to load resource error. Happens when i click on a form submit of a generic Java web application.
On click of the button there is no application log shows up, just the following error shows up and the process ends. This log is from browser console.
What are the possibilities for such an issue? Any help/suggestion will be really helpful.Help on internet was not much helpful.
Interestingly this is very intermediate i.e. it happens for a device and the next time you won't see it, comes back from grave again in few hrs or so.
Let me start by saying that I have not received this error, but based on your problem statement, it seems you could be having trouble with a file lock on your server. If you are attempting to call a resource that is loaded in another tool like an IDE, then it is possible the file is locked and the server is raising this error.
You may only receive this error if your server is in development mode vs production mode. Production mode will cache the file and not result in a conflict.
Here is a link that partially explains the problem.
HTTP Error 409 Conflict
Hope this helps!!