I have a meteor 1.5 application which had no problems last night, everything worked just fine, all routes defined by iron:router were working. Today, out of nowhere and without changing any code, i have this weird error page of iron:router saying that i have no routes defined.
I have done some research and found out that the latest version of iron:router have some conflicts and it causes problems, some suggested that i downgrade iron:router, i have done that but it didn't work.
i even tried removing all my routes and putting one simple route for the homepage with url " / " but still get the same error page
Any ideas ?
This message was also encountered by users when Chrome moved from version 50 to 51. Have you tried different browsers to see if the issue is specific to a certain vendor? (git topic)
Rgs, Paul
Related
I know it sounds a lot like other issues here in Stackoverflow, bear with me, it's not (not that I could tell)
I have a scraping app (using Puppeteer) that I use to scrape an Amazon public page.
It works great, I've debugged it by setting the headless: false and I see it works, and it gives me back the expected result.
The same app fails on Heroku, but the problem is not with launching or using Puppeteer (I have several indications), but probably because I'm being identified as a robot.
The error returned is:
waiting for selector `#link_continue input` failed: timeout 30000ms exceeded
Important to say that the error is a generic Puppeteer error that indicates that the selector I'm waiting for just doesn't appear on-page.
I know it should as it's a selector on the first page I navigate to, and it works locally (as mentioned before) - the selector always exists if the page loads.
I had the exactly same error when I've tried to run the scraping on my local machine before setting a User-Agent header. But at that time I could use the headless:false so I saw in my eyes that I'm being rejected due to illegal operations on their page (robots-like operations) so I was redirected to an error page that didn't contain this selector on it.
For this reason, I suspect it recognizes me as a robot, but I don't know how to debug it, it drives me crazy.
Now, if you'd like to reproduce the problem:
You need to wait for the mentioned selector on this site:
https://sellercentral.amazon.com/hz/fba/profitabilitycalculator/index
and then deploy it to Heroku and try to run it maybe 2-3 times
** Two questions: **
How can I proceed from here, I'm 99.9% sure it's the same issue I had previously, but I can't verify... any suggestions?
Given that this is actually the problem, can anyone suggest an easy-to-use/deploy hosting that also allow easy VPN configuration? I think Heroku doesn't give you to do that unless you have an enterprise account
Thanks
I would like to point out that Amazon is very good at blocking IPs. It is very likely that they already blacklisted IPs of cloud services like Heroku, Azure, etc... Previously I have observed services like Cloudflare, Akamai etc... blacklisting these known IPs.
In this scenario Rotating proxies could help you to avoid getting blocked.
I am using pjax 2.0.0 by cdn and after adding
this line of code $(document).pjax('a', '#pjax-container');
and when I go to check the header this is Request Url
http://laravelclear.dev/replace-page?_pjax=
and my response is 500 the current node list is empty
I have searched alot please help me.
I did a lot of search in many site and my problem was asked but there was no answer for this question for and finally I found a way for using pjax and my problem was crawler version for laravel middleware pjax I used
https://github.com/spatie/laravel-pjax this one and for crawler I used this command
composer require spatie/crawler
and this is working fine
I hope this can help others.
I have several Azure functions published and all but 2 of them work fine. The two that don't work are named "AdminData" and "AdminImage" and will be used to feed data to an administrative page. The code compiles fine, deploys fine, everything in the Azure dashboard looks fine but when I try calling them I always get the 404 Not Found error.
It turns out the issue is that an Azure function name can't start with "Admin". I only found this out after hours of trial and error because no errors or warnings were thrown when the code was compiled or deployed. A subsequent search once I knew what the problem was turned up an issue report in github so it's a known problem;
Cannot have [FunctionName] starting with 'admin' #141
This was a very frustrating problem that took a lot of time to figure out so hopefully, this post will help someone else avoid this trap.
The Heroku app i'm trying to get to work (code here):
https://github.com/heroku/facebook-template-nodejs
"Unsafe Javascript attempt to access frame with URL" errors occur when the page is loaded in chrome.
The login button takes you to facebook but does not actually log you into the app and gives the same errors.
Has anyone got this app to work on Chrome or can anyone advise as to how to patch it up?
P.S. it seems to work fine on Mozilla.
Almost certain this is a cross domain policy issue, as stated above. Generally speaking, you just need to add the correct header info to the response.
Access-Control-Allow-Origin: *
In Node, I think it is just a matter of adding it as another header in the response, using
response.writeHead
See http://nodejs.org/api/http.html#http_response_writehead_statuscode_reasonphrase_headers
Oh, and there's explicit instructions on how to do it if you're using Express. I see no reason why it can't work using plain old node then.
http://enable-cors.org/server_expressjs.html
So I looked at your link, in your case I think you just have to enter the header info prior to using any other express app methods.
As to why it works in Firefox and not Chrome, not sure. Both support CORS many versions back. Maybe you have some Chrome extension that's interfering.
I have a website, today looking at the log, I found some request to a page from my server with appended this variable &sa=U&ei.
Could you tell me guys what &sa=U&e could mean? Could be an attempt to find Nullable Scrips? Could be a security threat?
&sa=U&ei=XuRBT92UFseYhQf_w7HeBQ&ved=0CNYBEBYwYw&sig2=Rt1Cr_FCPD1-6VYu__Oavg&usg=AFQjCNFlHVaDQL--kgDbOn2vNgUqwUOsTA
The error in my log is:
A potentially dangerous Request.Path value was detected from the client (&)
http://www.bleepingcomputer.com/forums/topic442637.html
But nevermind. I found the offending culprit. Seems that for some
reason my "GoogleEnhancer" became "incompatible" with Firefox. It
worked fine even before I updated to 10, but go figure. And it wasn't
the whole add-on, it was the "Use Google Classic" radio button turned
on. I got this add-on after Google started making their search engine
so... oh, what's the word I'm looking for... umm... oh, yeah...
crappy!
http://www.ausforces.com/showthread.php?6595-Google-is-acting-odd...
I figured out what it is... I have an add-on called google enhancer
which obviously hasn't been updated properly of late. Disabled it and
it worked fine. Well that was a waste of a thread. Thanks for the help
though guys.
So, the unnecessary part of the request is created by an outdated add-on for Firefox, nothing serious. The visitors with that add-on have more problems than your website :)
Practically, it could just be simple url request.
And the random texts you are seeing could be an autogenerated random string to maintain sessions. As there seems nothing wrong with the URL, and those sa and ei simply means, that these get variables are assigned some values, which would then be used in your application, for may be session management or other purpose.
From the face of it, it doesn't appears to be any hackable stuff.