I was trying to run npm link on a nodeJs project but it fails on running nsp check command at the end, giving the following error:
(+) Response Error: 408 Request Timeout
How can i fix that ?
408 Error indicates that the server did not receive a complete request from the client within the server’s allotted timeout period.
Please check your Internet connection and try again.
Related
I HAVE A ERROR IN INSTALLING PYTORCH:
PLEASE HELP ME.
CondaHTTPError: HTT P000 CONNECTION FAILED for url https://conda.anaconda.org/pytorch/win-64/current_repodata.json
Elapsed: -
An HTTP error occurred when trying to retrieve this URL.
HTTP errors are often intermittent, and a simple retry will get you on your way.
'https//conda.anaconda.org/pytorch/win-64'
An HTTP error occurred when trying to retrieve this URL. HTTP errors are often intermittent, and a simple retry will get you on your way
The possible reason for HTTP error could be an unstable network connection or a corporate firewall.
If it was an unstable network connection, as mentioned in the Error message, retry the installation steps that failed.
If you are behind a corporate firewall, you might need additional steps to add your proxy server to the .condarc file on your machine.
Since you are on Windows, you could open the Anaconda prompt and run conda info to figure out where the .condarc file is located.
Find the proxy by running echo "$http_proxy" in your prompt. Copy the proxy.
Open the .condarc file and paste the proxy under proxy_servers section
For more details see: Anaconda Docs: Configure conda for use behind a proxy server (proxy_servers)
I'm trying to compile Linux for an embedded system. I'm using ptxdist command to build the system.
My colleague told me to use ptxdist get so we can get all dependencies downloaded before we start compiling.
We also sit behind a proxy and my /etc/environment is configured with the http/https/ftp proxy (same IP and port for all), as well as apt (meaning that wget works fine without extra parameters)
PROBLEM:
Running ptxdist get will download some of the packages, others it will just not download and the following error appears (example for 2 packages):
Here is an example when it worked fine:
---------------------------
target: cmake-3.13.4.tar.gz
---------------------------
--2019-07-23 20:54:28-- https://cmake.org/files/v3.13/cmake-3.13.4.tar.gz
Connecting to <PROXY IP HERE>:8080... connected.
Proxy request sent, awaiting response... 200 OK
Length: 8617881 (8.2M) [application/x-gzip]
Saving to: '/home/xxxxxxx/xxxxxxx/xxxxxxx/src/cmake-3.13.4.tar.gz.RSjrC7k7mC'
/home/xxxxxxx/x 100%[===================>] 8.22M 2.00MB/s in 4.7s
2019-07-23 20:54:34 (1.75 MB/s) - '/home/xxxxxxx/xxxxxxx/xxxxxxx/src/cmake-3.13.4.tar.gz.RSjrC7k7mC' saved [8617881/8617881]
And here an example when it failed:
-----------------------
target: lzo-2.08.tar.gz
-----------------------
--2019-07-23 20:55:13-- http://www.oberhumer.com/opensource/lzo/download/lzo-2.08.tar.gz
Connecting to <PROXY IP HERE>:8080... connected.
Proxy request sent, awaiting response... 502 Proxy Error ( The specified network name is no longer available. )
2019-07-23 20:55:13 ERROR 502: Proxy Error ( The specified network name is no longer available. ).
--2019-07-23 20:55:13-- http://www.pengutronix.de/software/ptxdist/temporary-src/lzo-2.08.tar.gz
Connecting to <PROXY IP HERE>:8080... connected.
Proxy request sent, awaiting response... 502 Proxy Error ( The specified network name is no longer available. )
2019-07-23 20:55:13 ERROR 502: Proxy Error ( The specified network name is no longer available. ).
make: *** [/home/xxxxxxxx/xxxxxxx/xxxxxxx/src/lzo-2.08.tar.gz] Error 1
Could not download package
URL: http://www.oberhumer.com/opensource/lzo/download/lzo-2.08.tar.gz
/usr/local/lib/ptxdist-2019.03.1/rules/post/ptxd_make_world_get.make:17: recipe for target '/home/xxxxxxx/xxxxxxx/xxxxxxx/src/lzo-2.08.tar.gz' failed
Even if I repeat ptxdist get, it stops at the same exact package that fails.
BUT:
If I just use
wget http://www.oberhumer.com/opensource/lzo/download/lzo-2.08.tar.gz
it downloads the file just fine, same happens to ANY other package that fails, using the URL that shows up above.
When I run ptxdist get then it moves on to other downloads as it sees the file already there, and then hangs on the next download that failes.
ALSO TRIED:
I also tried to run ptxdist setup and configure the proxy there. Nothing changes in the terminal, so no luck...
ALSO TRIED:
I also wrote a script that takes package names using ptxdist print PACKAGES, then gets the URL using ptxdist package-info <PKG> then downloads the file via wget <URL>, this solved most part of the errors while running ptxdist get as wget works fine, but it seems that there are many other packages that are needed and not listed with ptxdist print PACKAGES...
Any idea what is wrong and what can I do?
Since I'm supposed to automate this task, there is no way manual interaction (for using wget on failing packages) is an option...
Description
I'm crawling the website :bjx.com and all codes can be run in the local.Then I put the code on the Amazon service and run ,it failed.
What I have Done
I guess that maybe the website block the server and I have tried some ways :
1) curl http://guangfu.bjx.com.cn/xtgc/List.aspx?classid=583
2) wget http://guangfu.bjx.com.cn/xtgc/List.aspx?classid=583
err msg as follows:
Resolving news.bjx.com.cn (news.bjx.com.cn)... 114.113.145.103
Connecting to news.bjx.com.cn (news.bjx.com.cn)|114.113.145.103|:80... failed: Connection timed out.
Retrying.
--2019-04-23 05:45:00-- (try: 2) http://news.bjx.com.cn/list
Connecting to news.bjx.com.cn (news.bjx.com.cn)|114.113.145.103|:80...
some reference:
https://serverfault.com/questions/124952/testing-a-website-from-linux-command-line
My question :
how to confirm whether the website has blocked me and if blocked, what can I do to solve the issue and crawl the website, thanks
How about to make the program failed with particular timeout setting?
For example, to make curl failed if it can't get response within 10 seconds
curl -m 10
And, to go over these issues, you can try to run the spiders with Proxy of VPN networking
This call (imap_open()) has been failing consistently when I run the script inside a docker container run by bitbucket pipeline.
PHP Warning: imap_open(): Couldn't open stream {imap.gmail.com:993/imap/ssl/novalidate-cert}INBOX in /opt/atlassian/pipelines/agent/build/test/tools/plib/confirm.php on line 24
PHP Fatal error: Uncaught exception 'Exception' with message 'signup: confirm failed' in /opt/atlassian/pipelines/agent/build/test/tools/plib/signup.php:24
Stack trace:
#0 /opt/atlassian/pipelines/agent/build/test/tools/test.php(51): signup(Array)
#1 {main}
thrown in /opt/atlassian/pipelines/agent/build/test/tools/plib/signup.php on line 24
PHP Notice: Unknown: Retrying PLAIN authentication after [ALERT] Please log in via your web browser: https://support.google.com/mail/acco (errflg=1) in Unknown on line 0
PHP Notice: Unknown: Retrying PLAIN authentication after [ALERT] Please log in via your web browser: https://support.google.com/mail/acco (errflg=1) in Unknown on line 0
PHP Notice: Unknown: Can not authenticate to IMAP server: [ALERT] Please log in via your web browser: https://support.google.com/mail/acco (errflg=2) in Unknown on line 0
Cannot connect to Gmail: Can not authenticate to IMAP server: [ALERT] Please log in via your web browser: https://support.google.com/mail/acco
I have followed all these instructions:
Enabled less secure apps
Enabled from https://accounts.google.com/b/0/DisplayUnlockCaptcha
Tried with and without /novalidate-cert flag
Same script works fine when run locally within the Mac OS or even with AWS EC2 instance but fails when run by bitbucket or Heroku. There is no way I can run a browser on these instances, so cannot try web interface and apparently once enabled for access, it should work every where.
Too bad, the link in the error message isn't even complete.
Any idea how to overcome this? All I want to do is to simply simulate a click on the verify link in a signup email programmatically.
As mentioned , google does alert "Less secured apps" and authentication keeps failing even after that sometimes.
Best solution here is to move to OAUTH2 auth method, else even if you find a way to solve this "less secured alert" issue... it may eventually occur in future on and off.
Is there a way to verify Websphere Node status from Linux CLI using one of IBM shell scripts?
I've been using this way to get this information from CLI since i did not find a script that does this for me, but i get an 500 error since i did not login yet to DMGR (After login this will result with an answer:
wget "http://ap0tv01tap01ndsw0.ap0tv.tas.att.net:9060/ibm/console/status?text=true&type=node&node=ap0tv01tap02mlsw0"
--2013-07-04 18:00:07-- http://ap0tv01tap01ndsw0.ap0tv.tas.att.net:9060/ibm/console/status?text=true&type=node&node=ap0tv01tap02mlsw0
Resolving ap0tv01tap01ndsw0.ap0tv.tas.att.net... 172.20.31.12
Connecting to ap0tv01tap01ndsw0.ap0tv.tas.att.net|172.20.31.12|:9060... connected.
HTTP request sent, awaiting response... 500 Internal Server Error
2013-07-04 18:00:07 ERROR 500: Internal Server Error.
Alternative solution for me could be triggering what Web login triggers to operation system and do it from CLI.
Thanks!
You may use serverStatus command to query status of node agent. You may use both nodeagent as server name as an argument or use -all flag. If security is enabled in your cell don't forget to specify username and password as mentioned in the documentation.