Error creating IBM Containers 404: Resource not found - apache-spark

Good night , does anyone can help me , I'm trying to create an IBM Containers for my app , but is experiencing an error in the end.
404: Resource not found

You can get to the logs of your container in the Bluemix Dashboard.
First of all, in the Dashboard you see a green/yellow/red indicator on your container tile/icon indicating if your container is running or has an issue.
When you click on the container tile, in the left menu you will see a Monitoring and Logs item. In the logging tab make sure the time "window" is large enough, there is this pull down in the upper right part which by default displays only the last 30 minutes of the logs.
You can also get to the logs from the command line with "cf ic logs "
If the logs don't help you, can you be more specific in what you attempt to do?

If you get this error when running cf ic login you have to check the cf api https://api.[region].bluemix.net URL you are using, where [region] could have ng or eu-gb value according to the region used (us or UK).
If the error you got is requesting a page from your service running on the container, it is depending from your container, so first check the ip used, then the path requested

Related

Getting 500 - Internal server error in Azure Website after restart

I have a web application in azure(App Service), recently i have restarted the app, after restaring i am getting:
500 - Internal server error.
There is a problem with the resource you
are looking for, and it cannot be displayed.
Please let me know what could be the issue. application was working fine till that time. I didnt change any other settings.
How to resolve this issue?
Go to the app service, then pick Monitoring -> App Service Logs. Then enable Application Logging (FileSystem) and save the changes. Now go into Monitoring -> Log Stream and watch the Application Logs while the site is starting up, this will most likely give the detailed error message.
There are a few things you can do step by step
Step 1: Use the KUDU console at https://sitename.scm.azurewebsites.net. Check if there are errors. More info about KUDU
Step 2: Have you turned on the detailed error message in the Diagnostics logs tab of the Azure portal. If not enable and see if you can track any error.This will give you a .zip file. Browse to LogFiles->DetailedErrors and here you should see some error pages.
Step 3: You can add Application Insights to your application and log those errors when you start the application
If nothing works, try restarting app service, or access it from a different location (because it happened to me earlier)

You do not have permission to view this directory or page Azure and FTP (Filezilla error)

I am getting the following error message when I try to view an azure website I created: 'You do not have permission to view this directory or page.'
When I click on the URL I created in Microsoft Azure, it also disconnects me from the FTP that I have set up with this error messsage: 'Disconnected from server: ECONNABORTED - Connection aborted' I thought I fixed this setting by going into Filezilla FTP Settings and changing it from Passive to Active
I don't know if these things are related to each other or causing one of the things to happen, but I don't understand what to do or what is broken?
Update: I got the error message 'You do not have permission to view this directory or page.' to work by changing the name of my html file from 'index2.html' to 'index.html' but it isn't playing my html file correctly. It has changed into a small black box, rather than filling up the whole screen and it is just a black box with a play button, but it doesn't play.Azure website
I fixed it. I uploaded my video to Wistia and then added the Inline Embed code to my index.html file, then added my files to Filezilla, and then the hosting site from Azure finally worked.
I am glad you fixed the issue. If you see the issue again, could you please restart the website and check once.
For testing/debugging purposes you can turn on the Azure detailed messaging, and turn back off when its ready for production. To do so, you have to follow these two steps,
Login to Azure > App Services (left side menu) > Your Web App > Diagnostics logs (search box is at the top if you can't find it), then turn on Detailed Error Messages or turn on all of the loging options, up to you.
Now add the following in your Web Config file,
In your Web Config file add BEFORE system.web closing tag, . Similarly, add BEFORE . Finally, upload the Web Config to Azure and cross your fingers.
If you follow the steps correctly, that will show the error messages in detail and hopefully from there

azure 502 bad gateway

has anyone seen this before so I am getting a 502 bad gateway error on my app, the issue I have is that the detailed error information I am getting says my requested url is https://SOX:80/api however my site is configured to use https://sox.domain.com and the site largely works pulling the various JS files required
my app service name is SOX in the azure dashboard so I assume that is where it is picking up SOX from but I have no idea why it is using this.
So overall the issue had me perplexed... however with more testing I soon figured out what was going on.
my backend is Dotnet core Azure throwing the 502 bad gateway was its way of handling exceptions ultimately the problem was code based.
I am mentioning this purely so that it will help others
my first issue was based on cert handling it seems dotnet runs in a container that is specified by your app name as i mentioned above https://SOX:80
the below was causing my issues
sslPolicyErrors = X509StoreStoreHelper.ValidateSSLPolicy(cert.Thumbprint, cert);
after commenting this out for testing my problem went away(we are putting in a proper fix )
my second issue came from using an unsupported view in Azure SQL master.sys.master_files which again just threw a 502 bad gateway error referencing https://SOX:80
please note I have used https://SOX:80 as a reference to mask the real site.
hope this helps the next person.
Based on your description, I have checked your site (https://sox.azurewebsites.net/) and found that it contains three static files (index.html,generic.html,elements.html). I viewed your website in Chrome incognito window as follows:
I did not find any requests against https://SOX:80/api in your html page or JavaScript files. Please try to access your website in a new incognito window to isolate the cache issue or just press CTRL + F5 to refresh your current page to narrow this issue. Moreover, you need to check whether you have configured URL Rewrite. If you still could not solve this issue, you need to update your question with the details for us to reproduce this issue.

Azure function logging doesn't appear to be working

I'm able to successfully call my functions and make them do what I want them to do. The problem is that it doesn't look like the logs are being saved anywhere and I don't see how I can view them. Which I'll want to do in the event of an error. As a test I have my working function just doing a log.Info as soon as it's called. When testing locally it prints the message to the console. I believe I've enabled everything correctly but let me explain what I've done in case I didn't.
In my app service, under Monitoring -> Diagnostic Logs, I have enabled everything. Application Logging (filesystem) verbose, Application Logging (Blob) verbose (with the storage location set), detailed error messages and failed request tracing turned on.
In my function, I'm using the TraceWriter object that's passed to my run method (I started from a template).
Please note that functions are set to require authentication. If I click on the "Monitor" tab nothing appears. It just says "Loading..." forever and there's no information. Perhaps this is because of the authentication?
I used the Azure Storage Explorer to browse to my blob. The "log" blob exists, and I do see a set of nested directories that lead up to now. However it just contains a 354 byte file that contains a few lines of some random info. This file never seems to update or get larger.
I used FTP to try and browse to where the logs might be, but there's no directory on there that contains any log files.
I also went to KUDU for my function app ({myfunctionapp}.scm.azurewebsites.net/azurejobs/#/functions). While I do see that my function was called successfully, I don't see anything from the call to log.Info anywhere.
I tried using a different logger, and as a test did: System.Diagnostics.Trace.TraceError("test error");
I also don't see this message appearing anywhere.
Am I missing something as far as set up goes? Is the problem the fact that I require authentication? If it's the latter, is there still a way to view logs? I definitely have to have auth enabled. Thanks. And if it helps, below are links to what my settings and the monitor tab look like.
Settings: https://postimg.org/image/u57m2xbl5/
Monitor: https://postimg.org/image/uou10arch/
Authentication should not cause any problems with logging and Log.Info should work out of the box, no setup required.
I highly recommend that you enable AlwaysOn for your dedicated function app. The long loading of the Monitor tab could be because your site is in a 'cold' state, where it takes longer to start up.
If you go to {myfunctionapp}.scm.azurewebsites.net/DebugConsole and navigate to LogFiles/Application/Functions do you see any expected logs there? Also, when you run a function from the portal do you see logs in the log window?
Same thing happened to me if I had Fiddler open....close Fiddler and all is good.

Where can I see the nodejs logs after I deployed on Google App Engine?

I deployed a nodejs app on Google App engine following this tutorial https://github.com/GoogleCloudPlatform/appengine-nodejs-quickstart it was successful and now I want to check the logs of the nodejs server, like in development from the terminal console. The Vms are managed by google but even if I ssh to them I don't know where to look for the logs.
You can read the stdout of the docker container that your app runs by doing docker logs <container id> in the VM instance. You can get the container id from docker ps.
No need to SSH into the instance though. You can simply fetch the logs from the Developers Console under Monitoring > Logs.
As #tamberg mentioned in a comment, the easiest option I've found for looking at logs produced by Google App Engine instances running Node.js is to simply use the log viewer at:
https://console.cloud.google.com/logs/viewer?resource=gae_app
Detailed instructions from https://cloud.google.com/appengine/docs/standard/nodejs/building-app/viewing-service-logs are as follows:
Open the Logs Viewer in the GCP Console: https://console.cloud.google.com/logs/viewer?resource=gae_app
In the first filter dropdown at the top of the page, ensure GAE Application is selected, and choose Default Service.
Use the second filter dropdown to select only stdout and click OK. This limits the viewer to logs sent to standard output.
Use the text field above the dropdowns to search for the name you used when you submitted your form. You should see the logs corresponding to your submissions.
The default logging is really awful. None of my console.log messages show up! There are a few ways you can fix this.
1) Write logs to a log file.
For example, /var/log/app_engine/custom_logs/applogs.log
https://cloud.google.com/appengine/articles/logging
"Cloud Logging and Managed VMs apps Applications using App Engine
Managed VMs should write custom log files to the VM's log directory at
/var/log/app_engine/custom_logs. These files are automatically
collected and made available in the Logs Viewer. Custom log files
must have the suffix .log or .log.json. If the suffix is .log.json,
the logs must be in JSON format with one JSON object per line. If the
suffix is .log, log entries are treated as plain text."
2) Use winston with winston-gae.
Create a transport that will send the logs to appengine.
3) Use gcloud-logging module
Too verbose for my liking, but it is another option.

Resources