How to resolve Oracle RemoteOperationException: Error reading error from command - linux

I have recently been receiving the following error whenever I am asked to supply the host username and password in Oracle DataGuard's Enterprise Manager (EM) tool: RemoteOperationException: Error reading error from command. Any configuration or management that needs to be performed and requires the host credentials is throwing this error. This is one of the most unhelpful errors I have seen in a long time. I have checked all of the log files I can think of and I can't find any error logs or indications to the problem. I verified that I can log into the servers with the credentials that I was trying in EM, I verified the connection configuration through EM using the built in test tools, and I verified that all of the saved passwords in EM were correct. This was working on the initial installation. Sometime over the last couple of weeks it stopped working and I'm not sure why. The hosts are running Linux Red Hat 4 Enterprise and Oracle 10g.

Reinstalling the OEM agents fixed this issue. Somehow, all of the agents seemed to be corrupt.

Related

System crashes while using clearcase 8.0.1.x /9.0.1.x (checking out files) on windows 10 (1803) platform

After upgrading system to Windows 10 - os 1803 we are getting below issues while working with ClearCase 8.0.1.x/9.0.1.x
Unable to checkin/checkout.
Not able to create views.
Not able to add any file to source control.
The system hangs & crashes while performing any ClearCase operation.
There is no error message, but I have attached screenshot for reference.
Please let us know if there is any issue with the Windows 10 ver(1803), any security system enabled?
Or has ClearCase provided any fix?
We have tried 9.0.1.5 and issue still persists.
This is what we got from windows event log.
The computer has rebooted from a bugcheck.
The bugcheck was:
0x000000c2 (0x0000000000000004, 0x00000000535be990, 0x000000000004efd3, 0xfffff803e01848b1)
for most of them whoever has upgraded to windows 1803 ver :( for people who are still using ver1709 it is working perfectly fine
Then I would recommand contacting IBM support: only them can update their ClearCase 9/Windows 10 compatibility matrix and confirm if MVFS is supported on a more recent (1803) Windows 10 edition.
We also facing same problem and I have raised the case with IBM. Still not yet resolved. As IBM said there are some limitations to work ClearCase with windows 10 and windows 2016.
We tried all the options except Secure boot disable. If possible please do disable secure boot option in Windows 10 and try to checkin/checkout code from CleraCase
Note : It works for Snapshot views. That means the issue related to MVFS
I'm seconding #VonC's recommendation to open a ticket with IBM. When you do that, save a step and collect a clearbug2 and a kernel memory dump to send in as soon as the case is opened. It will save the turn-around time of us asking you for it. If the installed programs list doesn't list installed security software (DLP, Privilege management sw like Avecto, other endpoint security tools), please list those separately as well.
I would also love to know who # IBM told you there are "limitations" with Win10-1803.
There are a few issues with Windows 10 "version upgrades" breaking things, but they generally don't cause system crashes. Windows 10 upgrades are actually full OS installs that then (imperfectly) migrate application settings. Anything that uses custom network providers (ClearCase is one example) will find that the network providers will be broken or partially broken. Reinstalling is usually required. Again, that has not yet been reported as a cause of a BSOD.
If the upgrade/reinstall didn't fix view creation, please post a separate question on the view creation issue. There may be things we can do to the SMB 2 caches to allow view creation to work in cases where the view storage is not on the client host.
I noticed that the screen shot you posted is a Terminal Services disconnect screenshot. Does the issue only occur over a Terminal Services client connection or does it also happen on a local connection?

Trouble Logging into Server: call to lnusertemp failed (temporary directories full ?).Check your installation

We are running a Kubuntu Server, It has been running without any problems for a long time.
However, When attempting to login to the server through the GUI it throws the following error message:
lnusertemp failed (temporary directories full ?). Check your installation
We have much free room on the drives and cannot figure out the cause of this issue. We are able to login through a remote cli application (using putty).
To my knowledge, we haven't updated or modified any core files or settings to this server, so its sudden change in operation has given us with some concern. This server is an important unit to our working process.

Cannot connect with FreeNX to Ubuntu ec2 instance

I have recently spun up a new Ubuntu 12.04 instance in AWS. I had no issues connecting to and opening an SSH terminal to the server. Having connected to the instance, I was able to install the Ubuntu desktop and FreeNX without any problem, as well as enabling password authentication on the server instance.
I downloaded and installed the NX Client for Windows on a PC running Windows 8. After entering the user credentials I can connect to and authenticate into the server. I'm brand new to the Linux world, but at this point everything was going so smoothly I was about ready to throw my Windows licenses to the dogs - good thing I held off on that.
"Problem: At the client, the !M logo window appears, but after a few seconds that window just closes, even without showing any error message."
That problem statement is in quotes because it's precisely the issue described in FreeNX Ubuntu Community support documentation https://help.ubuntu.com/community/FreeNX#Troubleshooting.
So naturally I follow the solution in the guide:
"Solution: The issue is due custom VNC configuration. In the server, access your home directory and run these commands,"
sudo rm .Xauthority*
touch .Xauthority
chmod 600 .Xauthority
Unfortunately, this did absolutely nothing to resolve the issue. The problem would be easier to diagnose if I had an error message, but reference the Problem statement, there is no error message to be had. Several hours of googling yielded nothing, so wondering if anyone here has encountered this problem in the past, and if so, they would be willing to help.
Thanks!

Resolving Mass-loading problems in WebSphere Commerce Instance creation

I am trying to create an instance using the Configuration Manager of WCS 7. I am working on a Win 7 x64 machine with DB2 9.5 64 bit version.
I am struck with this Massloading error when the instance creation happens :
In createInstanceANT.log file :
[Massload] Massloading
C:\IBM\WebSphere\CommerceServer\schema\xml\wcs.keys.xml Error in
MassLoading, please check logs for details.
The error log shows the following error :
[jcc][10165][10044][4.3.111] Invalid database URL syntax:
jdbc:db2://:0/WCSDEMO. ERRORCODE=-4461, SQLSTATE=42815
C:\IBM\WEBSPH~1\COMMER~2\config\DEPLOY~1\xml\createBaseSchema.xml:185:
Error in massloading
WCSDEMO is the database name. The Massloader is not able to get the URL and port to connect. It is supposedly getting them from createInstance.properties file but it is not working. The createInstance.properties file has all the details of the DB to connect.
What could be the reason for this error and how to resolve it ? Is there any configuration change that I am missing ?
Can you provide some more details.
look inside the messages.txt file located in WC_install_dir/instances/instance_name/logs
and confirm what the exact issue is. If it is related to jdbc driver being wrong I may be able to help you.
I've been running into massloading problems with external systems. Eg. databases not on the same machine as the WAS installation.
In these cases I look for the
As you can see setting the loaderDBName to just the name of the database would look on the local machine. But by changing this statement so you load with the syntax
loaderDBName=[DATABASE_SERVER_NAME]:[PORT]/[DATABASE_NAME]
You'll be able to massload using the commerce standard scripts. These changes needs to be done in many scripts. Both for updating fixpacks and enabling features. If you run database updates without the changes it will crash at first and have done all the schema changes to the database that you then need to comment out before trying again.
IBM Software Support is your friend. They'll help you fix it.

Cognos 8.4.1 St9bad_alloc message : how to solve?

For eval purpose; I installed Cognos 8.4.1 on Ubuntu 10.04.
The DB for the content store is Oracle 11g with AL32 UTF8 character set.
Crypto keys are stored locally. Env is set as recommended.
When trying to start the service, even before creating the content store, I get this message :
CM-CFG-5076 A Content Manager configuration error occurred during the initialization of Cognos Access Manager. CAM-AAA-0071 An internal error occurred. St9bad_alloc
Any idea of the cause ?
Thanks,
Fred
Using a supported distribution would definitely be a good idea. That said, if you stop cognos, rename the cogserver.log file (by default this is in /logs under your installation directory), start cognos up again and post the generated log file we might get more detail about what's stopping it.

Resources