Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 4 years ago.
Improve this question
I tried to copy 17171 files, but whatever parameters I use, it always copied 17160 which 11 lost.
But same command for another directory, copied accurately. (16545 files).
I also tried use cp, but also lost 11 files.
When I check the folder with finder, it should be 17171 files there...
rsync -arvz src dst
cp src dst
Above is the command I've tried
There can be a number of issues at play:
One of the more common issues is that the target filename is illegal on the remote system, for example trying to copy a file with a colon : in the filename from UNIX to Windows.
There may also be permission issues reading the files that are not copied, check the permissions here.
Finally, you could try zipping (or taring) the bunch of files into a single file, and transfer just that instead. Typically you'll see the problem when unpacking that file on the remote system.
EDIT: Another thought - are the files that did not copy really-really large, too large to store remotely?
If you rsync with the -P option, it should only re-transfer files that were not copied. It will also print progress, that should give you a better idea of what's not copying.
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 6 years ago.
Improve this question
I see these files many times when I use the command ls -a . But I want to open that file and access that file. What material contain that file. How it can be done ?
in linux all files whose name starts with . is hidden files.
these files are not shown when you use 'ls', but doing 'ls -a' shows these hidden files too.
if you execute ls -a on your home directory you will see lot of such files. (.profile .bashrc .history etc)
such filenames are given to config files. so most of these files will be in a text format..and may be in other format depending on to which application it belongs to.
to know the format use the following command file filename
to see text files cat filename
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 7 years ago.
Improve this question
Today,I encounter a very tough problem which cost me nearly 6 hours.
When I remove a file called ha_wan.conf using rm -rf ha_wan.conf command under /etc directory,Success.When I use ls -al command to see the result,The file disappear.
But when I reboot the linux system,same file named ha_wan.conf come back,located under /etc/ directory.
I tried to delete it many many times,It is the same result.
What should I do,I want to permanently remove that file.Thanks.
There's no magic. You removed the file. If you still see it after a reboot, it means one of two things:
(very likely) Some service recreates the files on boot, or periodically. You can probably use standard system tools to find out which package contained that file. (for example dpkg -S ha_wan.conf in debian-like systems)
(unlikely) You're running some interesting system which uses a temporary filesystem in /etc. If you're using a standard desktop distribution, that's improbable. But if it's some kind of router / special device, then it could happen.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 1 year ago.
Improve this question
Is there any way to store any data in Linux tmp directory. As i know Linux clear its /tmp directory when system is rebooted. But I want to store data permanently.
Like #bereal said, this defeats the purpose of the /tmp directory. Let me quote the Linux Filesystem Hierarchy Standard:
The /tmp directory must be made available for programs that require temporary files.
Programs must not assume that any files or directories in /tmp are preserved between invocations of the program.
You'll find a better place to store permanent data.
Since it's linux you are free to do what you want to do (as root). When /tmp is cleared depends on your system and can be changed; there is no particular magic involved. A good summary seems to be here: https://serverfault.com/questions/377348/when-does-tmp-get-cleared.
Of course if you are root you can set up an entirely different global directory, say "/not-quite-tmp" or such. But I assume that some progs not under your control write to tmp and you want to inspect or in any case persist those files.
While you are trying to do wrong things, it’s still possible.
/tmp directory is cleared accordigly to TMPTIME setting. The default is apparently 0, what means “clear on every startup”.
The value might be changed in /etc/default/rcS (value is to be set in days.)
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 7 years ago.
Improve this question
I am just preparing a doc file. i am using LibreOffice in ubuntu.
The name of the file is
WebApplicationRequirements.doc
When i am saving that in a drive, an extra file shows in the hidden file.
The name is
.~lock.WebApplicationRequirements.doc#
When i am pushing that in a remote repository it is including that hidden file. If i delete that fill will it harm the original file. And why it is happening?
As the name suggests, that hidden file is a lock file used internally by LibreOffice. To prevent multiple LibreOffice application instances from writing to the same file at the same time. It's not generally harmful to delete that file. It should get re-created again next time you open that file again in LibreOffice.
You haven't indicated what remote repo system you are using. But also note that most repro systems (e.g. git) have the concept of ignore files which allow you to configure which files to ignore during commit. If your repro system has that you probably want to add a rule to ignore the lock files so that they are not committed/pushed.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 9 years ago.
Improve this question
I used the mv command on a file then this message appeared:
mv 'file.tar': Disk quota exceeded
but now there is a file.tar in both the new location and the old location. They're different sizes and I have no idea what the size of the original file was. (There are a lotttt of files in this tarball)
Can I just delete the file that failed to get moved or will I lose part of my original data if I do that?
When Unix/Linux mv(1)-es a file, if source and destination are on the same filesystem, it is just a rename. This can't give an "out of space" message (at least not AFAICS). When moving a file between filesystems (different partitions or disks), what is done is to copy the source to the destination, and when the operation suceeds, the source is deleted. During the copying the destination filesystem can run out of space, and the source won't be deleted. It is strange that the (partial) destination stays in your case, in my experience mv cleans that up (but it's been a while...). In this case the user to which the target file belongs is running out of allowed disk space.