Skip to main content

How to remove empty lines and weird characters from MANY files

Say you have a file (a print) and need to clean it up, remove weird control caracters: how do you do it?
First you look at an hex dump of the file with hexdump -C: then you take note of the control caracters and their hex code.
After that you simply replace them with blanks or strip them off, depending on taste or requirements.
In the end you put together a big for loop like the following:

for f in directory/* ; do

# replace control characters
perl -pi -e 's/[\x0b\x01\x0c]//g' $f;
dos2unix $f;
# remove empty lines or lines with only blanks
grep -v '^[ \t]*$' $f > tmp;
mv tmp $f;
unix2dos $f;



Popular posts from this blog

Indexing Apache access logs with ELK (Elasticsearch+Logstash+Kibana)

Who said that grepping Apache logs has to be boring?

The truth is that, as Enteprise applications move to the browser too, Apache access logs are a gold mine, it does not matter what your role is: developer, support or sysadmin. If you are not mining them you are most likely missing out a ton of information and, probably, making the wrong decisions.
ELK (Elasticsearch, Logstash, Kibana) is a terrific, Open Source stack for visually analyzing Apache (or nginx) logs (but also any other timestamped data).

From 0 to ZFS replication in 5m with syncoid

The ZFS filesystem has many features that once you try them you can never go back. One of the lesser known is probably the support for replicating a zfs filesystem by sending the changes over the network with zfs send/receive.
Technically the filesystem changes don't even need to be sent over a network: you could as well dump them on a removable disk, then receive  from the same removable disk.

RUNDECK job maintenance

Learn more about Rundeck.

Now that I have a fair number of jobs scheduled by Rundeck, how do I periodically prune the job execution history and keep only the last, say, 30 executions for each job?