Skip to main content

Quickly: Command history in Linux / Shell

If you are like me, doing a lot of command line in linux (Bash) shell, there are a few commands you do very often. The following commands will make your life a lot easier to quickly go back in command history.

[!!] or better yet [sudo !!] - [!!] is an alias for last run command. One of the best use of this is - Sometimes, you type a command and the shell hits you back with a root privileges required message. You dont have to copy the command and run again. just run [sudo !!]

[history] - just spits out ~/.bash_history with line numbers. A good use of this is piping it to grep to find something. The line numbers can be used to execute the command following it. Say for example, [history | grep find] lists the previous searches (okay, it also shows all commands with "find" anywhere in the line - but a filtered, smaller list).
$history | grep find
 457  find *.wav
 462  find *.wav -exec lame --preset fast extreme '{}' '/tmp/mp3/{}' \;
 487  find *.wav -exec mp3cvbr '{}' \;

To rerun the search just type !nnn where nnn is the number of the line shown, eg: !462.

And the best one..
[CTRL+R] Dont *type* it - press control + r on your keyboard at shell prompt. It opens a prompt that will start a reverse search on history of commands. Then, Type any part of the command (or arguments) that you have done in the past, it will fill in the whole command including the arguments. Hit Enter to execute it!!

Popular posts from this blog

Powered By

As it goes, We ought to give thanks to people who power us. This page will be updated, like the version page , to show all the tools, and people this site is Powered By! Ubuntu GIMP Firebug Blogger Google [AppEngine, Ajax and other Apis] AddtoAny Project Fondue jQuery

Decorator for Memcache Get/Set in python

I have suggested some time back that you could modularize and stitch together fragments of js and css to spit out in one HTTP connection. That makes the page load faster. I also indicated that there ways to tune them by adding cache-control headers. On the server-side however, you could have a memcache layer on the stitching operation. This saves a lot of Resources (CPU) on your server. I will demonstrate this using a python script I use currently on my site to generate the combined js and css fragments. So My stitching method is like this @memize(region="jscss") def joinAndPut(files, ext): res = files.split("/") o = StringIO.StringIO() for f in res: writeFileTo(o, ext + "/" + f + "." + ext) #writes file out ret = o.getvalue() o.close() return ret; The method joinAndPut is * decorated * by memize. What this means is, all calls to joinAndPut are now wrapped (at runtime) with the logic in memize. All you wa...

Faster webpages with fewer CSS and JS

Its easy, have lesser images, css and js files. I will cover reducing number of images in another post. But If you are like me, You always write js and css in a modular fashion. Grouping functions and classes into smaller files (and Following the DRY rule, Strictly!). But what happens is, when you start writing a page to have these css and js files, you are putting them in muliple link rel=style-sheet or script tags. Your server is being hit by (same) number of HTTP Requests for each page call. At this point, its not the size of files but the number server roundtrips on a page that slows your page down. Yslow shows how many server roundtrips happen for css and js. If you have more than one css call and one js call, You are not using your server well. How do you achieve this? By concatinating them and spitting out the content as one stream. So Lets say I have util.js, blog.js and so.js. If I have a blog template that depends on these three, I would call them in three script tags. Wh...