Skip to main content

Why Firefox 3.1 can wait

The FOSS community is rejoicing another release of FF - 3.1beta. Mozilla released FF3.1beta on tuesday. I was one of those beta-ers last time around during FF3 release. I even did countdown to the release. I even got my own download certificate :). However, I don't want to eager myself to it this time. Here is why its different this time.

Firstly, its not a major release, meaning nothing changed BIG! If at all, The only thing that will get me motivated is the CTRL+Tab feature. But my existing Tab-Mix Plus plugin gives a tab list that is adequate. A 3D preview is nice to have but not so much to hurry to.

All the goodies are unused, The geocoding, Video, Audio, HTML5 and CSS3 support have nothing to offer until there are some REAL websites that start utilizing them. Especially the Geocoding, As much as it sounds cool, until it becomes a standard - it would be another browser specific dumb-a$$-nerdy feature I would refrain myself as a webmaster. The others follow the same rationale.

The much boasted JIT javascript is not enabled by default. WHAT? Why? I think 3.0 has a huge performance improvement from 2.0 that I only doubt this is going to matter for an end user. Would I like to try this out? Yeah sure! But it can wait until my Ubuntu auto updater hits the new update!

Mozilla needs to start a different approach for distributing betas. The default beta should always be an independent zip or tar file. Like one of those Portable Apps. This will make me go out there and check it out. Right out of box, No - doubt, fear or uncertainty. You know what I mean..

Finally, I was more interested in the private browsing to come. I am disappointed it dint make it in this beta. I will wait. So will the betas :)

Popular posts from this blog

Powered By

As it goes, We ought to give thanks to people who power us. This page will be updated, like the version page , to show all the tools, and people this site is Powered By! Ubuntu GIMP Firebug Blogger Google [AppEngine, Ajax and other Apis] AddtoAny Project Fondue jQuery

Decorator for Memcache Get/Set in python

I have suggested some time back that you could modularize and stitch together fragments of js and css to spit out in one HTTP connection. That makes the page load faster. I also indicated that there ways to tune them by adding cache-control headers. On the server-side however, you could have a memcache layer on the stitching operation. This saves a lot of Resources (CPU) on your server. I will demonstrate this using a python script I use currently on my site to generate the combined js and css fragments. So My stitching method is like this @memize(region="jscss") def joinAndPut(files, ext): res = files.split("/") o = StringIO.StringIO() for f in res: writeFileTo(o, ext + "/" + f + "." + ext) #writes file out ret = o.getvalue() o.close() return ret; The method joinAndPut is * decorated * by memize. What this means is, all calls to joinAndPut are now wrapped (at runtime) with the logic in memize. All you wa...

Faster webpages with fewer CSS and JS

Its easy, have lesser images, css and js files. I will cover reducing number of images in another post. But If you are like me, You always write js and css in a modular fashion. Grouping functions and classes into smaller files (and Following the DRY rule, Strictly!). But what happens is, when you start writing a page to have these css and js files, you are putting them in muliple link rel=style-sheet or script tags. Your server is being hit by (same) number of HTTP Requests for each page call. At this point, its not the size of files but the number server roundtrips on a page that slows your page down. Yslow shows how many server roundtrips happen for css and js. If you have more than one css call and one js call, You are not using your server well. How do you achieve this? By concatinating them and spitting out the content as one stream. So Lets say I have util.js, blog.js and so.js. If I have a blog template that depends on these three, I would call them in three script tags. Wh...