Author Archives: specialk

Not bad for an old fart

 Today on Mont Ventoux, Lance Armstrong was able to hold on to his 3rd place finish.

Andy Schleck and Alberto Contador were clearly the best two riders of the tour, but Lance was very much the third best guy on the road.

The real fun, though, was listening to today’s broadcast because Bradley Wiggins of the UK was trying to desperately hold on to fourth.

Phil Liggett was practically jumping up and down cheering for Wiggins. Telling him to stay in there, push harder, and came up with the best phrase:

He’s holding on to 4th by the skin of his bike shorts!

Wiggins benefited from the fact that Frank Schleck didn’t have the legs, whereas Frank benefited from the fact that his brother was too far behind Alberto Contador to be able to mount a realistic challenge for the overall lead. So with Alberto and Andy stuck in first and second, it was a battle for the 3-4-5 positions of the overall lead. Andy tried to create attacks that Frank would follow but the magical legs that Frank had two days ago had disappeared. Instead after every attack Andy would have to stop and wait for his brother.

Furthermore, Frank was not going to benefit from the tactical screw up of the other day on stage 17 that lead to a surprising stage win:

What had happened was that  Lance was following Wiggins all day. On the climb to the Col de Romme, Andy attacked and Alberto and Kloden followed. Lance realizing that he wasn’t going to challenge for the overall GC and that Alberto had a team mate let that attack go. A few seconds later when the attack had not yet consolidated, the group behind the leaders more or less stopped. Lance looked behind him, watching both Wiggins and Frank. He didn’t want to be pacing his challengers for the podium. When Frank attacked, Lance was at too much of a dead stop to catch up. Making the tactical decision that he could limit the damage Frank could create on his overall lead, and believing in his performance in the time-trial, Lance let Frank go, deciding to keep his eyes on Wiggins. Lance, correctly, surmised, that the multi-minute gap was simply too big for Frank.

But today with only 30 or so seconds separating Lance and Frank, there was to be no such screw up. And indeed there was not.

At the end of the tour, when Versus finally caught up with the Texan, he looked pleased with himself. Like anyone who comes in third, and realizes there was no chance for first, he was happy with his accomplishment, and said:

Not bad for an old fart like me.

And the Spaniard Ends an Era

For even the unsurpassed, the great, Lance Armstrong, age finally arrived.

image

Amazingly able to keep up with great climbers a decade younger than him, Lance seemed poised to triumph in this year’s tour.

But the tour, has a way of separating the champions from the pretenders. For the first time in almost 11 years, Lance Armstrong felt like his competitors must have felt as he crushed them up mountains, leaving them gasping for air and wondering what went horribly wrong…

Alberto Contador looked to his left, saw the group of leaders and then blitzed to the top.

As we watched him sprint on ahead, we wondered, will Lance respond, and the old tired body that crossed over the finish line a 1.5 minutes later told us the story, no.

Lance no longer the Champion, he was now the domestique of a new champion.

George Orwell, once again, is right

http://www.nytimes.com/2009/07/18/technology/companies/18amazon.html?_r=1

In George Orwell’s “1984,” government censors erase all traces of news articles embarrassing to Big Brother by sending them down an incineration chute called the “memory hole.”

On Friday, it was “1984” and another Orwell book, “Animal Farm,” that were dropped down the memory hole — by Amazon.com.

In a move that angered customers and generated waves of online pique, Amazon remotely deleted some digital editions of the books from the Kindle devices of readers who had bought them.

        […]

Of all of the books of all of the authors to demonstrate the power of the technology…

George must be laughing wherever he might be.

A google theory of mind

This is a fascinating read:

http://dashes.com/anil/2007/12/google-and-theory-of-mind.html

Especially the section on the transformational power of the PageRank/AdSense mechanism on the link mechanism of the web.

2. Connecting PageRank to economic systems such as AdWords and AdSense corrupted the meaning and value of links by turning them into an economic exchange. Through the turn of the millennium, hyperlinking on the web was a social, aesthetic, and expressive editorial action. When Google introduced its advertising systems at the same time as it began to dominate the economy around search on the web, it transformed a basic form of online communication, without the permission of the web’s users, and without explaining that choice or offering an option to those users.

Worse, the transformation was retroactive and the eventual mechanisms for opting out were incomplete in that the economic value could not be decoupled from the informational value. Inevitably, spammers arose to take advantage of the ability to create high-economic-value links at very low cost, causing vast damage to the ability to use links as a purely informational exchange. In addition, this forced Google to become more and more opaque about the refinements and adjustments it makes to its indexing algorithms, making a key part of their business less and less transparent over time. The eventual result has been the virtual decimation of communications systems like TrackBack, and absurdities like blogs linking to their own tag search results for key words in lieu of useful links, in an attempt to appease a search algorithm that they will never be allowed to fully understand.

An awareness of how a transformation in the fundamental value of links from informational to economic could have led Google to develop a system that separated editorial and aesthetic choices from economic ones, preventing the eventual link-spam arms race.

What the author observes is that linking which was a globally distributed process of creating an outline of the world’s information has been distorted by the economies generated by Google. In effect, the global outline was perverted to be almost useless.

The first thoughtful post-web article

In this article:

http://www.businessinsider.com/4-reasons-why-the-iphone-app-store-is-bad-news-for-google-2009-6

Time spent in apps is competing with (and replacing) time spent on the mobile Web.
There are some Google display ads in iPhone apps, but no direct line to

Google Search or search ads, where Google stands to make the most
money. (The good news is that Google doesn’t make much money yet from
the mobile Web, so this isn’t immediately disruptive.)

Time spent in apps is competing with (and replacing) time spent on the REAL Web.
This is probably the scariest scenario for Google. If I am sitting on
my couch playing iPhone games or reading an article in Instapaper Pro
— instead of goofing around on the Web on my laptop — that is
potentially real lost revenue for Google. (Similarly, even if I’m using
my iPhone’s Web browser, using the mobile Web and Google search, Google
is probably not monetizing those searches nearly as well as they do on
the computer.)

Users are learning to go to an app to find the information they need as opposed to going to Google or the Web.
For example, if I want a restaurant review, I don’t go to Google to
type in the name of the restaurant. I go to the Yelp app. Or
Urbanspoon. Or whatever. Either way, no Google there.

The App Store search engine built into every iPhone is becoming a very important search engine.
One billion apps downloaded means hundreds of millions (or billions) of
searches conducted. Google doesn’t power it or sell ads there.

Think client applications are back, baby.

The creepiest thing by far

The Wall Street Journal reported on the following very creepy business practice …

Suppose you have a bunch of executives who are owed deferred compensation like a pension.

Rather than keep money and have it count against your bottom line, use the money to buy out life insurance for your employees.

The life insurance premiums count as an investment so don’t hurt the bottom line.

The payouts, when your employees die allow you to pay out your executives.

Wolfram and Hart, the evil law firm from Angel,  bragged that you had to sign a contract for all time, who knew that this wasn’t fiction.

So employers not only made money from you in life, but also made money off of your death.

And yes, your relatives got none of this money.

And yes, even if you quit the company or got laid off, the firm still made money off the premiums.

Wow. Wow. Wow.

Star Trek

Very rarely does a film and a director re-invent a genre,  creating a new story that is so utterly consistent with the original story to leave you breathless with the audacity and the vision.

And I will contend, I have never seen a story teller re-invent the story in a way that is utterly and completely totally consistent with the original story.

JJ Abrams may, over time, deliver a crappy series, and the rest of his Star Trek movies may suck, but this movie will remain a magical moment in the history of story telling.

Am I Mac?

So after almost two years of irritating ads by Steve Jobs and Co, I finally bought myself a mac-mini. Well, my wife bought me a mac-mini.

And it’s kind of fun.

New software to learn and what not.

Not sure if I am mac, yet.

WAFL Performance: Making writes go faster with fewer IOPS

cross posting from my corporate blog

Like every storage array, Data ONTAP attempts to make the latency of the average read and write operation be lower than the actual latency of the underlying disk drive.

Unlike Real FiberChannel systems, how Data ONTAP does achieves this is quite unique.

The Latency Challenge

If you’re a storage array life looks like a set of random read and write operations that you have to immediately respond to. The temptation is to take a request and immediately go to do disk to get it as fast as you possibly can.

But that’s kind-a-stupid.

When you look at read operations you realize that they tend to be clustered around the same piece of data. So although an application may be asking for one 4k block of data, the reality is that the next piece of data it will ask for is probably sequentially next to the first 4k block of data, or is in fact the same 4k block of data!  Given that, it makes sense to do two things:

  1. Keep the 4k around in memory
  2. Get more than 4k when you go to disk so that the next request is serviced from memory and not disk.

In effect by using memory you can reduce what would have been three IOPS into one actual disk IO.

Write operations are a little bit more complicated. Obviously when you acknowledge a write you’re making a promise to the application that the data is committed to stable storage. Superficially this would imply that every write operation has to go to disk since that’s the only available non-volatile storage.

image

In this picture each box represents a distinct write operations. The letters designate the order in which they were received. If the storage array were to write to the data to disk as they arrived, then the storage array would perform 8 IOs. Furthermore, the latency of each IO would be the latency of the disk drive itself.

Enter battery backed memory…

Except it turns out that we can turn DRAM into non-volatile memory if we’re willing to keep it powered!

What that means is that once the data is in memory, and the storage array is sure that data will eventually make to disk, the write can be acknowledged. Now that the write operations are in memory it’s possible to schedule the write operations so that they are done sequentially.

image

In this picture the order in which the write operations are committed to disk is a->b->h->i->e->f->c->d->g even though the order that the application thinks they were committed to disk is a->b->c->d->e->f->g->h->i. Because the array re-orders the write operations, they can be written in one single IO, effectively a single sequential write operation.

The challenge is that you need to buffer enough write operations to make this efficient.

But WAFL, of course, is different

What I just described is how Real FiberChannel works. Effectively the location of the blocks (a,b,c, etc) are fixed on disk. All the storage array can do is re-order when you write them to their fixed locations.

What WAFL does is determine where the data has to go when the data is ready to be written. So using the same picture:

image

Rather than trying to put each block (a, b, c, d etc) in some pre-arranged location, WAFL finds the first bit of available space and writes the data sequentially. Like Real FiberChannel, Better than Real Fiber Channel will transform the random write operations into sequential operations reducing the total number of IOPS required.

Now what’s interesting is that WAFL doesn’t require that much memory to buffer the writes. And what’s even more interesting is because the fixed overhead to flush data to disk is negligible, there is no real value in holding onto the data in memory for very long.

Read and Write and IOPS

If you’ve been following this story, you’ll have figured out something very interesting: a write operation can be deferred but a read operation must go to disk. In other words, the array can choose when it goes to disk to commit a write, but if the data is not in memory, the array must go right then and there to get the data from disk!

So what makes a disk behave in a random fashion? Why the read operations, because the writes are always done sequentially!

So why do you need a lot of IOPS? Not to service the writes, because those are being written sequentially, but to service the read operations. The more read operations the more IOPS. In fact if you are doing truly sequential write operations, then you don’t need that many IOPS ….

But it’s read operations to DISK not read operations in general!

Aha! Now I get it…

The reason the PAM card is only a read cache is that adding more memory for writes doesn’t improve WAFL write performance… We’re writing as fast as we can write data sequentially to disk.

But adding the PAM card absorbs more read operations which reduces the number of IOPS that the storage system requires. And the best way to express reduced number of IOPS is requiring both fewer and slower disks!

If you can’t comment because of  an inability to get past the CAPTCHA, try a different browser. I’ve had success with IE. We’re having some problems right now and hopefully they will get fixed soon.