Sometimes some simple question leads you down deep to basic discussions: This was something I had to solve 2 years ago. The hard part was not finding it out, that was pretty obvious at start. The problem was to it was to explain it.

Since then I saw it reoccurring several times. Once occasion two weeks ago, thus I decided to write this down. I will just take the question of the customer 'I had a SAN and my tar -x was running fine with a short runtime. Now you gave me a ZFSSA, I'm using NFS and the times for tar -x are horrible. 61 Minutes ." I'm writing about this now, because in different appearances I've seen this problem again and again over the last few months and I think it's time to write an article where I can just point when I'm seeing this problem again.

To start with it: This issue is not Solaris specific ... this more or less a NFS specific problem. So maybe it's interesting to users of others operating systems as well.

The misconception that leads to this reoccurrence is the idea that NFS is simple and it's like a local file system only shared in the network. But it isn't that easy. A filesystem that offers network wide multiple reader/multiple writer access to the same filesystem with caches on all nodes and locking can't be simple. At least not as simple as a filesystem that works just on block devices used by one node.

Okay, back to the customer: I thought " NFS and tar. This is a very well-known thing. I think I hit a problem there somewhere at the very start of this century. I looked into the ZFSSA configuration. There wasn't a single write accelerating SSD in it. Okay "Dear Customer, buy a Writezilla and you will be fine. Best Regards Moellenkamp". But before spending a non-nearzero amount of money, the customer can demand a longer explanation. And the customer got this explanation. Even when the customer things that such a long runtime for tar -x may be okay, it causes some fears in regard of the performance. People want to know why something behaves differently than they expect so they can set their expectation right and understand why this may not hit them in production.

The first thing you have to keep in mind, is that tar is a purely sequential workload. It has to be this way. A tar has no centralized directory, where it stores the position of each file. You have to go through the file sequentially. It was designed that way out of very good reasons because tar stands for Tape ARchive. And tapes are notoriously bad at random reads compared to disks. When you plan an archiving tool for such a media, just designing for sequential access is acceptable. It's somewhat a strange thing that we use a combination of program that can only compress a single file and a tape archive program to ship software, but the computer industry is full of such strangenesses.

At first I had a look into the tar ball. 290.000 Files. Well … 61 minutes divided by 290.000 were 12.62 milliseconds per file. That not that bad … sometimes simple tasks just take longer, because you don't do them parallel and in large numbers. 290.000 times doing something short sequentially is still a lot time.

But why is it much faster on local storage. And why took it 12 milliseconds per file? The customer told me it was a matter of one or two minutes to unpack this on local disks. The answer is quite simple: A local filesystem like UFS or ZFS is not a shared filesystem. You have only one server writing into it. You have only one server reading from it. Writing to local disk is fairly easy. Your system is the only source of truth. You can safely cache things because when you change things you know you know it can be only you. You can combine the data in your cache and the disk to give a correct answer. On a networked filesystem this looks differently. A different node can change information and your cached content is stale. A different node doesn't have your cache to create the truth about the filesystem state. So things work pretty differently with NFS and you have to take this into equation when thinking about performance.

So what happens when you are locally do an untar:
0.0165  0.0002 openat64(4, "testfile1", O_RDWR|O_CREAT|O_TRUNC, 0100600) = 5
0.0169  0.0003 fstat64(5, 0xFD4E2AA0)                          = 0
 0.0170  0.0001 write(5, "\0\0\0\0\0\0\0\0\0\0\0\0".., 1024)    = 1024
0.0174  0.0000 close(5)			    = 0
I would like to point your attention to the flags of the open. With O_RDWR|O_CREAT|O_TRUNC. The open doesn't define any special things in regard of the synchronized I/O. And of course those will be fast. A tar extract - so to say - is a sequential run of many of this open,write,closes with occasional creations of directories. Like the one depicted in the truss. The writes out of this won't hit the disk before the ZFS transaction group is committed to stable storage (simplified, the reality is a lot or more complex).

Okay, but local filesystems are not the problem here. When you unbar the same file into a directory provided by NFS, you will see the following output from a packet sniffer. I would like to add that no special mount options where used that could have some impact how things are working internally. This was made with a plain standard mount.

  1 10.0.2.11 -> 10.0.2.10    NFS 194 V3 LOOKUP Call, DH: 0xad487762/singlefile
  2 10.0.2.10 -> 10.0.2.11    NFS 186 V3 LOOKUP Reply (Call In 1) Error: NFS3ERR_NOENT
  3 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0xad487762
  4 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 3)  Directory mode: 0777 uid: 0 gid: 0
  5 10.0.2.11 -> 10.0.2.10    NFS 226 V3 MKDIR Call, DH: 0xad487762/singlefile
  6 10.0.2.10 -> 10.0.2.11    NFS 342 V3 MKDIR Reply (Call In 5)
  7 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0xad487762
  8 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 7)  Directory mode: 0777 uid: 0 gid: 0
  9 10.0.2.11 -> 10.0.2.10    NFS 182 V3 ACCESS Call, FH: 0x44856022, [Check: RD LU MD XT DL]
 10 10.0.2.10 -> 10.0.2.11    NFS 190 V3 ACCESS Reply (Call In 9), [Allowed: RD LU MD XT DL]
 11 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0x44856022
 12 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 11)  Directory mode: 0755 uid: 100 gid: 10
 13 10.0.2.11 -> 10.0.2.10    NFS 194 V3 LOOKUP Call, DH: 0x44856022/testfile1
 14 10.0.2.10 -> 10.0.2.11    NFS 186 V3 LOOKUP Reply (Call In 13) Error: NFS3ERR_NOENT
 15 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0x44856022
 16 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 15)  Directory mode: 0755 uid: 100 gid: 10
 17 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0x44856022
 18 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 17)  Directory mode: 0755 uid: 100 gid: 10
 19 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0x44856022
 20 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 19)  Directory mode: 0755 uid: 100 gid: 10
 21 10.0.2.11 -> 10.0.2.10    NFSACL 182 V3 GETACL Call
 22 10.0.2.10 -> 10.0.2.11    NFSACL 206 V3 GETACL Reply (Call In 
21)
 23 10.0.2.11 -> 10.0.2.10    NFS 194 V3 LOOKUP Call, DH: 0x44856022/testfile1
 24 10.0.2.10 -> 10.0.2.11    NFS 186 V3 LOOKUP Reply (Call In 23) Error: NFS3ERR_NOENT
 25 10.0.2.11 -> 10.0.2.10    NFS 238 V3 CREATE Call, DH: 0x44856022/testfile1 Mode: GUARDED
 26 10.0.2.10 -> 10.0.2.11    NFS 342 V3 CREATE Reply (Call In 25)
 27 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0xd98a8154
 28 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 27)  Regular File mode: 0600 uid: 100 gid: 10
 29 10.0.2.11 -> 10.0.2.10    NFS 1222 V3 WRITE Call, FH: 0xd98a8154 Offset: 0 Len: 1024 FILE_SYNC
 30 10.0.2.10 -> 10.0.2.11    NFS 230 V3 WRITE Reply (Call In 29) Len: 1024 FILE_SYNC
 31 10.0.2.11 -> 10.0.2.10    NFS 178 V3 GETATTR Call, FH: 0xd98a8154
 32 10.0.2.10 -> 10.0.2.11    NFS 182 V3 GETATTR Reply (Call In 31)  Regular File mode: 0600 uid: 100 gid: 10
 33 10.0.2.11 -> 10.0.2.10    NFS 222 V3 SETATTR Call, FH: 0xd98a8154
 34 10.0.2.10 -> 10.0.2.11    NFS 214 V3 SETATTR Reply (Call In 33)
 35 10.0.2.11 -> 10.0.2.10    NFS 222 V3 SETATTR Call, FH: 0x44856022
 36 10.0.2.10 -> 10.0.2.11    NFS 214 V3 SETATTR Reply (Call In 35)
As you see, a lot is going on. You may say, the performance difference between a local tar and a tar on a NFS filesystem is the network. Of course there is some latency in the story when you have to communicate that often over the network.

And so you are for sure partially right, but that is only half the story and it doesn't explain the 12ms (which are interestingly not that far away from the worst case rotational latency of a 5400 rpm disk). The rest of the story is that the load is transformed by using NFS.

There is an important passage in RFC 1813. It states:
Data-modifying operations in the NFS version 3 protocol are synchronous. When a procedure returns to the client, the client can assume that the operation has completed and any data associated with the request is now on stable storage.
Furthermore it elaborates:
The following data modifying procedures are synchronous: WRITE (with stable flag set to FILE_SYNC), CREATE,MKDIR, SYMLINK, MKNOD, REMOVE, RMDIR, RENAME, LINK, and COMMIT.

Not in this list is the SETATTR command, which forces the data as well to stable storage before completing at least to my knowledge and what's visible in the copies of nfs_srv.c out there in the internet.

When you are looking onto the tcpdump you see a significant number of the mentioned operations. Now imagine the same for a tar file of 290000 small files.


So essentially using NFS transform the load formerly consisting out of a multitude asynchronous write operations into one executing a significant number of operations that are synchronous by definition. What was a async write on application layer, is a sequence of synchronous writes when it hits the NFS server and thus hits the filesystem underneath. A rather total transformation of the load.

However, the situation is not that bad in reality. There are some optimizations. Let's have a look at a different tcpdump. I've created a tar file with files with four lengths in it: 1k, 32k, 33k and 64k. The tcpdumps are reduces to CREATE, WRITE and COMMIT calls.

We start with the 32k file. You see it's written as UNSTABLE. So the call is allowed to return without persisting the data to disk before. But at close there is a COMMIT forcing NFS to persist everything.
25 10.0.2.11 -> 10.0.2.10    NFS 234 V3 CREATE Call, DH: 0x24bc098f/32kfile Mode: GUARDED
 53 10.0.2.11 -> 10.0.2.10    NFS 846 V3 WRITE Call, FH: 0x58dd2c54 Offset: 0 Len: 32768 UNSTABLE
 65  10.0.2.11 -> 10.0.2.10    NFS 190 V3 COMMIT Call, FH: 0x58dd2c54

Let's have a look onto the 1k file.
 79 10.0.2.11 -> 10.0.2.10    NFS 234 V3 CREATE Call, DH: 0x24bc098f/1kfile Mode: GUARDED
 83 10.0.2.11 -> 10.0.2.10    NFS 1222 V3 WRITE Call, FH: 0xc5d2cd22 Offset: 0 Len: 1024 FILE_SYNC
The write is directly send as a FILE_SYNC write. But why? It's pretty obvious, when you look at what's missing. There is no COMMIT. At the end you save a command you don't have to send. One round-trip time less.

For the interest of curiosity, I look at the 33k size and the 64k sizes. At first 33k:
101  10.0.2.11 -> 10.0.2.10    NFS 234 V3 CREATE Call, DH: 0x24bc098f/33kfile Mode: GUARDED
131 10.0.2.11 -> 10.0.2.10    NFS 846 V3 WRITE Call, FH: 0x38903d60 Offset: 0 Len: 32768 UNSTABLE
135 10.0.2.11 -> 10.0.2.10    NFS 1222 V3 WRITE Call, FH: 0x38903d60 Offset: 32768 Len: 1024 FILE_SYNC
142 10.0.2.11 -> 10.0.2.10    NFS 190 V3 COMMIT Call, FH: 0x38903d60
Now at 64k :
156 10.0.2.11 -> 10.0.2.10    NFS 234 V3 CREATE Call, DH: 0x24bc098f/64kfile Mode: GUARDED
187 10.0.2.11 -> 10.0.2.10    NFS 846 V3 WRITE Call, FH: 0xa59fdc16 Offset: 0 Len: 32768 UNSTABLE
214 10.0.2.11 -> 10.0.2.10    NFS 846 V3 WRITE Call, FH: 0xa59fdc16 Offset: 32768 Len: 32768 UNSTABLE
227 10.0.2.11 -> 10.0.2.10    NFS 190 V3 COMMIT Call, FH: 0xa59fdc16
I'm not an expert for the NFS code, but my assumption about the assumption behind this behavior that it is expected that there will be more writes into the same file if you fill up the wsize. If you don't fill it, it's expected that it may be the only or the last write.

So it's not as easy as saying that NFS is always writing synchronously. It was the case in NFSv2 and was a source of a lot of NFS performance problems. NFSv3 introduced the concept of asynchronous NFS writes with a synchronous COMMIT at the end. But the "many small files"-tar counters this as you may have taken away from the tcpdumps. So: If you write just 290000 files mostly smaller than 32k for example (the configuration at that customer) you will see a lot of NFSv3 sync writes and almost no asynchronous writes and almost no commits. If you had one of few big files in the tar with the same total size you would write the amount of data much faster, because then you would have a lot asynchronous NFS writes and final synchronous NFS write with a closing NFS COMMIT.

I was able to show by this to the customer that the execution time of the tar -x and the difference between the local filesystem and a remote filesystem via NFS is the expected behavior given the configuration of their storage pool.

How do you improve this situation? The problem is there for quite a time. And there were a multitude of solutions to it.

Just to give you a perspective how long this problem exists: Perhaps some of you are old enough to remember things like the Prestoserv NFS accelerator, which was quite useful in NFSv2 as each and every WRITE Call had to be answered synchronously (which didn't know the method of a sequence of ASYNC write with a following COMMIT, obviously ... COMMIT was introduced in NFSv3). It solved this problem for example.

Basically the answer to the problem is always: Making write latency as fast as possible. And so it boils down to get rid of the need to hit rotating rust before being able to answer on the mentioned NFS calls. One solution for example used on ZFS Storage Appliances are SSD for the log devices. A write is considered as non-volatile as soon as it's on the log device.

Out of this reasons when working with customers I have two rules: 1. You don't get a ZFS Storage Appliance from me without having write-accelerating SSD as long as the you and me can't prove that your load won't be hit by the mentioned necessities of the protocol (which is quite hard, when it hides even in such a simple thing like at tar) or you tell me, that you understand the problem but still want it otherwise. At the end you are the customer and customer is king as we say in Germany. 2. When you tell me you run NFS from a server (let's say DIY ZFS Storage Appliance) you should either have a storage array behind with a non-volatile write cache (that should be really used and not switched off, just saying out of experience) or you should opt for log devices as well. Without it you have essentially a storage array with cache but with a wrecked up battery which goes into write through mode because of this and you get the performance you can expect out of such a setup.

And as a fun fact: You may think that tar -x is not that problematic because you do it only from time to time. But on the contrary: I've seen quite a number of customers in the past, where tar is production critical. For example unpacking sensor data, development file servers where quite a number of tar unpacks ran all day, diagnostic data of machines, something like that. This small little tape archive tool is used quite frequently for a multitude of tasks where you don't expect them.

What the conclusion out of this:
  • Well, things are most of the time not as easy as they seem.
  • Sometimes the simple loads are the most nasty and pathological ones.
  • It pays to read the RFCs

  • Disclaimer: The individual owning this blog works for Oracle in Germany. The opinions expressed here are his own, are not necessarily reviewed in advance by anyone but the individual author, and neither Oracle nor any other party necessarily agrees with them.

No comments

Add Comment

Enclosing asterisks marks text as bold (*word*), underscore are made via _word_.
Standard emoticons like :-) and ;-) are converted to images.

To prevent automated Bots from commentspamming, please enter the string you see in the image below in the appropriate input box. Your comment will only be submitted if the strings match. Please ensure that your browser supports and accepts cookies, or your comment cannot be verified correctly.
CAPTCHA