1*4882a593Smuzhiyun==================== 2*4882a593SmuzhiyunThe robust futex ABI 3*4882a593Smuzhiyun==================== 4*4882a593Smuzhiyun 5*4882a593Smuzhiyun:Author: Started by Paul Jackson <pj@sgi.com> 6*4882a593Smuzhiyun 7*4882a593Smuzhiyun 8*4882a593SmuzhiyunRobust_futexes provide a mechanism that is used in addition to normal 9*4882a593Smuzhiyunfutexes, for kernel assist of cleanup of held locks on task exit. 10*4882a593Smuzhiyun 11*4882a593SmuzhiyunThe interesting data as to what futexes a thread is holding is kept on a 12*4882a593Smuzhiyunlinked list in user space, where it can be updated efficiently as locks 13*4882a593Smuzhiyunare taken and dropped, without kernel intervention. The only additional 14*4882a593Smuzhiyunkernel intervention required for robust_futexes above and beyond what is 15*4882a593Smuzhiyunrequired for futexes is: 16*4882a593Smuzhiyun 17*4882a593Smuzhiyun 1) a one time call, per thread, to tell the kernel where its list of 18*4882a593Smuzhiyun held robust_futexes begins, and 19*4882a593Smuzhiyun 2) internal kernel code at exit, to handle any listed locks held 20*4882a593Smuzhiyun by the exiting thread. 21*4882a593Smuzhiyun 22*4882a593SmuzhiyunThe existing normal futexes already provide a "Fast Userspace Locking" 23*4882a593Smuzhiyunmechanism, which handles uncontested locking without needing a system 24*4882a593Smuzhiyuncall, and handles contested locking by maintaining a list of waiting 25*4882a593Smuzhiyunthreads in the kernel. Options on the sys_futex(2) system call support 26*4882a593Smuzhiyunwaiting on a particular futex, and waking up the next waiter on a 27*4882a593Smuzhiyunparticular futex. 28*4882a593Smuzhiyun 29*4882a593SmuzhiyunFor robust_futexes to work, the user code (typically in a library such 30*4882a593Smuzhiyunas glibc linked with the application) has to manage and place the 31*4882a593Smuzhiyunnecessary list elements exactly as the kernel expects them. If it fails 32*4882a593Smuzhiyunto do so, then improperly listed locks will not be cleaned up on exit, 33*4882a593Smuzhiyunprobably causing deadlock or other such failure of the other threads 34*4882a593Smuzhiyunwaiting on the same locks. 35*4882a593Smuzhiyun 36*4882a593SmuzhiyunA thread that anticipates possibly using robust_futexes should first 37*4882a593Smuzhiyunissue the system call:: 38*4882a593Smuzhiyun 39*4882a593Smuzhiyun asmlinkage long 40*4882a593Smuzhiyun sys_set_robust_list(struct robust_list_head __user *head, size_t len); 41*4882a593Smuzhiyun 42*4882a593SmuzhiyunThe pointer 'head' points to a structure in the threads address space 43*4882a593Smuzhiyunconsisting of three words. Each word is 32 bits on 32 bit arch's, or 64 44*4882a593Smuzhiyunbits on 64 bit arch's, and local byte order. Each thread should have 45*4882a593Smuzhiyunits own thread private 'head'. 46*4882a593Smuzhiyun 47*4882a593SmuzhiyunIf a thread is running in 32 bit compatibility mode on a 64 native arch 48*4882a593Smuzhiyunkernel, then it can actually have two such structures - one using 32 bit 49*4882a593Smuzhiyunwords for 32 bit compatibility mode, and one using 64 bit words for 64 50*4882a593Smuzhiyunbit native mode. The kernel, if it is a 64 bit kernel supporting 32 bit 51*4882a593Smuzhiyuncompatibility mode, will attempt to process both lists on each task 52*4882a593Smuzhiyunexit, if the corresponding sys_set_robust_list() call has been made to 53*4882a593Smuzhiyunsetup that list. 54*4882a593Smuzhiyun 55*4882a593Smuzhiyun The first word in the memory structure at 'head' contains a 56*4882a593Smuzhiyun pointer to a single linked list of 'lock entries', one per lock, 57*4882a593Smuzhiyun as described below. If the list is empty, the pointer will point 58*4882a593Smuzhiyun to itself, 'head'. The last 'lock entry' points back to the 'head'. 59*4882a593Smuzhiyun 60*4882a593Smuzhiyun The second word, called 'offset', specifies the offset from the 61*4882a593Smuzhiyun address of the associated 'lock entry', plus or minus, of what will 62*4882a593Smuzhiyun be called the 'lock word', from that 'lock entry'. The 'lock word' 63*4882a593Smuzhiyun is always a 32 bit word, unlike the other words above. The 'lock 64*4882a593Smuzhiyun word' holds 2 flag bits in the upper 2 bits, and the thread id (TID) 65*4882a593Smuzhiyun of the thread holding the lock in the bottom 30 bits. See further 66*4882a593Smuzhiyun below for a description of the flag bits. 67*4882a593Smuzhiyun 68*4882a593Smuzhiyun The third word, called 'list_op_pending', contains transient copy of 69*4882a593Smuzhiyun the address of the 'lock entry', during list insertion and removal, 70*4882a593Smuzhiyun and is needed to correctly resolve races should a thread exit while 71*4882a593Smuzhiyun in the middle of a locking or unlocking operation. 72*4882a593Smuzhiyun 73*4882a593SmuzhiyunEach 'lock entry' on the single linked list starting at 'head' consists 74*4882a593Smuzhiyunof just a single word, pointing to the next 'lock entry', or back to 75*4882a593Smuzhiyun'head' if there are no more entries. In addition, nearby to each 'lock 76*4882a593Smuzhiyunentry', at an offset from the 'lock entry' specified by the 'offset' 77*4882a593Smuzhiyunword, is one 'lock word'. 78*4882a593Smuzhiyun 79*4882a593SmuzhiyunThe 'lock word' is always 32 bits, and is intended to be the same 32 bit 80*4882a593Smuzhiyunlock variable used by the futex mechanism, in conjunction with 81*4882a593Smuzhiyunrobust_futexes. The kernel will only be able to wakeup the next thread 82*4882a593Smuzhiyunwaiting for a lock on a threads exit if that next thread used the futex 83*4882a593Smuzhiyunmechanism to register the address of that 'lock word' with the kernel. 84*4882a593Smuzhiyun 85*4882a593SmuzhiyunFor each futex lock currently held by a thread, if it wants this 86*4882a593Smuzhiyunrobust_futex support for exit cleanup of that lock, it should have one 87*4882a593Smuzhiyun'lock entry' on this list, with its associated 'lock word' at the 88*4882a593Smuzhiyunspecified 'offset'. Should a thread die while holding any such locks, 89*4882a593Smuzhiyunthe kernel will walk this list, mark any such locks with a bit 90*4882a593Smuzhiyunindicating their holder died, and wakeup the next thread waiting for 91*4882a593Smuzhiyunthat lock using the futex mechanism. 92*4882a593Smuzhiyun 93*4882a593SmuzhiyunWhen a thread has invoked the above system call to indicate it 94*4882a593Smuzhiyunanticipates using robust_futexes, the kernel stores the passed in 'head' 95*4882a593Smuzhiyunpointer for that task. The task may retrieve that value later on by 96*4882a593Smuzhiyunusing the system call:: 97*4882a593Smuzhiyun 98*4882a593Smuzhiyun asmlinkage long 99*4882a593Smuzhiyun sys_get_robust_list(int pid, struct robust_list_head __user **head_ptr, 100*4882a593Smuzhiyun size_t __user *len_ptr); 101*4882a593Smuzhiyun 102*4882a593SmuzhiyunIt is anticipated that threads will use robust_futexes embedded in 103*4882a593Smuzhiyunlarger, user level locking structures, one per lock. The kernel 104*4882a593Smuzhiyunrobust_futex mechanism doesn't care what else is in that structure, so 105*4882a593Smuzhiyunlong as the 'offset' to the 'lock word' is the same for all 106*4882a593Smuzhiyunrobust_futexes used by that thread. The thread should link those locks 107*4882a593Smuzhiyunit currently holds using the 'lock entry' pointers. It may also have 108*4882a593Smuzhiyunother links between the locks, such as the reverse side of a double 109*4882a593Smuzhiyunlinked list, but that doesn't matter to the kernel. 110*4882a593Smuzhiyun 111*4882a593SmuzhiyunBy keeping its locks linked this way, on a list starting with a 'head' 112*4882a593Smuzhiyunpointer known to the kernel, the kernel can provide to a thread the 113*4882a593Smuzhiyunessential service available for robust_futexes, which is to help clean 114*4882a593Smuzhiyunup locks held at the time of (a perhaps unexpectedly) exit. 115*4882a593Smuzhiyun 116*4882a593SmuzhiyunActual locking and unlocking, during normal operations, is handled 117*4882a593Smuzhiyunentirely by user level code in the contending threads, and by the 118*4882a593Smuzhiyunexisting futex mechanism to wait for, and wakeup, locks. The kernels 119*4882a593Smuzhiyunonly essential involvement in robust_futexes is to remember where the 120*4882a593Smuzhiyunlist 'head' is, and to walk the list on thread exit, handling locks 121*4882a593Smuzhiyunstill held by the departing thread, as described below. 122*4882a593Smuzhiyun 123*4882a593SmuzhiyunThere may exist thousands of futex lock structures in a threads shared 124*4882a593Smuzhiyunmemory, on various data structures, at a given point in time. Only those 125*4882a593Smuzhiyunlock structures for locks currently held by that thread should be on 126*4882a593Smuzhiyunthat thread's robust_futex linked lock list a given time. 127*4882a593Smuzhiyun 128*4882a593SmuzhiyunA given futex lock structure in a user shared memory region may be held 129*4882a593Smuzhiyunat different times by any of the threads with access to that region. The 130*4882a593Smuzhiyunthread currently holding such a lock, if any, is marked with the threads 131*4882a593SmuzhiyunTID in the lower 30 bits of the 'lock word'. 132*4882a593Smuzhiyun 133*4882a593SmuzhiyunWhen adding or removing a lock from its list of held locks, in order for 134*4882a593Smuzhiyunthe kernel to correctly handle lock cleanup regardless of when the task 135*4882a593Smuzhiyunexits (perhaps it gets an unexpected signal 9 in the middle of 136*4882a593Smuzhiyunmanipulating this list), the user code must observe the following 137*4882a593Smuzhiyunprotocol on 'lock entry' insertion and removal: 138*4882a593Smuzhiyun 139*4882a593SmuzhiyunOn insertion: 140*4882a593Smuzhiyun 141*4882a593Smuzhiyun 1) set the 'list_op_pending' word to the address of the 'lock entry' 142*4882a593Smuzhiyun to be inserted, 143*4882a593Smuzhiyun 2) acquire the futex lock, 144*4882a593Smuzhiyun 3) add the lock entry, with its thread id (TID) in the bottom 30 bits 145*4882a593Smuzhiyun of the 'lock word', to the linked list starting at 'head', and 146*4882a593Smuzhiyun 4) clear the 'list_op_pending' word. 147*4882a593Smuzhiyun 148*4882a593SmuzhiyunOn removal: 149*4882a593Smuzhiyun 150*4882a593Smuzhiyun 1) set the 'list_op_pending' word to the address of the 'lock entry' 151*4882a593Smuzhiyun to be removed, 152*4882a593Smuzhiyun 2) remove the lock entry for this lock from the 'head' list, 153*4882a593Smuzhiyun 3) release the futex lock, and 154*4882a593Smuzhiyun 4) clear the 'lock_op_pending' word. 155*4882a593Smuzhiyun 156*4882a593SmuzhiyunOn exit, the kernel will consider the address stored in 157*4882a593Smuzhiyun'list_op_pending' and the address of each 'lock word' found by walking 158*4882a593Smuzhiyunthe list starting at 'head'. For each such address, if the bottom 30 159*4882a593Smuzhiyunbits of the 'lock word' at offset 'offset' from that address equals the 160*4882a593Smuzhiyunexiting threads TID, then the kernel will do two things: 161*4882a593Smuzhiyun 162*4882a593Smuzhiyun 1) if bit 31 (0x80000000) is set in that word, then attempt a futex 163*4882a593Smuzhiyun wakeup on that address, which will waken the next thread that has 164*4882a593Smuzhiyun used to the futex mechanism to wait on that address, and 165*4882a593Smuzhiyun 2) atomically set bit 30 (0x40000000) in the 'lock word'. 166*4882a593Smuzhiyun 167*4882a593SmuzhiyunIn the above, bit 31 was set by futex waiters on that lock to indicate 168*4882a593Smuzhiyunthey were waiting, and bit 30 is set by the kernel to indicate that the 169*4882a593Smuzhiyunlock owner died holding the lock. 170*4882a593Smuzhiyun 171*4882a593SmuzhiyunThe kernel exit code will silently stop scanning the list further if at 172*4882a593Smuzhiyunany point: 173*4882a593Smuzhiyun 174*4882a593Smuzhiyun 1) the 'head' pointer or an subsequent linked list pointer 175*4882a593Smuzhiyun is not a valid address of a user space word 176*4882a593Smuzhiyun 2) the calculated location of the 'lock word' (address plus 177*4882a593Smuzhiyun 'offset') is not the valid address of a 32 bit user space 178*4882a593Smuzhiyun word 179*4882a593Smuzhiyun 3) if the list contains more than 1 million (subject to 180*4882a593Smuzhiyun future kernel configuration changes) elements. 181*4882a593Smuzhiyun 182*4882a593SmuzhiyunWhen the kernel sees a list entry whose 'lock word' doesn't have the 183*4882a593Smuzhiyuncurrent threads TID in the lower 30 bits, it does nothing with that 184*4882a593Smuzhiyunentry, and goes on to the next entry. 185