Blob Blame History Raw
diff -ur nfsometer-1.9.orig/nfsometerlib/cmd.py nfsometer-1.9.wip/nfsometerlib/cmd.py
--- nfsometer-1.9.orig/nfsometerlib/cmd.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/cmd.py	2019-05-13 16:26:40.269894942 +0100
@@ -12,9 +12,9 @@
 """
 
 import os
-import posix
 import sys
 import subprocess
+import six
 
 # command wrappers
 def simplecmd(args):
@@ -41,11 +41,11 @@
 def cmd(args, raiseerrorcode=True, raiseerrorout=True, instr='',
         env=None, pass_output=False):
 
-    #print "command> %s" % args
+    #print("command> %s" % args)
 
     if env:
-        curenv = dict(posix.environ)
-        for k,v in env.iteritems():
+        curenv = dict(os.environ)
+        for k,v in six.iteritems(env):
             curenv[k] = v
         env = curenv
 
@@ -82,11 +82,13 @@
                             (args, errstr))
 
     if outstr:
-        o_str = outstr.split('\n')
+        o_str = outstr.decode('utf-8').split('\n')
     else:
         o_str = ''
 
     if errstr:
+        if isinstance(errstr, six.binary_type):
+          errstr = errstr.decode('utf-8')
         e_str = errstr.split('\n')
     else:
         e_str = ''
diff -ur nfsometer-1.9.orig/nfsometerlib/collection.py nfsometer-1.9.wip/nfsometerlib/collection.py
--- nfsometer-1.9.orig/nfsometerlib/collection.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/collection.py	2019-05-20 13:39:47.967197141 +0100
@@ -14,11 +14,12 @@
 import os
 import numpy as np
 from subprocess import call
+import six
 
-from config import *
-from selector import Selector
-import parse
-from trace import TraceAttrs
+from .config import *
+from .selector import Selector
+from . import parse
+from .trace import TraceAttrs
 
 class Stat:
     """
@@ -64,7 +65,7 @@
         return "Stat(name=%r, values=%r, tracedirs=%r)" % \
                     (self.name, self._values, self._tracedirs)
 
-    def __nonzero__(self):
+    def __bool__(self):
         return not self.empty()
 
     def num_runs(self):
@@ -75,7 +76,7 @@
         """ return the value for the run associated with tracedir """
         try:
             run = self._tracedirs.index(tracedir)
-        except ValueError, e:
+        except ValueError as e:
             if args:
                 assert len(args) == 1
                 return args[0]
@@ -83,7 +84,7 @@
 
         try:
             return self._values[run]
-        except IndexError, e:
+        except IndexError as e:
             if args:
                 assert len(args) == 1
                 return args[0]
@@ -161,12 +162,12 @@
         self._empty = None
         self._num_runs = None
 
-    def __nonzero__(self):
+    def __bool__(self):
         return not self.empty()
 
     def _sort(self):
         if not self._sorted:
-            self._stats.sort(lambda x,y: -1 * cmp(x.mean(), y.mean()))
+            self._stats.sort(key=lambda x: x.mean(), reverse=True)
             self._sorted = True
 
     def foreach(self):
@@ -182,12 +183,12 @@
 
     def mean(self):
         if self._mean == None:
-            self._mean = np.mean(self._sum_by_tracedir.values())
+            self._mean = np.mean(list(self._sum_by_tracedir.values()))
         return self._mean
 
     def std(self):
         if self._std == None:
-            self._std = np.std(self._sum_by_tracedir.values())
+            self._std = np.std(list(self._sum_by_tracedir.values()))
         return self._std
 
     def max(self):
@@ -223,7 +224,7 @@
             if not d in self._tracedirs:
                 self._tracedirs.append(d)
 
-            if not self._sum_by_tracedir.has_key(d):
+            if d not in self._sum_by_tracedir:
                 self._sum_by_tracedir[d] = 0.0
             self._sum_by_tracedir[d] += vals[i]
 
@@ -240,15 +241,16 @@
         self._num_runs = None
 
     def add_attr(self, name, value):
-        if not self._attrs.has_key(name):
-            self._attrs[name] = set()
-        self._attrs[name].add(value)
+        if name not in self._attrs:
+            self._attrs[name] = set([value])
+        else:
+            self._attrs[name].add(value)
 
     def get_attr(self, name):
         return self._attrs[name]
 
     def has_attr(self, name):
-        return self._attrs.has_key(name)
+        return name in self._attrs
 
     def merge_attrs(self, new):
         str_attrs = ['workload_command', 'workload_description']
@@ -264,7 +266,7 @@
         """ add a value for the key.  should be called once on each key for
             every workload result directory """
 
-        if not self._values.has_key(key):
+        if key not in self._values:
             self._values[key] = Stat(key)
 
         self._values[key].add_value(float(value), filename, tracedir)
@@ -293,7 +295,7 @@
             every workload result directory """
         assert isinstance(stat, Stat), repr(stat)
 
-        if not self._values.has_key(bucket_name):
+        if bucket_name not in self._values:
             self._values[bucket_name] = Bucket(bucket_name)
 
         self._values[bucket_name].add_stat_to_bucket(stat)
@@ -337,7 +339,7 @@
                 # new
                 elif ent.startswith(TRACE_DIR_PREFIX) and os.path.isdir(ent):
                     self.load_tracedir(ent)
-            except IOError, e:
+            except IOError as e:
                 self.warn(ent, str(e))
 
         os.chdir(cwd)
@@ -351,7 +353,7 @@
         servers = set()
         paths = set()
 
-        for sel, tracestat in self._tracestats.iteritems():
+        for sel, tracestat in six.iteritems(self._tracestats):
             parse.gather_buckets(self, tracestat)
 
             workloads.add(sel.workload)
@@ -391,7 +393,7 @@
     def notes_get(self):
         notes_file = os.path.join(self.resultsdir, NOTES_FILE)
         try:
-            return file(notes_file).readlines()
+            return open(notes_file).readlines()
         except IOError:
             return []
 
@@ -401,19 +403,19 @@
         if msg.startswith('[Errno '):
             msg = msg[msg.find(']') + 1:]
 
-        if not self._warnings.has_key(tracedir):
+        if tracedir not in self._warnings:
             self._warnings[tracedir] = []
         self._warnings[tracedir].append(msg.replace(tracedir, '[dir]/'))
         warn(tracedir + ': ' + msg)
 
     def warnings(self):
-        return [ (d, tuple(self._warnings[d])) for d in self._warnings.keys() ]
+        return [ (d, tuple(self._warnings[d])) for d in list(self._warnings.keys()) ]
 
     def empty(self):
         return len(self._tracestats) == 0
 
     def set_stat_info(self, key, info):
-        if not self._stat_info.has_key(key):
+        if key not in self._stat_info:
             self._stat_info[key] = info
         else:
             assert self._stat_info[key] == info, \
@@ -450,7 +452,7 @@
 
         assert sel.is_valid_key(), "Invalid key: %r" % sel
 
-        if not self._tracestats.has_key(sel):
+        if sel not in self._tracestats:
             self._tracestats[sel] = TraceStats(self)
 
         return self._tracestats[sel]
@@ -462,7 +464,7 @@
         """ return True if this collection has any traces matching 'selection',
             otherwise returns False """
         for x in selection.foreach():
-            if self._tracestats.has_key(x):
+            if x in self._tracestats:
                 return True
         return False
 
@@ -474,7 +476,7 @@
         attr_file = os.path.join(tracedir, 'arguments')
         trace_attrs = TraceAttrs(filename=attr_file).to_dict()
 
-        for k, v in trace_attrs.iteritems():
+        for k, v in six.iteritems(trace_attrs):
             attr[k] = v
 
         return attr
@@ -485,7 +487,7 @@
             returns empty string if nothing is found
         """
         def _check_lines(f):
-            return '\n'.join([ x[2:] for x in file(f).readlines()
+            return '\n'.join([ x[2:] for x in open(f).readlines()
                 if x.startswith('>') and x.lower().find('nfs:') >= 0 ])
 
         diff = os.path.join(tracedir, 'dmesg.diff')
@@ -528,14 +530,18 @@
         for subsel in selection.foreach():
             try:
                 tracestat = self.get_trace(subsel)
-            except KeyError:
+            except KeyError as e:
                 continue
 
             if tracestat.has_attr(attr_name):
                 trace_attr = tracestat.get_attr(attr_name)
                 attr = attr.union(trace_attr)
 
-        attr = list(attr)
+        if not attr:
+            attr = []
+        else:
+            attr = list(attr)
+
         attr.sort()
 
         return tuple(attr)
@@ -568,9 +574,9 @@
                 if not mdt in map_order:
                     map_order.append(mdt)
 
-                if not tmpmap.has_key(mdt):
+                if mdt not in tmpmap:
                     tmpmap[mdt] = {}
-                if not tmpmap[mdt].has_key(nruns):
+                if nruns not in tmpmap[mdt]:
                     tmpmap[mdt][nruns] = []
 
                 tmpmap[mdt][nruns].append(subsel.workload)
@@ -578,10 +584,10 @@
             wmap = {}
             worder = []
             for mdt in map_order:
-                if not tmpmap.has_key(mdt):
+                if mdt not in tmpmap:
                     continue
 
-                runs = tmpmap[mdt].keys()
+                runs = list(tmpmap[mdt].keys())
                 runs.sort()
 
                 for r in runs:
@@ -633,7 +639,7 @@
         order = ['workload', 'client', 'server', 'mountopt', 'detect', 'tag', 'kernel', 'path']
 
         for subsel in selection.foreach(order):
-            assert not vals.has_key(subsel)
+            assert subsel not in vals
             vals[subsel] = {}
 
             try:
diff -ur nfsometer-1.9.orig/nfsometerlib/config.py nfsometer-1.9.wip/nfsometerlib/config.py
--- nfsometer-1.9.orig/nfsometerlib/config.py	2017-06-29 04:41:45.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/config.py	2019-05-13 16:26:40.270894947 +0100
@@ -12,7 +12,7 @@
 """
 
 import re
-import os, posix, stat, sys
+import os, stat, sys
 import socket
 
 NFSOMETER_VERSION='1.9'
@@ -20,7 +20,7 @@
 NFSOMETER_MANPAGE='nfsometer.1'
 
 NFSOMETERLIB_DIR=os.path.split(__file__)[0]
-NFSOMETER_DIR=os.path.join(posix.environ['HOME'], '.nfsometer')
+NFSOMETER_DIR=os.path.join(os.environ['HOME'], '.nfsometer')
 
 
 #
@@ -47,7 +47,7 @@
 #
 MOUNTDIR=os.path.join(RUNNING_TRACE_DIR, 'mnt')
 WORKLOADFILES_ROOT=os.path.join(NFSOMETER_DIR, 'workload_files')
-RESULTS_DIR=os.path.join(posix.environ['HOME'], 'nfsometer_results')
+RESULTS_DIR=os.path.join(os.environ['HOME'], 'nfsometer_results')
 HOSTNAME=socket.getfqdn()
 RUNROOT='%s/nfsometer_runroot_%s' % (MOUNTDIR, HOSTNAME)
 HTML_DIR="%s/html" % NFSOMETERLIB_DIR
@@ -139,7 +139,7 @@
 _TEMPLATE_CACHE={}
 def html_template(filename):
     global _TEMPLATE_CACHE
-    if not _TEMPLATE_CACHE.has_key(filename):
+    if filename not in _TEMPLATE_CACHE:
         _TEMPLATE_CACHE[filename] = Template(filename=filename)
     return _TEMPLATE_CACHE[filename]
 
@@ -244,7 +244,7 @@
     gmap = {}
     for g in groups:
         vers = mountopts_version(g.mountopt)
-        if not gmap.has_key(vers):
+        if vers not in gmap:
             gmap[vers] = []
         gmap[vers].append(g)
     return gmap
diff -ur nfsometer-1.9.orig/nfsometerlib/graph.py nfsometer-1.9.wip/nfsometerlib/graph.py
--- nfsometer-1.9.orig/nfsometerlib/graph.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/graph.py	2019-05-21 10:06:41.797380624 +0100
@@ -11,15 +11,15 @@
 FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
 """
 
-#!/usr/bin/env python
+#!/usr/bin/env python3
 
 import multiprocessing
-import cPickle
-import os, sys, time
+from six.moves import cPickle
+import os, sys, time, errno
 
-from collection import *
-from config import *
-import selector
+from .collection import *
+from .config import *
+from . import selector
 
 _GRAPH_COLLECTION = None
 
@@ -52,8 +52,8 @@
 
         try:
             os.mkdir(self.imagedir)
-        except OSError, e:
-            assert e.errno == os.errno.EEXIST
+        except OSError as e:
+            assert e.errno == errno.EEXIST
 
         self._entries = set(os.listdir(imagedir))
 
@@ -104,8 +104,8 @@
         if classes:
            other_attrs.append('class="%s"' % ' '.join(classes))
 
-        if attrs.has_key('groups'):
-            if not attrs.has_key('gmap'):
+        if 'groups' in attrs:
+            if 'gmap' not in attrs:
                 attrs['gmap'] = groups_by_nfsvers(attrs['groups'])
             gmap = attrs['gmap']
 
@@ -116,7 +116,7 @@
             cur = 0
             sub_src = []
             for vers in NFS_VERSIONS:
-                if not gmap.has_key(vers):
+                if vers not in gmap:
                     continue
 
                 assert cur < num
@@ -288,26 +288,26 @@
 
         inform('\rGraph Summary:                                  ')
         if self.gen_count:
-            print '  %u images generated' % self.gen_count
+            print('  %u images generated' % self.gen_count)
         if self.cached_count:
-            print '  %u cached images' % self.cached_count
+            print('  %u cached images' % self.cached_count)
         if self.prune_count:
-            print '  %u files pruned' % self.prune_count
+            print('  %u files pruned' % self.prune_count)
 
 def _fmt_data(x, scale):
     assert not isinstance(x, (list, tuple))
-    if isinstance(x, Stat):
+    if isinstance(x, Stat) or type(x).__name__ == 'Stat':
         return x.mean() / scale, x.std() / scale
 
     # disallow?
-    elif isinstance(x, (float, int, long)):
+    elif isinstance(x, (float,)+six.integer_types):
         return x, 0.0
 
     elif x == None:
         # when graphing, no data can just be zero
         return 0.0, 0.0
 
-    raise ValueError('Unexpected data type for %r' % (val,))
+    raise ValueError('Unexpected data type for %r' % (x,))
 
 def _graphize_units(units):
     if not units:
@@ -321,7 +321,7 @@
         graph_f(imgfile, attrs)
     except KeyboardInterrupt:
         return False
-    except Exception, e:
+    except Exception as e:
         return e
     return True
 
@@ -365,7 +365,7 @@
     ax1 = fig.add_subplot(111)
     ax1.set_autoscale_on(True)
     ax1.autoscale_view(True,True,True)
-    for i in ax1.spines.itervalues():
+    for i in six.itervalues(ax1.spines):
         i.set_linewidth(0.0)
 
     # width of bars within a group
@@ -409,14 +409,14 @@
 
         val = vals[g].get(key, None)
         hidx = 0 # default hatch
-        if isinstance(val, Bucket):
+        if isinstance(val, Bucket) or type(val).__name__ == 'Bucket':
             for s in val.foreach():
                 x_v, x_s = _fmt_data(s, scale)
                 hidx = hatch_map[s.name]
 
-                assert not valmap[key].has_key(hidx), \
+                assert hidx not in valmap[key], \
                     '%u, %r' % (hidx, val)
-                assert not errmap[key].has_key(hidx), \
+                assert hidx not in errmap[key], \
                     '%u, %r' % (hidx, val)
                 valmap[key][hidx] = x_v
                 errmap[key][hidx] = x_s
@@ -496,7 +496,7 @@
     ax1 = fig.add_subplot(111)
     ax1.set_autoscale_on(True)
     ax1.autoscale_view(True,True,True)
-    for i in ax1.spines.itervalues():
+    for i in six.itervalues(ax1.spines):
         i.set_linewidth(0.0)
 
     ax1.get_xaxis().set_visible(False)
diff -ur nfsometer-1.9.orig/nfsometerlib/options.py nfsometer-1.9.wip/nfsometerlib/options.py
--- nfsometer-1.9.orig/nfsometerlib/options.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/options.py	2019-05-17 12:58:31.852257799 +0100
@@ -14,8 +14,9 @@
 import os, posix, sys
 import getopt
 import re
+import six
 
-from config import *
+from .config import *
 
 _progname = sys.argv[0]
 
@@ -356,7 +357,7 @@
         try:
             opts, args = getopt.getopt(sys.argv[1:], shortstr, longlist)
 
-        except getopt.GetoptError, err:
+        except getopt.GetoptError as err:
             self.usage(str(err))
 
         # parse options
@@ -464,7 +465,7 @@
         for x in mountopts:
             try:
                 vers = mountopts_version(x)
-            except ValueError, e:
+            except ValueError as e:
                 self.usage(str(e))
             self.mountopts.append(x)
 
@@ -490,7 +491,7 @@
             err = False
             for name in ('NFSOMETER_CMD', 'NFSOMETER_NAME', 'NFSOMETER_DESC',):
                 if not name in posix.environ:
-                    print >>sys.stderr, "%s not set" % name
+                    six.print_("%s not set" % name, file=sys.stderr)
                     err = True
 
             if err:
@@ -564,10 +565,9 @@
         return lines
 
     def error(self, msg=''):
-        print >>sys.stderr, msg
-        print >>sys.stderr, \
-            '\nrun "%s --help" and "%s examples" for more info' % \
-            (_progname, _progname)
+        six.print_(msg, file=sys.stderr)
+        six.print_('\nrun "%s --help" and "%s examples" for more info' % \
+            (_progname, _progname), file=sys.stderr)
         sys.stderr.flush()
         sys.exit(1)
 
@@ -593,19 +593,19 @@
         return self._synopsis_fmt % script
 
     def examples(self):
-        print >>sys.stdout, self._examples()
+        print(self._examples())
 
     def usage(self, msg=''):
-        print >>sys.stderr, "usage: %s" % self._synopsis(_progname)
-        print >>sys.stderr, self._modes_description(_progname)
+        six.print_("usage: %s" % self._synopsis(_progname), file=sys.stderr)
+        six.print_(self._modes_description(_progname), file=sys.stderr)
 
-        print >>sys.stderr
-        print >>sys.stderr, "Options:"
-        print >>sys.stderr, '  %s' % '\n  '.join(self._option_help())
+        six.print_("", file=sys.stderr)
+        six.print_("Options:", file=sys.stderr)
+        six.print_('  %s' % '\n  '.join(self._option_help()), file=sys.stderr)
 
         if msg:
-            print >>sys.stderr
-            print >>sys.stderr, "Error: " + msg
+            six.print_('', file=sys.stderr)
+            six.print_("Error: " + msg, file=sys.stderr);
 
         sys.exit(1)
 
@@ -635,5 +635,6 @@
         for i in range(len(o)):
             o[i] = o[i].strip().replace('-', '\\-')
 
-        file(output_path, 'w+').write('\n'.join(o))
+        with open(output_path, 'w+') as f:
+          f.write('\n'.join(o))
 
diff -ur nfsometer-1.9.orig/nfsometerlib/parse.py nfsometer-1.9.wip/nfsometerlib/parse.py
--- nfsometer-1.9.orig/nfsometerlib/parse.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/parse.py	2019-05-20 13:40:30.733352852 +0100
@@ -13,8 +13,9 @@
 
 import os
 import re
+import six
 
-from config import *
+from .config import *
 
 #
 # Regular Expressions section
@@ -113,7 +114,7 @@
         return r
 
     def add_key(self, bucket_name, key, display):
-        if self._key2bucket.has_key(key) or key in self._other_keys:
+        if key in self._key2bucket or key in self._other_keys:
             return
 
         if display:
@@ -189,7 +190,7 @@
     ),
 }
 nfsstat_op_map = {}
-for b, ops in nfsstat_op_map_def.iteritems():
+for b, ops in six.iteritems(nfsstat_op_map_def):
     for o in ops:
         nfsstat_op_map[o] = b
 
@@ -245,7 +246,7 @@
 }
 
 mountstat_op_map = {}
-for b, ops in mountstat_op_map_def.iteritems():
+for b, ops in six.iteritems(mountstat_op_map_def):
     for o in ops:
         mountstat_op_map[o] = b
 
@@ -299,7 +300,7 @@
 
         try:
             p(tracestat, tracedir, attrs)
-        except Exception, e:
+        except Exception as e:
             collection.warn(tracedir, str(e))
 
 
@@ -310,7 +311,7 @@
 
     path = os.path.join(tracedir, filename)
 
-    lines = [ x.strip() for x in file(path) if x.strip() ]
+    lines = [ x.strip() for x in open(path) if x.strip() ]
     assert len(lines) == 3
 
     def _parse_time(minutes, seconds):
@@ -376,8 +377,7 @@
 
     path = os.path.join(tracedir, filename)
 
-    f = file(path)
-
+    f = open(path)
     for line in f:
         found = False
 
@@ -388,7 +388,7 @@
 
         m = RE['ms_read_norm'].match(line)
         if m:
-            val = long(m.group(1))
+            val = int(m.group(1))
             tracestat.add_stat(prefix + 'read_normal',
                         val, 'B',
                         'Bytes read through the read() syscall',
@@ -400,7 +400,7 @@
 
         m = RE['ms_write_norm'].match(line)
         if m:
-            val = long(m.group(1))
+            val = int(m.group(1))
             tracestat.add_stat(prefix + 'write_normal',
                         val, 'B',
                         'Bytes written through write() syscall',
@@ -412,7 +412,7 @@
 
         m = RE['ms_read_odir'].match(line)
         if m:
-            val = long(m.group(1))
+            val = int(m.group(1))
             tracestat.add_stat(prefix + 'read_odirect',
                         val, 'B',
                         'Bytes read through read(O_DIRECT) syscall',
@@ -424,7 +424,7 @@
 
         m = RE['ms_write_odir'].match(line)
         if m:
-            val = long(m.group(1))
+            val = int(m.group(1))
             tracestat.add_stat(prefix + 'write_odirect',
                         val, 'B',
                         'Bytes written through write(O_DIRECT) syscall',
@@ -436,7 +436,7 @@
 
         m = RE['ms_read_nfs'].match(line)
         if m:
-            val = long(m.group(1))
+            val = int(m.group(1))
             tracestat.add_stat(prefix + 'read_nfs',
                         val, 'B',
                         'Bytes read via NFS RPCs',
@@ -448,7 +448,7 @@
 
         m = RE['ms_write_nfs'].match(line)
         if m:
-            val = long(m.group(1))
+            val = int(m.group(1))
             tracestat.add_stat(prefix + 'write_nfs',
                         val, 'B',
                         'Bytes written via NFS RPCs',
@@ -461,21 +461,21 @@
         m = RE['ms_rpc_line'].match(line)
         if m:
             tracestat.add_stat(prefix + 'rpc_requests',
-                        long(m.group(1)), 'RPCs',
+                        int(m.group(1)), 'RPCs',
                         'Count of RPC requests',
                         BETTER_LESS_IF_IO_BOUND,
                         None,
                         filename,
                         tracedir)
             tracestat.add_stat(prefix + 'rpc_replies',
-                        long(m.group(2)), 'RPCs',
+                        int(m.group(2)), 'RPCs',
                         'Count of RPC replies',
                         BETTER_LESS_IF_IO_BOUND,
                         None,
                         filename,
                         tracedir)
             tracestat.add_stat(prefix + 'xid_not_found',
-                        long(m.group(3)), 'RPCs',
+                        int(m.group(3)), 'RPCs',
                         'Count of RPC replies that couldn\'t be matched ' +
                         'with a request',
                         BETTER_ALWAYS_LESS,
@@ -487,7 +487,7 @@
         m = RE['ms_rpc_backlog'].match(line)
         if m:
             tracestat.add_stat(prefix + 'backlog_queue_avg',
-                        long(m.group(1)), 'RPCs',
+                        int(m.group(1)), 'RPCs',
                         'Average number of outgoing requests on the backlog ' +
                         'queue',
                         BETTER_ALWAYS_LESS,
@@ -500,9 +500,10 @@
     op = None
     oplineno = 0
     for line in f:
-        m = RE['ms_ops_header'].match(line.strip())
+        ls = line.strip()
+        m = RE['ms_ops_header'].match(ls)
         if m:
-            assert op == None
+            #assert op is None,"failed op==None, m==%s" % m
             op = m.group(1)
             op_bucket = mountstat_op_map.get(op, BUCKET_OTHER)
             oplineno = 1
@@ -511,7 +512,7 @@
         if oplineno == 1:
             m = RE['ms_ops_line1'].match(line)
             if m:
-                assert op != None
+                assert op is not None,"failed op != None"
                 oplineno += 1
                 continue
 
@@ -563,6 +564,8 @@
         elif op:
             raise ParseError("Didn't match line: %s" % line)
 
+    f.close()
+
 def parse_nfsiostat(tracestat, tracedir, attrs):
     prefix = 'nfsiostat:'
     stat_desc = 'output of nfsiostat(1)'
@@ -570,7 +573,7 @@
 
     path = os.path.join(tracedir, filename)
 
-    lines = file(path).readlines()
+    lines = open(path).readlines()
 
     # skip until we find our mount
     name=None
@@ -656,7 +659,7 @@
 
     path = os.path.join(tracedir, filename)
 
-    lines = file(path).readlines()
+    lines = open(path).readlines()
 
     m = RE['ns_rpc_title'].match(lines[0])
 
@@ -675,7 +678,7 @@
         raise ParseError("Can't find RPC call count")
 
     tracestat.add_stat(prefix + 'rpc_calls',
-                long(m.group(1)), 'Calls',
+                int(m.group(1)), 'Calls',
                 'Count of RPC calls',
                 BETTER_LESS_IF_IO_BOUND,
                 None,
@@ -709,12 +712,12 @@
             m = RE['ns_count_data'].match(line)
             if m:
                 for i, t in enumerate(titles):
-                    assert not op_counts.has_key(t), "dup op count %s" % t
-                    op_counts[t] = long(m.group(i+1))
+                    assert t not in op_counts, "dup op count %s" % t
+                    op_counts[t] = int(m.group(i+1))
 
             titles = None
 
-    for op, count in op_counts.iteritems():
+    for op, count in six.iteritems(op_counts):
         if count:
             op_bucket = nfsstat_op_map.get(op, BUCKET_OTHER)
             tracestat.add_stat(prefix + op.upper() + ' Count',
@@ -734,7 +737,8 @@
 
     # NOTE: BETTER_* based on fact that filebench output is only ever time bound
     found = False
-    for line in file(path):
+    with open(path) as f:
+      for line in f:
         m = RE['filebench_stats'].match(line)
         if m:
             tracestat.add_stat(prefix + 'op_count',
@@ -784,16 +788,15 @@
 
     path = os.path.join(tracedir, filename)
 
-    f = file(path)
-
-    found = False
-    for line in f:
+    with open(path) as f:
+      found = False
+      for line in f:
         m = RE['pms_events'].match(line)
         if m:
             found = True
 
             tracestat.add_stat(prefix + 'inode_revalidate',
-                        long(m.group(1)), 'events',
+                        int(m.group(1)), 'events',
                         'Count of inode_revalidate events',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -801,7 +804,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'dentry_revalidate',
-                        long(m.group(2)), 'events',
+                        int(m.group(2)), 'events',
                         'Count of dentry_revalidate events',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -809,7 +812,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'data_invalidate',
-                        long(m.group(3)), 'events',
+                        int(m.group(3)), 'events',
                         'Count of data_invalidate events',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -817,7 +820,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'attr_invalidate',
-                        long(m.group(4)), 'events',
+                        int(m.group(4)), 'events',
                         'Count of attr_invalidate events',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -825,7 +828,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_open',
-                        long(m.group(5)), 'events',
+                        int(m.group(5)), 'events',
                         'Count of file and directory opens',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -833,7 +836,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_lookup',
-                        long(m.group(6)), 'events',
+                        int(m.group(6)), 'events',
                         'Count of lookups',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -841,7 +844,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_access',
-                        long(m.group(7)), 'events',
+                        int(m.group(7)), 'events',
                         'Count of access calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -849,7 +852,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_updatepage',
-                        long(m.group(8)), 'events',
+                        int(m.group(8)), 'events',
                         'Count of updatepage calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -857,7 +860,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_readpage',
-                        long(m.group(9)), 'events',
+                        int(m.group(9)), 'events',
                         'Count of readpage calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -865,7 +868,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_readpages',
-                        long(m.group(10)), 'events',
+                        int(m.group(10)), 'events',
                         'Count of readpages calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -873,7 +876,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_writepage',
-                        long(m.group(11)), 'events',
+                        int(m.group(11)), 'events',
                         'Count of writepage calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -881,7 +884,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_writepages',
-                        long(m.group(12)), 'events',
+                        int(m.group(12)), 'events',
                         'Count of writepages calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -889,7 +892,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_getdents',
-                        long(m.group(13)), 'events',
+                        int(m.group(13)), 'events',
                         'Count of getdents calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -897,7 +900,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_setattr',
-                        long(m.group(14)), 'events',
+                        int(m.group(14)), 'events',
                         'Count of setattr calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -905,7 +908,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_flush',
-                        long(m.group(15)), 'events',
+                        int(m.group(15)), 'events',
                         'Count of flush calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -913,7 +916,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_fsync',
-                        long(m.group(16)), 'events',
+                        int(m.group(16)), 'events',
                         'Count of fsync calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -921,7 +924,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_lock',
-                        long(m.group(17)), 'events',
+                        int(m.group(17)), 'events',
                         'Count of lock calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -929,7 +932,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'vfs_release',
-                        long(m.group(18)), 'events',
+                        int(m.group(18)), 'events',
                         'Count of release calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -937,7 +940,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'congestion_wait',
-                        long(m.group(19)), 'events',
+                        int(m.group(19)), 'events',
                         'Count of congestion_wait',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -945,7 +948,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'setattr_trunc',
-                        long(m.group(20)), 'events',
+                        int(m.group(20)), 'events',
                         'Count of setattr_trunc',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -953,7 +956,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'extend_write',
-                        long(m.group(21)), 'events',
+                        int(m.group(21)), 'events',
                         'Count of extend_write',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -961,7 +964,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'silly_rename',
-                        long(m.group(22)), 'events',
+                        int(m.group(22)), 'events',
                         'Count of silly_rename',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -969,7 +972,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'short_read',
-                        long(m.group(23)), 'events',
+                        int(m.group(23)), 'events',
                         'Count of short_read',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -977,7 +980,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'short_write',
-                        long(m.group(24)), 'events',
+                        int(m.group(24)), 'events',
                         'Count of short_write',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -985,7 +988,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'delay',
-                        long(m.group(25)), 'events',
+                        int(m.group(25)), 'events',
                         'Count of delays (v3: JUKEBOX, v4: ERR_DELAY, grace period, key expired)',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -993,7 +996,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'pnfs_read',
-                        long(m.group(26)), 'events',
+                        int(m.group(26)), 'events',
                         'Count of pnfs_read calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -1001,7 +1004,7 @@
                         tracedir)
 
             tracestat.add_stat(prefix + 'pnfs_write',
-                        long(m.group(27)), 'events',
+                        int(m.group(27)), 'events',
                         'Count of pnfs_write calls',
                         BETTER_ALWAYS_LESS | BETTER_NO_VARIANCE,
                         None,
@@ -1016,7 +1019,7 @@
             if len(values) > 10:
                 # older mountstats don't have so many values
                 tracestat.add_stat(prefix + 'xprt_max_slots',
-                        long(values[10]), 'slots',
+                        int(values[10]), 'slots',
                         'Max slots used by rpc transport',
                         BETTER_ALWAYS_LESS,
                         None,
@@ -1031,7 +1034,7 @@
             if len(values) > 8:
                 # older mountstats don't have so many values
                 tracestat.add_stat(prefix + 'xprt_max_slots',
-                        long(values[8]), 'slots',
+                        int(values[8]), 'slots',
                         'Max slots used by rpc transport',
                         BETTER_ALWAYS_LESS,
                         None,
@@ -1049,7 +1052,6 @@
     filename = 'test.log'
 
     path = os.path.join(tracedir, filename)
-    f = file(path)
 
     rpt_name = None
     rpt_col_hdr = []
@@ -1057,7 +1059,8 @@
     # maps name -> (%u_%u) -> value
     newkeys = []
 
-    for line in f:
+    with open(path) as f:
+      for line in f:
         line = line.strip()
         if rpt_name:
             if not line:
@@ -1097,7 +1100,7 @@
         y = int(skey[-1])
 
         tracestat.add_stat(prefix + key + ' iozone',
-            long(value), 'KB/s',
+            int(value), 'KB/s',
             '%s: size kb: %u, reclen: %u' % (report, x, y),
             BETTER_ALWAYS_MORE,
             (iozone_bucket_def, report + ' iozone'),
diff -ur nfsometer-1.9.orig/nfsometerlib/report.py nfsometer-1.9.wip/nfsometerlib/report.py
--- nfsometer-1.9.orig/nfsometerlib/report.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/report.py	2019-05-20 13:40:47.224411050 +0100
@@ -15,11 +15,11 @@
 from math import sqrt, pow
 import time
 
-import graph
-from collection import *
-from selector import Selector, SELECTOR_ORDER
-from config import *
-from workloads import *
+from . import graph
+from .collection import *
+from .selector import Selector, SELECTOR_ORDER
+from .config import *
+from .workloads import *
 
 ENABLE_PIE_GRAPHS=False
 
@@ -238,6 +238,8 @@
             cell = val
         elif isinstance(val, (Stat, Bucket)):
             cell = html_fmt_value(val.mean(), val.std(), units=self.units)
+        elif type(val).__name__ == "Bucket" or type(val).__name__ == 'Stat':
+            cell = html_fmt_value(val.mean(), val.std(), units=self.units)
         else:
             assert val == None, "Not a string, Stat or Bucket: %r\ng = %s, k = %s" % (val, g, k)
 
@@ -279,7 +281,7 @@
         cur = 0
 
         for vers in NFS_VERSIONS:
-            if not gmap.has_key(vers):
+            if vers not in gmap:
                 continue
 
             assert cur < num
@@ -392,7 +394,7 @@
             value_map[g] = {}
             v = vals.get(g, {}).get(key, None)
             if v != None:
-                if isinstance(v, Bucket):
+                if isinstance(v, Bucket) or type(v).__name__ == 'Bucket':
                     self.all_buckets = True
                     for stat in v.foreach():
                         value_map[g][stat.name] = stat.mean()
@@ -405,8 +407,9 @@
         self.bucket_pie = ''
 
         # does ordering matter here?
-        bk_order = [ (k,v) for k, v in self.hatch_map.iteritems() ]
-        bk_order.sort(lambda x,y: cmp(x[1], y[1]))
+        bk_order = [ (k,v) for k, v in six.iteritems(self.hatch_map) ]
+        if bk_order:
+            bk_order.sort(key=lambda x: x[1])
 
         table_values = {}
         bucket_names = []
@@ -483,7 +486,7 @@
             self.make_comparison_vals(vals, key, groups, select_order)
 
         self.gmap = groups_by_nfsvers(groups)
-        self.nfs_versions = [ v for v in NFS_VERSIONS if self.gmap.has_key(v) ]
+        self.nfs_versions = [ v for v in NFS_VERSIONS if v in self.gmap ]
 
         # ensure the order of groups is in nfs_version order
         groups = []
@@ -549,7 +552,7 @@
                  ' style="display: none;">%s</div>' % (hits,)
 
 
-        for compare, compvals in self.comparison_vals_map.iteritems():
+        for compare, compvals in six.iteritems(self.comparison_vals_map):
             if compvals:
                 c += '<div class="compare_%s" ' \
                      'style="display: none;">' \
@@ -563,7 +566,7 @@
             table_rows = []
             color_idx = COLORS.index(self.color_map[g])
 
-            if isinstance(stat, Bucket):
+            if isinstance(stat, Bucket) or type(stat).__name__ == 'Bucket':
                 table_hdrs.append('run')
                 for x in stat.foreach():
                     hidx = self.hatch_map[x.name]
@@ -585,7 +588,7 @@
                 row = []
                 row.append('<a href="%s">%s</a>' % (tracedir, run))
 
-                if isinstance(stat, Bucket):
+                if isinstance(stat, Bucket) or type(stat).__name__ == 'Bucket':
                     for x in stat.foreach():
                         row.append('<a href="%s/%s">%s</a>' %
                             (tracedir, stat.filename(),
@@ -619,26 +622,26 @@
             if stat == None:
                 continue
 
-            if isinstance(stat, Bucket):
+            if isinstance(stat, Bucket) or type(stat).__name__ == 'Bucket':
                 for sub in stat.foreach():
-                    if not key2val.has_key(sub.name):
+                    if sub.name not in key2val:
                         key2val[sub.name] = 0.0
                     key2val[sub.name] += sub.mean()
                     total_val += sub.mean()
             else:
                 # a basic Stat - makes hatch map with one entry
-                if not key2val.has_key(stat.name):
+                if stat.name not in key2val:
                     key2val[stat.name] = 0.0
                 key2val[stat.name] += stat.mean()
                 total_val += stat.mean()
 
-        ordered = [ (k, v) for k, v in key2val.iteritems() ]
-        ordered.sort(lambda x,y: cmp(x[1], y[1]))
-        ordered.reverse()
+        ordered = [ (k, v) for k, v in six.iteritems(key2val) ]
+        if ordered:
+             ordered.sort(key=lambda x: x[1], reverse=True)
 
         k2h = {}
         for i, kv in enumerate(ordered):
-            assert not k2h.has_key(kv[0])
+            assert kv[0] not in k2h
             k2h[kv[0]] = i
 
         return k2h, key2val, total_val
@@ -656,7 +659,7 @@
         for g in groups:
             idx = None
             for i, cg in enumerate(compare_groups):
-                if g.compare_order(cg[0], select_order) == 0:
+                if g.match_order(cg[0], select_order):
                     # found a group!
                     idx = i
                     break
@@ -671,7 +674,7 @@
             ref_val = None
             ref_g = None
             for g in cg:
-                if not newvals.has_key(g):
+                if g not in newvals:
                     newvals[g] = {}
 
                 # handle no data
@@ -727,9 +730,9 @@
 
     def fmt_cell_hits(self, value):
         classes = ('hatch_hit',)
-        if isinstance(value, Bucket):
+        if isinstance(value, Bucket) or type(value).__name__ == 'Bucket':
             stat_list = [ x for x in value.foreach() ]
-            stat_list.sort(lambda x,y: -1 * cmp(x.mean(), y.mean()))
+            stat_list.sort(key=lambda x: x.mean(), reverse=True)
 
             units = self.report.collection.stat_units(value.name)
 
@@ -1045,8 +1048,9 @@
                              bucket_def=self.bucket_def,
                              units=units)
 
-        bucket_info = [ (k, v) for k, v in bucket_totals.iteritems() ]
-        bucket_info.sort(lambda x, y: cmp(x[1], y[1]) * -1)
+        bucket_info = [ (k, v) for k, v in six.iteritems(bucket_totals) ]
+        if bucket_info:
+            bucket_info.sort(key=lambda x: x[1], reverse=True)
 
         bucket_info = [ x for x in bucket_info if x[1] ]
 
@@ -1166,12 +1170,18 @@
 
         for wsel in topsel.foreach('workload'):
             workload_name = wsel.fmt('workload')
+            workload_command = ''
+            workload_description = ''
 
             # XXX 0?
-            workload_command = \
-                self.collection.get_attr(wsel, 'workload_command')[0]
-            workload_description = \
-                self.collection.get_attr(wsel, 'workload_description')[0]
+            col = self.collection
+            atr = col.get_attr(wsel, 'workload_command')
+            if atr:
+              workload_command = atr[0]
+
+            atr = col.get_attr(wsel, 'workload_description')
+            if atr:
+              workload_description = atr[0]
 
             wdesc = '<span class="workload_name">%s</span>' \
                     '<span class="workload_description">%s</span>' % \
@@ -1228,7 +1238,7 @@
                     if not mdt in seen_mdts:
                         seen_mdts.append(mdt)
 
-                    if not mount_options.has_key(mdt):
+                    if mdt not in mount_options:
                         mount_options[mdt] = set()
                     mount_options[mdt] = \
                         mount_options[mdt].union(real_info['mount_options'])
@@ -1236,7 +1246,7 @@
                     # lowlite (opposite of hilite) values same as prev row.
                     info = {}
                     ignore = ('runs',)
-                    for k in real_info.keys():
+                    for k in list(real_info.keys()):
                         if not k in ignore and last_info and \
                            real_info[k] == last_info[k]:
                             info[k] = '<span class="lowlite">%s</span>' % \
@@ -1425,14 +1435,14 @@
 
     def _write_report(self, r):
         abs_path = os.path.join(self.reportdir, r.path)
-        file(abs_path, 'w+').write(r.html())
-        print " %s" % r.path
+        with open(abs_path, 'w+') as f:
+          f.write(r.html())
 
     def _write_index(self):
         path = 'index.html'
         abs_path = os.path.join(self.reportdir, path)
-        file(abs_path, 'w+').write(self.html_index())
-        print " %s" % path
+        with open(abs_path, 'w+') as f:
+          f.write(self.html_index())
 
     def _step_through_reports(self, cb_f):
         for x in self.collection.selection.foreach('workload'):
diff -ur nfsometer-1.9.orig/nfsometerlib/selector.py nfsometer-1.9.wip/nfsometerlib/selector.py
--- nfsometer-1.9.orig/nfsometerlib/selector.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/selector.py	2019-05-20 13:40:54.882438074 +0100
@@ -11,7 +11,7 @@
 FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
 """
 
-from config import *
+from .config import *
 
 SELECTOR_ORDER=(
     'workload',
@@ -70,19 +70,29 @@
 
         return hash(tuple(args))
 
-    def __cmp__(self, other):
+    def __eq__(self, other):
         for name in SELECTOR_ORDER:
-            r = cmp(getattr(self, name + 's'), getattr(other, name + 's'))
-            if r:
-                return r
-        return 0
+            if getattr(self, name + 's') != getattr(other, name + 's'):
+                return False
+        return True
 
-    def compare_order(self, other, order):
+    def __lt__(self, other):
+        return NotImplemented
+
+    def __le__(self, other):
+        return NotImplemented
+
+    def __gt__(self, other):
+        return NotImplemented
+
+    def __ge__(self, other):
+        return NotImplemented
+
+    def match_order(self, other, order):
         for name in order:
-            r = cmp(getattr(self, name + 's'), getattr(other, name + 's'))
-            if r != 0:
-                return r
-        return 0
+            if getattr(self, name + 's') != getattr(other, name + 's'):
+                return False
+        return True
 
     def __repr__(self):
         args = []
@@ -102,7 +112,7 @@
         elif hasattr(superself, attr):
             return getattr(superself, attr)
         else:
-            raise AttributeError, "invalid attribute: %r" % attr
+            raise AttributeError("invalid attribute: %r" % attr)
 
     def __add__(self, other):
         new = Selector()
diff -ur nfsometer-1.9.orig/nfsometerlib/trace.py nfsometer-1.9.wip/nfsometerlib/trace.py
--- nfsometer-1.9.orig/nfsometerlib/trace.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/trace.py	2019-05-20 13:41:11.549496892 +0100
@@ -22,10 +22,10 @@
 import atexit
 import pwd
 
-from cmd import *
-from config import *
-from workloads import *
-import selector
+from .cmd import *
+from .config import *
+from .workloads import *
+from . import selector
 
 _server_path_v4 = re.compile('([^:]+):(\S+)')
 _server_path_v6 = re.compile('(\[\S+\]):(\S+)')
@@ -70,8 +70,8 @@
 
         if not temp and not new:
             try:
-                f = file(self.__attrfile)
-            except IOError, e:
+                f = open(self.__attrfile)
+            except IOError as e:
                 raise IOError('Attr file not found')
             for line in f:
                 if line.strip():
@@ -79,13 +79,13 @@
                     self.__attrs[name.strip()] = \
                         val.strip().replace('\\n', '\n')
 
-        if not self.__attrs.has_key('tracedir'):
+        if 'tracedir' not in self.__attrs:
             self.__attrs['tracedir'] = RUNNING_TRACE_DIR
 
-        if not self.__attrs.has_key('stoptime'):
+        if 'stoptime' not in self.__attrs:
             self.__attrs['stoptime'] = 'ongoing'
 
-        if not self.__attrs.has_key('tracedir_version'):
+        if 'tracedir_version' not in self.__attrs:
             if new or temp:
                 self.__attrs['tracedir_version'] = TRACE_DIR_VERSION
             else:
@@ -101,7 +101,7 @@
             if tracedir_vers == 1:
                 # move tags from separate attrs to one 'tags' attr
                 v1_tags = [ 'delegations_enabled', 'pnfs_enabled', 'remote' ]
-                tag_names = [ x for x in self.__attrs.keys() if x in v1_tags ]
+                tag_names = [ x for x in list(self.__attrs.keys()) if x in v1_tags ]
 
                 for name in tag_names:
                     assert int(self.__attrs[name]) == 1
@@ -202,12 +202,12 @@
         self.__attrs['tracedir_version'] = TRACE_DIR_VERSION
 
     def _sorted_names(self):
-        names = self.__attrs.keys()
+        names = list(self.__attrs.keys())
         names.sort()
         return names
 
     def get(self, name, *args):
-        if self.__attrs.has_key(name):
+        if name in self.__attrs:
             return self.__attrs[name]
 
         # handle optional default value
@@ -233,14 +233,14 @@
     def write(self):
         if self.__temp:
             return
-        f = file(self.__attrfile, 'w+')
-        for k, v in self.__attrs.iteritems():
+        with open(self.__attrfile, 'w+') as f:
+          for k, v in six.iteritems(self.__attrs):
             f.write('%s = %s\n' % (k, str(v).replace('\n', '\\n')))
 
 def _dir_create():
     try:
         os.mkdir(RUNNING_TRACE_DIR)
-    except OSError, e:
+    except OSError as e:
         if e.errno == errno.EEXIST:
             raise IOError('An NFS trace is already running')
         raise
@@ -292,7 +292,7 @@
         for old_syntax in (False, True):
             try:
                 _mount(attrs, old_syntax=old_syntax)
-            except Exception, e:
+            except Exception as e:
                 if not quiet:
                     sys.stdout.write('.')
                     sys.stdout.flush()
@@ -310,7 +310,7 @@
         sys.stdout.flush()
 
     if err:
-        raise e
+        raise err
 
 def _is_mounted(attrs):
     try:
@@ -358,7 +358,7 @@
     for i in range(tries):
         try:
             _unmount(attrs)
-        except Exception, e:
+        except Exception as e:
             if not quiet:
                 sys.stdout.write('.')
                 sys.stdout.flush()
@@ -441,12 +441,12 @@
     for c in commands:
         stats.append(c['file'])
         out = cmd(c['cmd'])
-        f = file(os.path.join(RUNNING_TRACE_DIR, c['file']), 'w+')
-        f.write('\n'.join(out[0]))
+        with open(os.path.join(RUNNING_TRACE_DIR, c['file']), 'w+') as f:
+          f.write('\n'.join(out[0]))
 
 def probe_detect(probe_trace_dir, mountopt):
     lines = [ x.strip()
-             for x in file(os.path.join(probe_trace_dir,
+             for x in open(os.path.join(probe_trace_dir,
                            'proc_mountstats.stop')) ]
 
     # find this mountpoint
@@ -519,7 +519,7 @@
     return detect
 
 def _is_auth_gss():
-    lines = [ x.strip() for x in file('/proc/self/mountstats') ]
+    lines = [ x.strip() for x in open('/proc/self/mountstats') ]
     mounted_on = ' mounted on %s with ' % MOUNTDIR
     start = -1
     end = -1
@@ -554,7 +554,7 @@
 
 def _has_tkt(server):
     princ = re.compile('nfs/' + server + '\S+$')
-    lines = [ x.strip() for x in file(os.path.join(RUNNING_TRACE_DIR,
+    lines = [ x.strip() for x in open(os.path.join(RUNNING_TRACE_DIR,
                                                    'klist_user.start')) ]
     for i, line in enumerate(lines):
         if re.search(princ, line):
@@ -625,7 +625,7 @@
     attrs.set('server', server)
     attrs.set('path', path)
     attrs.set('localpath', MOUNTDIR)
-    attrs.set('starttime', long(time.time()))
+    attrs.set('starttime', int(time.time()))
     attrs.set('workload', workload)
     attrs.set('workload_command', workload_command(workload, pretty=True))
     attrs.set('workload_description', workload_description(workload))
@@ -655,8 +655,8 @@
     attrs.set('stoptime', time.time())
     attrs.write()
 
-    is_setup = long(attrs.get('is_setup', 0))
-    is_probe = long(attrs.get('is_probe', 0))
+    is_setup = int(attrs.get('is_setup', 0))
+    is_probe = int(attrs.get('is_probe', 0))
 
     if not is_setup:
         _save_stop_stats(attrs)
@@ -669,11 +669,11 @@
     if resdir != None:
         cmd('mv %s %s' % (RUNNING_TRACE_DIR, resdir))
         if not is_probe:
-            print 'Results copied to: %s' % (os.path.split(resdir)[-1],)
+            print('Results copied to: %s' % (os.path.split(resdir)[-1],))
     else:
         cmd('rm -rf %s' % (RUNNING_TRACE_DIR))
         if not is_setup:
-            print 'Results thrown away'
+            print('Results thrown away')
 
 def find_mounted_serverpath(mountdir):
     try:
@@ -723,16 +723,16 @@
                 workloads[name] = obj
                 new.append(name)
             except KeyError:
-                print
+                print('')
                 warn('Invalid workload: "%s"' % w)
-                print
-                print "Available workloads:"
-                print "  %s" % '\n  '.join(available_workloads())
+                print('')
+                print("Available workloads:")
+                print("  %s" % '\n  '.join(available_workloads()))
                 sys.exit(2)
         workloads_requested = new
 
     else:
-        for w, workload_obj in WORKLOADS.iteritems():
+        for w, workload_obj in six.iteritems(WORKLOADS):
             if not workload_obj.check():
                 name = workload_obj.name()
                 workloads[name] = workload_obj
@@ -745,7 +745,7 @@
     current_kernel = get_current_kernel()
     client = get_current_hostname()
 
-    for w, workload_obj in workloads.iteritems():
+    for w, workload_obj in six.iteritems(workloads):
         for mountopt, detects, tags in mountopts_detects_tags:
             sel = selector.Selector(w, current_kernel, mountopt,
                                     detects, tags,
@@ -820,7 +820,7 @@
 
         try:
             cmd('mkdir -p "%s"' % RUNROOT)
-        except CmdErrorCode, e:
+        except CmdErrorCode as e:
             msg = str.format('"mkdir -p {:s}" failed.', RUNROOT)
             # try to hint why it failed
             if e.code == errno.EPERM:
@@ -831,9 +831,8 @@
             # and bail out right now
             sys.exit(1)
 
-        f = file(fpath, 'w+')
-        f.write('nfsometer probe to determine server features: %s' % m)
-        f.close()
+        with open(fpath, 'w+') as f:
+            f.write('nfsometer probe to determine server features: %s' % m)
 
         # force delegation if supported
         fd1 = os.open(fpath, os.O_RDWR)
@@ -869,14 +868,14 @@
                        mountopts_detects_tags, opts.num_runs, opts.server,
                        opts.path)
 
-    for w, workload_obj in workloads.iteritems():
+    for w, workload_obj in six.iteritems(workloads):
         workload_obj.fetch()
 
     if fetch_only:
         return
 
     # check each workload to make sure we'll be able to run it
-    for w, workload_obj in workloads.iteritems():
+    for w, workload_obj in six.iteritems(workloads):
         check_mesg = workload_obj.check()
 
         if check_mesg:
@@ -884,12 +883,12 @@
 
     this_trace = 0
 
-    print
-    print "Requested: %u workloads X %u options X %u runs = %u traces" % \
-        (len(workloads), len(mountopts_detects_tags), int(opts.num_runs), requested)
+    print('')
+    print("Requested: %u workloads X %u options X %u runs = %u traces" % \
+        (len(workloads), len(mountopts_detects_tags), int(opts.num_runs), requested))
     if skipped:
-        print "Results directory already has %u matching traces" % (skipped,)
-    print "Need to run %u of %u requested traces" % (total, requested)
+        print("Results directory already has %u matching traces" % (skipped,))
+    print("Need to run %u of %u requested traces" % (total, requested))
 
     for workload_obj, mountopt, detects, tags, nruns in trace_list:
         mdt = mountopt
@@ -897,8 +896,8 @@
             mdt += ' ' + detects
         if tags:
             mdt += ' ' + tags
-        print " %s - needs %u runs of %s" % (workload_obj.name(), nruns, mdt)
-    print
+        print(" %s - needs %u runs of %s" % (workload_obj.name(), nruns, mdt))
+    print('')
 
     dir_remove_old_asides()
 
@@ -916,7 +915,7 @@
         for run in range(nruns):
             this_trace += 1
 
-            print
+            print('')
             mdt = mountopt
             if detects:
                 mdt += ' ' + detects
@@ -925,7 +924,7 @@
 
             inform("Trace %u/%u: %u of %u for %s: %s" %
                    (this_trace, total, run+1, nruns, workload_obj.name(), mdt))
-            print
+            print('')
 
             sys.stdout.write("< SETUP WORKLOAD >\n")
             sys.stdout.flush()
@@ -936,7 +935,7 @@
 
             stop()
 
-            print
+            print('')
 
             sys.stdout.write("< RUN WORKLOAD >\n")
             sys.stdout.flush()
@@ -983,7 +982,7 @@
             inform("loadgen %u: %s stop" % (num, workload))
             stop = True
 
-        except Exception, e:
+        except Exception as e:
             warn("loadgen %u: %s error:\n%s" % (num, workload, e))
             time.sleep(1.0)
 
@@ -1029,7 +1028,7 @@
         workpool.terminate()
         workpool.join()
 
-    except Exception, e:
+    except Exception as e:
         workpool.terminate()
         workpool.join()
         raise e
diff -ur nfsometer-1.9.orig/nfsometerlib/workloads.py nfsometer-1.9.wip/nfsometerlib/workloads.py
--- nfsometer-1.9.orig/nfsometerlib/workloads.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometerlib/workloads.py	2019-05-20 13:41:27.334552597 +0100
@@ -14,15 +14,15 @@
 import errno
 import re
 
-from cmd import *
-from config import *
+from .cmd import *
+from .config import *
 
 _re_which = re.compile('[\s\S]*which: no (\S+) in \([\s\S]*')
 
 def _mkdir_quiet(path):
     try:
         os.mkdir(path)
-    except OSError, e:
+    except OSError as e:
         if e.errno != errno.EEXIST:
             raise e
 
@@ -94,15 +94,15 @@
 
             if not os.path.exists(url_out):
                 if url.startswith('git://'):
-                    print "Fetching git: %s" % url
+                    print("Fetching git: %s" % url)
                     fetch_cmd = 'git clone "%s" "%s"' % (url, url_out)
                 else:
-                    print "Fetching url: %s" % url
+                    print("Fetching url: %s" % url)
                     fetch_cmd = 'wget -O "%s" "%s"' % (url_out, url)
 
                 try:
                     cmd(fetch_cmd, pass_output=True, raiseerrorout=True)
-                except Exception, e:
+                except Exception as e:
                     cmd('rm -rf "%s"' % url_out)
                 finally:
                     if not os.path.exists(url_out):
@@ -115,7 +115,7 @@
             assert not url and not url_out
 
     def check(self):
-        if not self._cache.has_key('check'):
+        if not 'check' in self._cache:
             res = cmd('%s check %s' % (self.script, self.defname))
             res = ', '.join([ x.strip() for x in res[0]]).strip()
             self._cache['check'] = res
@@ -123,7 +123,7 @@
         return self._cache['check']
 
     def command(self):
-        if not self._cache.has_key('command'):
+        if not 'command' in self._cache:
             res = cmd('%s command %s' % (self.script, self.defname))
             res = '\n'.join(res[0]).strip()
             assert not '\n' in res
@@ -132,7 +132,7 @@
         return self._cache['command']
 
     def description(self):
-        if not self._cache.has_key('description'):
+        if not 'description' in self._cache:
             res = cmd('%s description %s' % (self.script, self.defname))
             res = '\n'.join(res[0]).strip()
             assert not '\n' in res
@@ -141,7 +141,7 @@
         return self._cache['description']
 
     def name(self):
-        if not self._cache.has_key('name'):
+        if not 'name' in self._cache:
             res = cmd('%s name %s' % (self.script, self.defname))
             res = '\n'.join(res[0]).strip()
             assert not '\n' in res
@@ -150,7 +150,7 @@
         return self._cache['name']
 
     def url(self):
-        if not self._cache.has_key('url'):
+        if not 'url' in self._cache:
             res = cmd('%s url %s' % (self.script, self.defname))
             res = '\n'.join(res[0]).strip()
             assert not '\n' in res
@@ -159,7 +159,7 @@
         return self._cache['url']
 
     def url_out(self):
-        if not self._cache.has_key('url_out'):
+        if not 'url_out' in self._cache:
             res = cmd('%s url_out %s' % (self.script, self.defname))
             res = '\n'.join(res[0]).strip()
             assert not '\n' in res
@@ -174,14 +174,15 @@
 
         command = self.command()
 
-        print "Running command: %s" % command
+        print("Running command: %s" % command)
         sys.stdout.flush()
 
         oldcwd = os.getcwd()
         os.chdir(self.rundir)
 
         # write command to file
-        file(cmdfile, 'w+').write(command)
+        with open(cmdfile, 'w+') as f:
+          f.write(command)
 
         sh_cmd = "sh %s > %s 2>&1" % (cmdfile, logfile)
         wrapped_cmd = '( time ( %s ) ) 2> %s' % (sh_cmd, timefile)
@@ -193,7 +194,7 @@
             # re-raise
             raise KeyboardInterrupt
 
-        except Exception, e:
+        except Exception as e:
             os.chdir(oldcwd)
             # re-raise
             raise e
@@ -207,14 +208,15 @@
         cmdfile = os.path.join(self.rundir, 'command.sh')
         command = self.command()
 
-        print "Running command without trace: %s" % command
+        print("Running command without trace: %s" % command)
         sys.stdout.flush()
 
         oldcwd = os.getcwd()
         os.chdir(self.rundir)
 
         # write command to file
-        file(cmdfile, 'w+').write(command)
+        with open(cmdfile, 'w+') as f:
+          f.write(command)
 
         sh_cmd = "sh %s > %s 2>&1" % (cmdfile, logfile)
 
@@ -226,7 +228,7 @@
             # re-raise
             raise KeyboardInterrupt
 
-        except Exception, e:
+        except Exception as e:
             os.chdir(oldcwd)
             # re-raise
             raise e
@@ -245,7 +247,7 @@
     WORKLOADS[w] = Workload(w)
 
 def workload_command(workload, pretty=False):
-    if workload == posix.environ.get('NFSOMETER_NAME', None):
+    if workload == os.getenv('NFSOMETER_NAME', None):
         workload = 'custom'
     try:
         obj = WORKLOADS[workload]
@@ -262,7 +264,7 @@
     return cmdstr
 
 def workload_description(workload):
-    if workload == posix.environ.get('NFSOMETER_NAME', None):
+    if workload == os.getenv('NFSOMETER_NAME', None):
         workload = 'custom'
     try:
         obj = WORKLOADS[workload]
@@ -273,8 +275,7 @@
 
 def available_workloads():
     o = []
-    defnames = WORKLOADS.keys()
-    defnames.sort()
+    defnames = sorted( WORKLOADS.keys() )
     for defname in defnames:
         check_mesg = WORKLOADS[defname].check()
 
@@ -287,7 +288,7 @@
     """ return a string containing a comma separated list of the available 
         workload """
     o = []
-    defnames = WORKLOADS.keys()
+    defnames = list(WORKLOADS.keys())
     defnames.sort()
     for defname in defnames:
         check_mesg = WORKLOADS[defname].check()
diff -ur nfsometer-1.9.orig/nfsometer.py nfsometer-1.9.wip/nfsometer.py
--- nfsometer-1.9.orig/nfsometer.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/nfsometer.py	2019-05-13 16:26:40.269894942 +0100
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+#!/usr/bin/env python3
 """
 Copyright 2012 NetApp, Inc. All Rights Reserved,
 contribution by Weston Andros Adamson <dros@netapp.com>
@@ -15,6 +15,7 @@
 import posix
 import sys
 import os
+import six
 
 from nfsometerlib import trace
 from nfsometerlib import options
@@ -46,19 +47,19 @@
     collection = TraceCollection(opts.resultdir)
     collection.notes_edit()
 
-    print 'Saved notes for results %s' % (opts.resultdir)
+    print('Saved notes for results %s' % (opts.resultdir))
 
 def mode_list(opts):
     collection = TraceCollection(opts.resultdir)
 
-    print 'Result directory \'%s\' contains:\n\n%s' % \
-        (opts.resultdir, '\n'.join(collection.show_contents(pre='')))
+    print('Result directory \'%s\' contains:\n\n%s' % \
+        (opts.resultdir, '\n'.join(collection.show_contents(pre=''))))
 
 def mode_workloads(opts):
-    print "Available workloads:"
-    print "  %s" % '\n  '.join(available_workloads())
-    print "Unavailable workloads:"
-    print "  %s" % '\n  '.join(unavailable_workloads())
+    print("Available workloads:")
+    print("  %s" % '\n  '.join(available_workloads()))
+    print("Unavailable workloads:")
+    print("  %s" % '\n  '.join(unavailable_workloads()))
 
 def mode_loadgen(opts):
     # XXX check idle?
@@ -75,7 +76,7 @@
     check_idle_before_start(opts)
     collection = TraceCollection(opts.resultdir)
     trace.run_traces(collection, opts, fetch_only=fetch_only)
-    print
+    print('')
 
 def mode_report(opts):
     collection = TraceCollection(opts.resultdir)
@@ -83,7 +84,7 @@
         rpt = ReportSet(collection, opts.serial_graph_gen)
         rpt.generate_reports()
     else:
-        print "No tracedirs found"
+        print("No tracedirs found")
 
 def main():
     opts = options.Options()
@@ -129,5 +130,5 @@
     try:
         main()
     except KeyboardInterrupt:
-        print >>sys.stderr, "\nCancelled by user...\n"
+        six.print_("\nCancelled by user...\n", file=sys.stderr)
 
diff -ur nfsometer-1.9.orig/setup.py nfsometer-1.9.wip/setup.py
--- nfsometer-1.9.orig/setup.py	2017-06-29 04:39:18.000000000 +0100
+++ nfsometer-1.9.wip/setup.py	2019-05-14 17:44:38.065138924 +0100
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+#!/usr/bin/env python3
 """
 Copyright 2012 NetApp, Inc. All Rights Reserved,
 contribution by Weston Andros Adamson <dros@netapp.com>
@@ -21,7 +21,7 @@
 class sdist(_sdist):
     def run(self):
         if not self.dry_run:
-            print "generating manpage %s" % NFSOMETER_MANPAGE
+            print("generating manpage %s" % NFSOMETER_MANPAGE)
             o = nfsometerlib.options.Options()
             o.generate_manpage(NFSOMETER_MANPAGE)
 
@@ -37,7 +37,7 @@
             manpath = self.root + manpath
             gzpath = self.root + gzpath
 
-        print "gzipping manpage %s" % (gzpath,)
+        print("gzipping manpage %s" % (gzpath,))
         os.system('mkdir -p %s' % manpath)
         os.system('gzip -f --stdout "%s" > "%s"' % (manpage, gzpath))
 
@@ -52,7 +52,7 @@
             old = self.root + old
             new = self.root + new
 
-        print "stripping .py from script %s" % (old,)
+        print("stripping .py from script %s" % (old,))
         os.rename(old, new)
 
     def run(self):