diff options
author | RajithaY <rajithax.yerrumsetty@intel.com> | 2017-04-25 03:31:15 -0700 |
---|---|---|
committer | Rajitha Yerrumchetty <rajithax.yerrumsetty@intel.com> | 2017-05-22 06:48:08 +0000 |
commit | bb756eebdac6fd24e8919e2c43f7d2c8c4091f59 (patch) | |
tree | ca11e03542edf2d8f631efeca5e1626d211107e3 /qemu/scripts/analyze-migration.py | |
parent | a14b48d18a9ed03ec191cf16b162206998a895ce (diff) |
Adding qemu as a submodule of KVMFORNFV
This Patch includes the changes to add qemu as a submodule to
kvmfornfv repo and make use of the updated latest qemu for the
execution of all testcase
Change-Id: I1280af507a857675c7f81d30c95255635667bdd7
Signed-off-by:RajithaY<rajithax.yerrumsetty@intel.com>
Diffstat (limited to 'qemu/scripts/analyze-migration.py')
-rwxr-xr-x | qemu/scripts/analyze-migration.py | 610 |
1 files changed, 0 insertions, 610 deletions
diff --git a/qemu/scripts/analyze-migration.py b/qemu/scripts/analyze-migration.py deleted file mode 100755 index 14553876a..000000000 --- a/qemu/scripts/analyze-migration.py +++ /dev/null @@ -1,610 +0,0 @@ -#!/usr/bin/env python -# -# Migration Stream Analyzer -# -# Copyright (c) 2015 Alexander Graf <agraf@suse.de> -# -# This library is free software; you can redistribute it and/or -# modify it under the terms of the GNU Lesser General Public -# License as published by the Free Software Foundation; either -# version 2 of the License, or (at your option) any later version. -# -# This library is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# Lesser General Public License for more details. -# -# You should have received a copy of the GNU Lesser General Public -# License along with this library; if not, see <http://www.gnu.org/licenses/>. - -import numpy as np -import json -import os -import argparse -import collections -import pprint - -def mkdir_p(path): - try: - os.makedirs(path) - except OSError: - pass - -class MigrationFile(object): - def __init__(self, filename): - self.filename = filename - self.file = open(self.filename, "rb") - - def read64(self): - return np.asscalar(np.fromfile(self.file, count=1, dtype='>i8')[0]) - - def read32(self): - return np.asscalar(np.fromfile(self.file, count=1, dtype='>i4')[0]) - - def read16(self): - return np.asscalar(np.fromfile(self.file, count=1, dtype='>i2')[0]) - - def read8(self): - return np.asscalar(np.fromfile(self.file, count=1, dtype='>i1')[0]) - - def readstr(self, len = None): - if len is None: - len = self.read8() - if len == 0: - return "" - return np.fromfile(self.file, count=1, dtype=('S%d' % len))[0] - - def readvar(self, size = None): - if size is None: - size = self.read8() - if size == 0: - return "" - value = self.file.read(size) - if len(value) != size: - raise Exception("Unexpected end of %s at 0x%x" % (self.filename, self.file.tell())) - return value - - def tell(self): - return self.file.tell() - - # The VMSD description is at the end of the file, after EOF. Look for - # the last NULL byte, then for the beginning brace of JSON. - def read_migration_debug_json(self): - QEMU_VM_VMDESCRIPTION = 0x06 - - # Remember the offset in the file when we started - entrypos = self.file.tell() - - # Read the last 10MB - self.file.seek(0, os.SEEK_END) - endpos = self.file.tell() - self.file.seek(max(-endpos, -10 * 1024 * 1024), os.SEEK_END) - datapos = self.file.tell() - data = self.file.read() - # The full file read closed the file as well, reopen it - self.file = open(self.filename, "rb") - - # Find the last NULL byte, then the first brace after that. This should - # be the beginning of our JSON data. - nulpos = data.rfind("\0") - jsonpos = data.find("{", nulpos) - - # Check backwards from there and see whether we guessed right - self.file.seek(datapos + jsonpos - 5, 0) - if self.read8() != QEMU_VM_VMDESCRIPTION: - raise Exception("No Debug Migration device found") - - jsonlen = self.read32() - - # Seek back to where we were at the beginning - self.file.seek(entrypos, 0) - - return data[jsonpos:jsonpos + jsonlen] - - def close(self): - self.file.close() - -class RamSection(object): - RAM_SAVE_FLAG_COMPRESS = 0x02 - RAM_SAVE_FLAG_MEM_SIZE = 0x04 - RAM_SAVE_FLAG_PAGE = 0x08 - RAM_SAVE_FLAG_EOS = 0x10 - RAM_SAVE_FLAG_CONTINUE = 0x20 - RAM_SAVE_FLAG_XBZRLE = 0x40 - RAM_SAVE_FLAG_HOOK = 0x80 - - def __init__(self, file, version_id, ramargs, section_key): - if version_id != 4: - raise Exception("Unknown RAM version %d" % version_id) - - self.file = file - self.section_key = section_key - self.TARGET_PAGE_SIZE = ramargs['page_size'] - self.dump_memory = ramargs['dump_memory'] - self.write_memory = ramargs['write_memory'] - self.sizeinfo = collections.OrderedDict() - self.data = collections.OrderedDict() - self.data['section sizes'] = self.sizeinfo - self.name = '' - if self.write_memory: - self.files = { } - if self.dump_memory: - self.memory = collections.OrderedDict() - self.data['memory'] = self.memory - - def __repr__(self): - return self.data.__repr__() - - def __str__(self): - return self.data.__str__() - - def getDict(self): - return self.data - - def read(self): - # Read all RAM sections - while True: - addr = self.file.read64() - flags = addr & (self.TARGET_PAGE_SIZE - 1) - addr &= ~(self.TARGET_PAGE_SIZE - 1) - - if flags & self.RAM_SAVE_FLAG_MEM_SIZE: - while True: - namelen = self.file.read8() - # We assume that no RAM chunk is big enough to ever - # hit the first byte of the address, so when we see - # a zero here we know it has to be an address, not the - # length of the next block. - if namelen == 0: - self.file.file.seek(-1, 1) - break - self.name = self.file.readstr(len = namelen) - len = self.file.read64() - self.sizeinfo[self.name] = '0x%016x' % len - if self.write_memory: - print self.name - mkdir_p('./' + os.path.dirname(self.name)) - f = open('./' + self.name, "wb") - f.truncate(0) - f.truncate(len) - self.files[self.name] = f - flags &= ~self.RAM_SAVE_FLAG_MEM_SIZE - - if flags & self.RAM_SAVE_FLAG_COMPRESS: - if flags & self.RAM_SAVE_FLAG_CONTINUE: - flags &= ~self.RAM_SAVE_FLAG_CONTINUE - else: - self.name = self.file.readstr() - fill_char = self.file.read8() - # The page in question is filled with fill_char now - if self.write_memory and fill_char != 0: - self.files[self.name].seek(addr, os.SEEK_SET) - self.files[self.name].write(chr(fill_char) * self.TARGET_PAGE_SIZE) - if self.dump_memory: - self.memory['%s (0x%016x)' % (self.name, addr)] = 'Filled with 0x%02x' % fill_char - flags &= ~self.RAM_SAVE_FLAG_COMPRESS - elif flags & self.RAM_SAVE_FLAG_PAGE: - if flags & self.RAM_SAVE_FLAG_CONTINUE: - flags &= ~self.RAM_SAVE_FLAG_CONTINUE - else: - self.name = self.file.readstr() - - if self.write_memory or self.dump_memory: - data = self.file.readvar(size = self.TARGET_PAGE_SIZE) - else: # Just skip RAM data - self.file.file.seek(self.TARGET_PAGE_SIZE, 1) - - if self.write_memory: - self.files[self.name].seek(addr, os.SEEK_SET) - self.files[self.name].write(data) - if self.dump_memory: - hexdata = " ".join("{0:02x}".format(ord(c)) for c in data) - self.memory['%s (0x%016x)' % (self.name, addr)] = hexdata - - flags &= ~self.RAM_SAVE_FLAG_PAGE - elif flags & self.RAM_SAVE_FLAG_XBZRLE: - raise Exception("XBZRLE RAM compression is not supported yet") - elif flags & self.RAM_SAVE_FLAG_HOOK: - raise Exception("RAM hooks don't make sense with files") - - # End of RAM section - if flags & self.RAM_SAVE_FLAG_EOS: - break - - if flags != 0: - raise Exception("Unknown RAM flags: %x" % flags) - - def __del__(self): - if self.write_memory: - for key in self.files: - self.files[key].close() - - -class HTABSection(object): - HASH_PTE_SIZE_64 = 16 - - def __init__(self, file, version_id, device, section_key): - if version_id != 1: - raise Exception("Unknown HTAB version %d" % version_id) - - self.file = file - self.section_key = section_key - - def read(self): - - header = self.file.read32() - - if (header > 0): - # First section, just the hash shift - return - - # Read until end marker - while True: - index = self.file.read32() - n_valid = self.file.read16() - n_invalid = self.file.read16() - - if index == 0 and n_valid == 0 and n_invalid == 0: - break - - self.file.readvar(n_valid * self.HASH_PTE_SIZE_64) - - def getDict(self): - return "" - - -class ConfigurationSection(object): - def __init__(self, file): - self.file = file - - def read(self): - name_len = self.file.read32() - name = self.file.readstr(len = name_len) - -class VMSDFieldGeneric(object): - def __init__(self, desc, file): - self.file = file - self.desc = desc - self.data = "" - - def __repr__(self): - return str(self.__str__()) - - def __str__(self): - return " ".join("{0:02x}".format(ord(c)) for c in self.data) - - def getDict(self): - return self.__str__() - - def read(self): - size = int(self.desc['size']) - self.data = self.file.readvar(size) - return self.data - -class VMSDFieldInt(VMSDFieldGeneric): - def __init__(self, desc, file): - super(VMSDFieldInt, self).__init__(desc, file) - self.size = int(desc['size']) - self.format = '0x%%0%dx' % (self.size * 2) - self.sdtype = '>i%d' % self.size - self.udtype = '>u%d' % self.size - - def __repr__(self): - if self.data < 0: - return ('%s (%d)' % ((self.format % self.udata), self.data)) - else: - return self.format % self.data - - def __str__(self): - return self.__repr__() - - def getDict(self): - return self.__str__() - - def read(self): - super(VMSDFieldInt, self).read() - self.sdata = np.fromstring(self.data, count=1, dtype=(self.sdtype))[0] - self.udata = np.fromstring(self.data, count=1, dtype=(self.udtype))[0] - self.data = self.sdata - return self.data - -class VMSDFieldUInt(VMSDFieldInt): - def __init__(self, desc, file): - super(VMSDFieldUInt, self).__init__(desc, file) - - def read(self): - super(VMSDFieldUInt, self).read() - self.data = self.udata - return self.data - -class VMSDFieldIntLE(VMSDFieldInt): - def __init__(self, desc, file): - super(VMSDFieldIntLE, self).__init__(desc, file) - self.dtype = '<i%d' % self.size - -class VMSDFieldBool(VMSDFieldGeneric): - def __init__(self, desc, file): - super(VMSDFieldBool, self).__init__(desc, file) - - def __repr__(self): - return self.data.__repr__() - - def __str__(self): - return self.data.__str__() - - def getDict(self): - return self.data - - def read(self): - super(VMSDFieldBool, self).read() - if self.data[0] == 0: - self.data = False - else: - self.data = True - return self.data - -class VMSDFieldStruct(VMSDFieldGeneric): - QEMU_VM_SUBSECTION = 0x05 - - def __init__(self, desc, file): - super(VMSDFieldStruct, self).__init__(desc, file) - self.data = collections.OrderedDict() - - # When we see compressed array elements, unfold them here - new_fields = [] - for field in self.desc['struct']['fields']: - if not 'array_len' in field: - new_fields.append(field) - continue - array_len = field.pop('array_len') - field['index'] = 0 - new_fields.append(field) - for i in xrange(1, array_len): - c = field.copy() - c['index'] = i - new_fields.append(c) - - self.desc['struct']['fields'] = new_fields - - def __repr__(self): - return self.data.__repr__() - - def __str__(self): - return self.data.__str__() - - def read(self): - for field in self.desc['struct']['fields']: - try: - reader = vmsd_field_readers[field['type']] - except: - reader = VMSDFieldGeneric - - field['data'] = reader(field, self.file) - field['data'].read() - - if 'index' in field: - if field['name'] not in self.data: - self.data[field['name']] = [] - a = self.data[field['name']] - if len(a) != int(field['index']): - raise Exception("internal index of data field unmatched (%d/%d)" % (len(a), int(field['index']))) - a.append(field['data']) - else: - self.data[field['name']] = field['data'] - - if 'subsections' in self.desc['struct']: - for subsection in self.desc['struct']['subsections']: - if self.file.read8() != self.QEMU_VM_SUBSECTION: - raise Exception("Subsection %s not found at offset %x" % ( subsection['vmsd_name'], self.file.tell())) - name = self.file.readstr() - version_id = self.file.read32() - self.data[name] = VMSDSection(self.file, version_id, subsection, (name, 0)) - self.data[name].read() - - def getDictItem(self, value): - # Strings would fall into the array category, treat - # them specially - if value.__class__ is ''.__class__: - return value - - try: - return self.getDictOrderedDict(value) - except: - try: - return self.getDictArray(value) - except: - try: - return value.getDict() - except: - return value - - def getDictArray(self, array): - r = [] - for value in array: - r.append(self.getDictItem(value)) - return r - - def getDictOrderedDict(self, dict): - r = collections.OrderedDict() - for (key, value) in dict.items(): - r[key] = self.getDictItem(value) - return r - - def getDict(self): - return self.getDictOrderedDict(self.data) - -vmsd_field_readers = { - "bool" : VMSDFieldBool, - "int8" : VMSDFieldInt, - "int16" : VMSDFieldInt, - "int32" : VMSDFieldInt, - "int32 equal" : VMSDFieldInt, - "int32 le" : VMSDFieldIntLE, - "int64" : VMSDFieldInt, - "uint8" : VMSDFieldUInt, - "uint16" : VMSDFieldUInt, - "uint32" : VMSDFieldUInt, - "uint32 equal" : VMSDFieldUInt, - "uint64" : VMSDFieldUInt, - "int64 equal" : VMSDFieldInt, - "uint8 equal" : VMSDFieldInt, - "uint16 equal" : VMSDFieldInt, - "float64" : VMSDFieldGeneric, - "timer" : VMSDFieldGeneric, - "buffer" : VMSDFieldGeneric, - "unused_buffer" : VMSDFieldGeneric, - "bitmap" : VMSDFieldGeneric, - "struct" : VMSDFieldStruct, - "unknown" : VMSDFieldGeneric, -} - -class VMSDSection(VMSDFieldStruct): - def __init__(self, file, version_id, device, section_key): - self.file = file - self.data = "" - self.vmsd_name = "" - self.section_key = section_key - desc = device - if 'vmsd_name' in device: - self.vmsd_name = device['vmsd_name'] - - # A section really is nothing but a FieldStruct :) - super(VMSDSection, self).__init__({ 'struct' : desc }, file) - -############################################################################### - -class MigrationDump(object): - QEMU_VM_FILE_MAGIC = 0x5145564d - QEMU_VM_FILE_VERSION = 0x00000003 - QEMU_VM_EOF = 0x00 - QEMU_VM_SECTION_START = 0x01 - QEMU_VM_SECTION_PART = 0x02 - QEMU_VM_SECTION_END = 0x03 - QEMU_VM_SECTION_FULL = 0x04 - QEMU_VM_SUBSECTION = 0x05 - QEMU_VM_VMDESCRIPTION = 0x06 - QEMU_VM_CONFIGURATION = 0x07 - QEMU_VM_SECTION_FOOTER= 0x7e - - def __init__(self, filename): - self.section_classes = { ( 'ram', 0 ) : [ RamSection, None ], - ( 'spapr/htab', 0) : ( HTABSection, None ) } - self.filename = filename - self.vmsd_desc = None - - def read(self, desc_only = False, dump_memory = False, write_memory = False): - # Read in the whole file - file = MigrationFile(self.filename) - - # File magic - data = file.read32() - if data != self.QEMU_VM_FILE_MAGIC: - raise Exception("Invalid file magic %x" % data) - - # Version (has to be v3) - data = file.read32() - if data != self.QEMU_VM_FILE_VERSION: - raise Exception("Invalid version number %d" % data) - - self.load_vmsd_json(file) - - # Read sections - self.sections = collections.OrderedDict() - - if desc_only: - return - - ramargs = {} - ramargs['page_size'] = self.vmsd_desc['page_size'] - ramargs['dump_memory'] = dump_memory - ramargs['write_memory'] = write_memory - self.section_classes[('ram',0)][1] = ramargs - - while True: - section_type = file.read8() - if section_type == self.QEMU_VM_EOF: - break - elif section_type == self.QEMU_VM_CONFIGURATION: - section = ConfigurationSection(file) - section.read() - elif section_type == self.QEMU_VM_SECTION_START or section_type == self.QEMU_VM_SECTION_FULL: - section_id = file.read32() - name = file.readstr() - instance_id = file.read32() - version_id = file.read32() - section_key = (name, instance_id) - classdesc = self.section_classes[section_key] - section = classdesc[0](file, version_id, classdesc[1], section_key) - self.sections[section_id] = section - section.read() - elif section_type == self.QEMU_VM_SECTION_PART or section_type == self.QEMU_VM_SECTION_END: - section_id = file.read32() - self.sections[section_id].read() - elif section_type == self.QEMU_VM_SECTION_FOOTER: - read_section_id = file.read32() - if read_section_id != section_id: - raise Exception("Mismatched section footer: %x vs %x" % (read_section_id, section_id)) - else: - raise Exception("Unknown section type: %d" % section_type) - file.close() - - def load_vmsd_json(self, file): - vmsd_json = file.read_migration_debug_json() - self.vmsd_desc = json.loads(vmsd_json, object_pairs_hook=collections.OrderedDict) - for device in self.vmsd_desc['devices']: - key = (device['name'], device['instance_id']) - value = ( VMSDSection, device ) - self.section_classes[key] = value - - def getDict(self): - r = collections.OrderedDict() - for (key, value) in self.sections.items(): - key = "%s (%d)" % ( value.section_key[0], key ) - r[key] = value.getDict() - return r - -############################################################################### - -class JSONEncoder(json.JSONEncoder): - def default(self, o): - if isinstance(o, VMSDFieldGeneric): - return str(o) - return json.JSONEncoder.default(self, o) - -parser = argparse.ArgumentParser() -parser.add_argument("-f", "--file", help='migration dump to read from', required=True) -parser.add_argument("-m", "--memory", help='dump RAM contents as well', action='store_true') -parser.add_argument("-d", "--dump", help='what to dump ("state" or "desc")', default='state') -parser.add_argument("-x", "--extract", help='extract contents into individual files', action='store_true') -args = parser.parse_args() - -jsonenc = JSONEncoder(indent=4, separators=(',', ': ')) - -if args.extract: - dump = MigrationDump(args.file) - - dump.read(desc_only = True) - print "desc.json" - f = open("desc.json", "wb") - f.truncate() - f.write(jsonenc.encode(dump.vmsd_desc)) - f.close() - - dump.read(write_memory = True) - dict = dump.getDict() - print "state.json" - f = open("state.json", "wb") - f.truncate() - f.write(jsonenc.encode(dict)) - f.close() -elif args.dump == "state": - dump = MigrationDump(args.file) - dump.read(dump_memory = args.memory) - dict = dump.getDict() - print jsonenc.encode(dict) -elif args.dump == "desc": - dump = MigrationDump(args.file) - dump.read(desc_only = True) - print jsonenc.encode(dump.vmsd_desc) -else: - raise Exception("Please specify either -x, -d state or -d dump") |