1272 lines
		
	
	
		
			45 KiB
		
	
	
	
		
			Python
		
	
	
	
			
		
		
	
	
			1272 lines
		
	
	
		
			45 KiB
		
	
	
	
		
			Python
		
	
	
	
| # SPDX-License-Identifier: GPL-2.0+
 | |
| # Copyright (c) 2016 Google, Inc
 | |
| #
 | |
| # Base class for all entries
 | |
| #
 | |
| 
 | |
| from collections import namedtuple
 | |
| import importlib
 | |
| import os
 | |
| import pathlib
 | |
| import sys
 | |
| import time
 | |
| 
 | |
| from binman import bintool
 | |
| from dtoc import fdt_util
 | |
| from patman import tools
 | |
| from patman.tools import to_hex, to_hex_size
 | |
| from patman import tout
 | |
| 
 | |
| modules = {}
 | |
| 
 | |
| # This is imported if needed
 | |
| state = None
 | |
| 
 | |
| # An argument which can be passed to entries on the command line, in lieu of
 | |
| # device-tree properties.
 | |
| EntryArg = namedtuple('EntryArg', ['name', 'datatype'])
 | |
| 
 | |
| # Information about an entry for use when displaying summaries
 | |
| EntryInfo = namedtuple('EntryInfo', ['indent', 'name', 'etype', 'size',
 | |
|                                      'image_pos', 'uncomp_size', 'offset',
 | |
|                                      'entry'])
 | |
| 
 | |
| class Entry(object):
 | |
|     """An Entry in the section
 | |
| 
 | |
|     An entry corresponds to a single node in the device-tree description
 | |
|     of the section. Each entry ends up being a part of the final section.
 | |
|     Entries can be placed either right next to each other, or with padding
 | |
|     between them. The type of the entry determines the data that is in it.
 | |
| 
 | |
|     This class is not used by itself. All entry objects are subclasses of
 | |
|     Entry.
 | |
| 
 | |
|     Attributes:
 | |
|         section: Section object containing this entry
 | |
|         node: The node that created this entry
 | |
|         offset: Offset of entry within the section, None if not known yet (in
 | |
|             which case it will be calculated by Pack())
 | |
|         size: Entry size in bytes, None if not known
 | |
|         pre_reset_size: size as it was before ResetForPack(). This allows us to
 | |
|             keep track of the size we started with and detect size changes
 | |
|         uncomp_size: Size of uncompressed data in bytes, if the entry is
 | |
|             compressed, else None
 | |
|         contents_size: Size of contents in bytes, 0 by default
 | |
|         align: Entry start offset alignment relative to the start of the
 | |
|             containing section, or None
 | |
|         align_size: Entry size alignment, or None
 | |
|         align_end: Entry end offset alignment relative to the start of the
 | |
|             containing section, or None
 | |
|         pad_before: Number of pad bytes before the contents when it is placed
 | |
|             in the containing section, 0 if none. The pad bytes become part of
 | |
|             the entry.
 | |
|         pad_after: Number of pad bytes after the contents when it is placed in
 | |
|             the containing section, 0 if none. The pad bytes become part of
 | |
|             the entry.
 | |
|         data: Contents of entry (string of bytes). This does not include
 | |
|             padding created by pad_before or pad_after. If the entry is
 | |
|             compressed, this contains the compressed data.
 | |
|         uncomp_data: Original uncompressed data, if this entry is compressed,
 | |
|             else None
 | |
|         compress: Compression algoithm used (e.g. 'lz4'), 'none' if none
 | |
|         orig_offset: Original offset value read from node
 | |
|         orig_size: Original size value read from node
 | |
|         missing: True if this entry is missing its contents
 | |
|         allow_missing: Allow children of this entry to be missing (used by
 | |
|             subclasses such as Entry_section)
 | |
|         allow_fake: Allow creating a dummy fake file if the blob file is not
 | |
|             available. This is mainly used for testing.
 | |
|         external: True if this entry contains an external binary blob
 | |
|         bintools: Bintools used by this entry (only populated for Image)
 | |
|         missing_bintools: List of missing bintools for this entry
 | |
|         update_hash: True if this entry's "hash" subnode should be
 | |
|             updated with a hash of the entry contents
 | |
|         comp_bintool: Bintools used for compress and decompress data
 | |
|         fake_fname: Fake filename, if one was created, else None
 | |
|         required_props (dict of str): Properties which must be present. This can
 | |
|             be added to by subclasses
 | |
|     """
 | |
|     fake_dir = None
 | |
| 
 | |
|     def __init__(self, section, etype, node, name_prefix=''):
 | |
|         # Put this here to allow entry-docs and help to work without libfdt
 | |
|         global state
 | |
|         from binman import state
 | |
| 
 | |
|         self.section = section
 | |
|         self.etype = etype
 | |
|         self._node = node
 | |
|         self.name = node and (name_prefix + node.name) or 'none'
 | |
|         self.offset = None
 | |
|         self.size = None
 | |
|         self.pre_reset_size = None
 | |
|         self.uncomp_size = None
 | |
|         self.data = None
 | |
|         self.uncomp_data = None
 | |
|         self.contents_size = 0
 | |
|         self.align = None
 | |
|         self.align_size = None
 | |
|         self.align_end = None
 | |
|         self.pad_before = 0
 | |
|         self.pad_after = 0
 | |
|         self.offset_unset = False
 | |
|         self.image_pos = None
 | |
|         self.extend_size = False
 | |
|         self.compress = 'none'
 | |
|         self.missing = False
 | |
|         self.faked = False
 | |
|         self.external = False
 | |
|         self.allow_missing = False
 | |
|         self.allow_fake = False
 | |
|         self.bintools = {}
 | |
|         self.missing_bintools = []
 | |
|         self.update_hash = True
 | |
|         self.fake_fname = None
 | |
|         self.required_props = []
 | |
|         self.comp_bintool = None
 | |
| 
 | |
|     @staticmethod
 | |
|     def FindEntryClass(etype, expanded):
 | |
|         """Look up the entry class for a node.
 | |
| 
 | |
|         Args:
 | |
|             node_node: Path name of Node object containing information about
 | |
|                        the entry to create (used for errors)
 | |
|             etype:   Entry type to use
 | |
|             expanded: Use the expanded version of etype
 | |
| 
 | |
|         Returns:
 | |
|             The entry class object if found, else None if not found and expanded
 | |
|                 is True, else a tuple:
 | |
|                     module name that could not be found
 | |
|                     exception received
 | |
|         """
 | |
|         # Convert something like 'u-boot@0' to 'u_boot' since we are only
 | |
|         # interested in the type.
 | |
|         module_name = etype.replace('-', '_')
 | |
| 
 | |
|         if '@' in module_name:
 | |
|             module_name = module_name.split('@')[0]
 | |
|         if expanded:
 | |
|             module_name += '_expanded'
 | |
|         module = modules.get(module_name)
 | |
| 
 | |
|         # Also allow entry-type modules to be brought in from the etype directory.
 | |
| 
 | |
|         # Import the module if we have not already done so.
 | |
|         if not module:
 | |
|             try:
 | |
|                 module = importlib.import_module('binman.etype.' + module_name)
 | |
|             except ImportError as e:
 | |
|                 if expanded:
 | |
|                     return None
 | |
|                 return module_name, e
 | |
|             modules[module_name] = module
 | |
| 
 | |
|         # Look up the expected class name
 | |
|         return getattr(module, 'Entry_%s' % module_name)
 | |
| 
 | |
|     @staticmethod
 | |
|     def Lookup(node_path, etype, expanded, missing_etype=False):
 | |
|         """Look up the entry class for a node.
 | |
| 
 | |
|         Args:
 | |
|             node_node (str): Path name of Node object containing information
 | |
|                 about the entry to create (used for errors)
 | |
|             etype (str):   Entry type to use
 | |
|             expanded (bool): Use the expanded version of etype
 | |
|             missing_etype (bool): True to default to a blob etype if the
 | |
|                 requested etype is not found
 | |
| 
 | |
|         Returns:
 | |
|             The entry class object if found, else None if not found and expanded
 | |
|                 is True
 | |
| 
 | |
|         Raise:
 | |
|             ValueError if expanded is False and the class is not found
 | |
|         """
 | |
|         # Convert something like 'u-boot@0' to 'u_boot' since we are only
 | |
|         # interested in the type.
 | |
|         cls = Entry.FindEntryClass(etype, expanded)
 | |
|         if cls is None:
 | |
|             return None
 | |
|         elif isinstance(cls, tuple):
 | |
|             if missing_etype:
 | |
|                 cls = Entry.FindEntryClass('blob', False)
 | |
|             if isinstance(cls, tuple): # This should not fail
 | |
|                 module_name, e = cls
 | |
|                 raise ValueError(
 | |
|                     "Unknown entry type '%s' in node '%s' (expected etype/%s.py, error '%s'" %
 | |
|                     (etype, node_path, module_name, e))
 | |
|         return cls
 | |
| 
 | |
|     @staticmethod
 | |
|     def Create(section, node, etype=None, expanded=False, missing_etype=False):
 | |
|         """Create a new entry for a node.
 | |
| 
 | |
|         Args:
 | |
|             section (entry_Section):  Section object containing this node
 | |
|             node (Node): Node object containing information about the entry to
 | |
|                 create
 | |
|             etype (str): Entry type to use, or None to work it out (used for
 | |
|                 tests)
 | |
|             expanded (bool): Use the expanded version of etype
 | |
|             missing_etype (bool): True to default to a blob etype if the
 | |
|                 requested etype is not found
 | |
| 
 | |
|         Returns:
 | |
|             A new Entry object of the correct type (a subclass of Entry)
 | |
|         """
 | |
|         if not etype:
 | |
|             etype = fdt_util.GetString(node, 'type', node.name)
 | |
|         obj = Entry.Lookup(node.path, etype, expanded, missing_etype)
 | |
|         if obj and expanded:
 | |
|             # Check whether to use the expanded entry
 | |
|             new_etype = etype + '-expanded'
 | |
|             can_expand = not fdt_util.GetBool(node, 'no-expanded')
 | |
|             if can_expand and obj.UseExpanded(node, etype, new_etype):
 | |
|                 etype = new_etype
 | |
|             else:
 | |
|                 obj = None
 | |
|         if not obj:
 | |
|             obj = Entry.Lookup(node.path, etype, False, missing_etype)
 | |
| 
 | |
|         # Call its constructor to get the object we want.
 | |
|         return obj(section, etype, node)
 | |
| 
 | |
|     def ReadNode(self):
 | |
|         """Read entry information from the node
 | |
| 
 | |
|         This must be called as the first thing after the Entry is created.
 | |
| 
 | |
|         This reads all the fields we recognise from the node, ready for use.
 | |
|         """
 | |
|         self.ensure_props()
 | |
|         if 'pos' in self._node.props:
 | |
|             self.Raise("Please use 'offset' instead of 'pos'")
 | |
|         if 'expand-size' in self._node.props:
 | |
|             self.Raise("Please use 'extend-size' instead of 'expand-size'")
 | |
|         self.offset = fdt_util.GetInt(self._node, 'offset')
 | |
|         self.size = fdt_util.GetInt(self._node, 'size')
 | |
|         self.orig_offset = fdt_util.GetInt(self._node, 'orig-offset')
 | |
|         self.orig_size = fdt_util.GetInt(self._node, 'orig-size')
 | |
|         if self.GetImage().copy_to_orig:
 | |
|             self.orig_offset = self.offset
 | |
|             self.orig_size = self.size
 | |
| 
 | |
|         # These should not be set in input files, but are set in an FDT map,
 | |
|         # which is also read by this code.
 | |
|         self.image_pos = fdt_util.GetInt(self._node, 'image-pos')
 | |
|         self.uncomp_size = fdt_util.GetInt(self._node, 'uncomp-size')
 | |
| 
 | |
|         self.align = fdt_util.GetInt(self._node, 'align')
 | |
|         if tools.not_power_of_two(self.align):
 | |
|             raise ValueError("Node '%s': Alignment %s must be a power of two" %
 | |
|                              (self._node.path, self.align))
 | |
|         if self.section and self.align is None:
 | |
|             self.align = self.section.align_default
 | |
|         self.pad_before = fdt_util.GetInt(self._node, 'pad-before', 0)
 | |
|         self.pad_after = fdt_util.GetInt(self._node, 'pad-after', 0)
 | |
|         self.align_size = fdt_util.GetInt(self._node, 'align-size')
 | |
|         if tools.not_power_of_two(self.align_size):
 | |
|             self.Raise("Alignment size %s must be a power of two" %
 | |
|                        self.align_size)
 | |
|         self.align_end = fdt_util.GetInt(self._node, 'align-end')
 | |
|         self.offset_unset = fdt_util.GetBool(self._node, 'offset-unset')
 | |
|         self.extend_size = fdt_util.GetBool(self._node, 'extend-size')
 | |
|         self.missing_msg = fdt_util.GetString(self._node, 'missing-msg')
 | |
| 
 | |
|         # This is only supported by blobs and sections at present
 | |
|         self.compress = fdt_util.GetString(self._node, 'compress', 'none')
 | |
| 
 | |
|     def GetDefaultFilename(self):
 | |
|         return None
 | |
| 
 | |
|     def GetFdts(self):
 | |
|         """Get the device trees used by this entry
 | |
| 
 | |
|         Returns:
 | |
|             Empty dict, if this entry is not a .dtb, otherwise:
 | |
|             Dict:
 | |
|                 key: Filename from this entry (without the path)
 | |
|                 value: Tuple:
 | |
|                     Entry object for this dtb
 | |
|                     Filename of file containing this dtb
 | |
|         """
 | |
|         return {}
 | |
| 
 | |
|     def gen_entries(self):
 | |
|         """Allow entries to generate other entries
 | |
| 
 | |
|         Some entries generate subnodes automatically, from which sub-entries
 | |
|         are then created. This method allows those to be added to the binman
 | |
|         definition for the current image. An entry which implements this method
 | |
|         should call state.AddSubnode() to add a subnode and can add properties
 | |
|         with state.AddString(), etc.
 | |
| 
 | |
|         An example is 'files', which produces a section containing a list of
 | |
|         files.
 | |
|         """
 | |
|         pass
 | |
| 
 | |
|     def AddMissingProperties(self, have_image_pos):
 | |
|         """Add new properties to the device tree as needed for this entry
 | |
| 
 | |
|         Args:
 | |
|             have_image_pos: True if this entry has an image position. This can
 | |
|                 be False if its parent section is compressed, since compression
 | |
|                 groups all entries together into a compressed block of data,
 | |
|                 obscuring the start of each individual child entry
 | |
|         """
 | |
|         for prop in ['offset', 'size']:
 | |
|             if not prop in self._node.props:
 | |
|                 state.AddZeroProp(self._node, prop)
 | |
|         if have_image_pos and 'image-pos' not in self._node.props:
 | |
|             state.AddZeroProp(self._node, 'image-pos')
 | |
|         if self.GetImage().allow_repack:
 | |
|             if self.orig_offset is not None:
 | |
|                 state.AddZeroProp(self._node, 'orig-offset', True)
 | |
|             if self.orig_size is not None:
 | |
|                 state.AddZeroProp(self._node, 'orig-size', True)
 | |
| 
 | |
|         if self.compress != 'none':
 | |
|             state.AddZeroProp(self._node, 'uncomp-size')
 | |
| 
 | |
|         if self.update_hash:
 | |
|             err = state.CheckAddHashProp(self._node)
 | |
|             if err:
 | |
|                 self.Raise(err)
 | |
| 
 | |
|     def SetCalculatedProperties(self):
 | |
|         """Set the value of device-tree properties calculated by binman"""
 | |
|         state.SetInt(self._node, 'offset', self.offset)
 | |
|         state.SetInt(self._node, 'size', self.size)
 | |
|         base = self.section.GetRootSkipAtStart() if self.section else 0
 | |
|         if self.image_pos is not None:
 | |
|             state.SetInt(self._node, 'image-pos', self.image_pos - base)
 | |
|         if self.GetImage().allow_repack:
 | |
|             if self.orig_offset is not None:
 | |
|                 state.SetInt(self._node, 'orig-offset', self.orig_offset, True)
 | |
|             if self.orig_size is not None:
 | |
|                 state.SetInt(self._node, 'orig-size', self.orig_size, True)
 | |
|         if self.uncomp_size is not None:
 | |
|             state.SetInt(self._node, 'uncomp-size', self.uncomp_size)
 | |
| 
 | |
|         if self.update_hash:
 | |
|             state.CheckSetHashValue(self._node, self.GetData)
 | |
| 
 | |
|     def ProcessFdt(self, fdt):
 | |
|         """Allow entries to adjust the device tree
 | |
| 
 | |
|         Some entries need to adjust the device tree for their purposes. This
 | |
|         may involve adding or deleting properties.
 | |
| 
 | |
|         Returns:
 | |
|             True if processing is complete
 | |
|             False if processing could not be completed due to a dependency.
 | |
|                 This will cause the entry to be retried after others have been
 | |
|                 called
 | |
|         """
 | |
|         return True
 | |
| 
 | |
|     def SetPrefix(self, prefix):
 | |
|         """Set the name prefix for a node
 | |
| 
 | |
|         Args:
 | |
|             prefix: Prefix to set, or '' to not use a prefix
 | |
|         """
 | |
|         if prefix:
 | |
|             self.name = prefix + self.name
 | |
| 
 | |
|     def SetContents(self, data):
 | |
|         """Set the contents of an entry
 | |
| 
 | |
|         This sets both the data and content_size properties
 | |
| 
 | |
|         Args:
 | |
|             data: Data to set to the contents (bytes)
 | |
|         """
 | |
|         self.data = data
 | |
|         self.contents_size = len(self.data)
 | |
| 
 | |
|     def ProcessContentsUpdate(self, data):
 | |
|         """Update the contents of an entry, after the size is fixed
 | |
| 
 | |
|         This checks that the new data is the same size as the old. If the size
 | |
|         has changed, this triggers a re-run of the packing algorithm.
 | |
| 
 | |
|         Args:
 | |
|             data: Data to set to the contents (bytes)
 | |
| 
 | |
|         Raises:
 | |
|             ValueError if the new data size is not the same as the old
 | |
|         """
 | |
|         size_ok = True
 | |
|         new_size = len(data)
 | |
|         if state.AllowEntryExpansion() and new_size > self.contents_size:
 | |
|             # self.data will indicate the new size needed
 | |
|             size_ok = False
 | |
|         elif state.AllowEntryContraction() and new_size < self.contents_size:
 | |
|             size_ok = False
 | |
| 
 | |
|         # If not allowed to change, try to deal with it or give up
 | |
|         if size_ok:
 | |
|             if new_size > self.contents_size:
 | |
|                 self.Raise('Cannot update entry size from %d to %d' %
 | |
|                         (self.contents_size, new_size))
 | |
| 
 | |
|             # Don't let the data shrink. Pad it if necessary
 | |
|             if size_ok and new_size < self.contents_size:
 | |
|                 data += tools.get_bytes(0, self.contents_size - new_size)
 | |
| 
 | |
|         if not size_ok:
 | |
|             tout.debug("Entry '%s' size change from %s to %s" % (
 | |
|                 self._node.path, to_hex(self.contents_size),
 | |
|                 to_hex(new_size)))
 | |
|         self.SetContents(data)
 | |
|         return size_ok
 | |
| 
 | |
|     def ObtainContents(self, skip_entry=None, fake_size=0):
 | |
|         """Figure out the contents of an entry.
 | |
| 
 | |
|         Args:
 | |
|             skip_entry (Entry): Entry to skip when obtaining section contents
 | |
|             fake_size (int): Size of fake file to create if needed
 | |
| 
 | |
|         Returns:
 | |
|             True if the contents were found, False if another call is needed
 | |
|             after the other entries are processed.
 | |
|         """
 | |
|         # No contents by default: subclasses can implement this
 | |
|         return True
 | |
| 
 | |
|     def ResetForPack(self):
 | |
|         """Reset offset/size fields so that packing can be done again"""
 | |
|         self.Detail('ResetForPack: offset %s->%s, size %s->%s' %
 | |
|                     (to_hex(self.offset), to_hex(self.orig_offset),
 | |
|                      to_hex(self.size), to_hex(self.orig_size)))
 | |
|         self.pre_reset_size = self.size
 | |
|         self.offset = self.orig_offset
 | |
|         self.size = self.orig_size
 | |
| 
 | |
|     def Pack(self, offset):
 | |
|         """Figure out how to pack the entry into the section
 | |
| 
 | |
|         Most of the time the entries are not fully specified. There may be
 | |
|         an alignment but no size. In that case we take the size from the
 | |
|         contents of the entry.
 | |
| 
 | |
|         If an entry has no hard-coded offset, it will be placed at @offset.
 | |
| 
 | |
|         Once this function is complete, both the offset and size of the
 | |
|         entry will be know.
 | |
| 
 | |
|         Args:
 | |
|             Current section offset pointer
 | |
| 
 | |
|         Returns:
 | |
|             New section offset pointer (after this entry)
 | |
|         """
 | |
|         self.Detail('Packing: offset=%s, size=%s, content_size=%x' %
 | |
|                     (to_hex(self.offset), to_hex(self.size),
 | |
|                      self.contents_size))
 | |
|         if self.offset is None:
 | |
|             if self.offset_unset:
 | |
|                 self.Raise('No offset set with offset-unset: should another '
 | |
|                            'entry provide this correct offset?')
 | |
|             self.offset = tools.align(offset, self.align)
 | |
|         needed = self.pad_before + self.contents_size + self.pad_after
 | |
|         needed = tools.align(needed, self.align_size)
 | |
|         size = self.size
 | |
|         if not size:
 | |
|             size = needed
 | |
|         new_offset = self.offset + size
 | |
|         aligned_offset = tools.align(new_offset, self.align_end)
 | |
|         if aligned_offset != new_offset:
 | |
|             size = aligned_offset - self.offset
 | |
|             new_offset = aligned_offset
 | |
| 
 | |
|         if not self.size:
 | |
|             self.size = size
 | |
| 
 | |
|         if self.size < needed:
 | |
|             self.Raise("Entry contents size is %#x (%d) but entry size is "
 | |
|                        "%#x (%d)" % (needed, needed, self.size, self.size))
 | |
|         # Check that the alignment is correct. It could be wrong if the
 | |
|         # and offset or size values were provided (i.e. not calculated), but
 | |
|         # conflict with the provided alignment values
 | |
|         if self.size != tools.align(self.size, self.align_size):
 | |
|             self.Raise("Size %#x (%d) does not match align-size %#x (%d)" %
 | |
|                   (self.size, self.size, self.align_size, self.align_size))
 | |
|         if self.offset != tools.align(self.offset, self.align):
 | |
|             self.Raise("Offset %#x (%d) does not match align %#x (%d)" %
 | |
|                   (self.offset, self.offset, self.align, self.align))
 | |
|         self.Detail('   - packed: offset=%#x, size=%#x, content_size=%#x, next_offset=%x' %
 | |
|                     (self.offset, self.size, self.contents_size, new_offset))
 | |
| 
 | |
|         return new_offset
 | |
| 
 | |
|     def Raise(self, msg):
 | |
|         """Convenience function to raise an error referencing a node"""
 | |
|         raise ValueError("Node '%s': %s" % (self._node.path, msg))
 | |
| 
 | |
|     def Info(self, msg):
 | |
|         """Convenience function to log info referencing a node"""
 | |
|         tag = "Info '%s'" % self._node.path
 | |
|         tout.detail('%30s: %s' % (tag, msg))
 | |
| 
 | |
|     def Detail(self, msg):
 | |
|         """Convenience function to log detail referencing a node"""
 | |
|         tag = "Node '%s'" % self._node.path
 | |
|         tout.detail('%30s: %s' % (tag, msg))
 | |
| 
 | |
|     def GetEntryArgsOrProps(self, props, required=False):
 | |
|         """Return the values of a set of properties
 | |
| 
 | |
|         Args:
 | |
|             props: List of EntryArg objects
 | |
| 
 | |
|         Raises:
 | |
|             ValueError if a property is not found
 | |
|         """
 | |
|         values = []
 | |
|         missing = []
 | |
|         for prop in props:
 | |
|             python_prop = prop.name.replace('-', '_')
 | |
|             if hasattr(self, python_prop):
 | |
|                 value = getattr(self, python_prop)
 | |
|             else:
 | |
|                 value = None
 | |
|             if value is None:
 | |
|                 value = self.GetArg(prop.name, prop.datatype)
 | |
|             if value is None and required:
 | |
|                 missing.append(prop.name)
 | |
|             values.append(value)
 | |
|         if missing:
 | |
|             self.GetImage().MissingArgs(self, missing)
 | |
|         return values
 | |
| 
 | |
|     def GetPath(self):
 | |
|         """Get the path of a node
 | |
| 
 | |
|         Returns:
 | |
|             Full path of the node for this entry
 | |
|         """
 | |
|         return self._node.path
 | |
| 
 | |
|     def GetData(self, required=True):
 | |
|         """Get the contents of an entry
 | |
| 
 | |
|         Args:
 | |
|             required: True if the data must be present, False if it is OK to
 | |
|                 return None
 | |
| 
 | |
|         Returns:
 | |
|             bytes content of the entry, excluding any padding. If the entry is
 | |
|                 compressed, the compressed data is returned
 | |
|         """
 | |
|         self.Detail('GetData: size %s' % to_hex_size(self.data))
 | |
|         return self.data
 | |
| 
 | |
|     def GetPaddedData(self, data=None):
 | |
|         """Get the data for an entry including any padding
 | |
| 
 | |
|         Gets the entry data and uses its section's pad-byte value to add padding
 | |
|         before and after as defined by the pad-before and pad-after properties.
 | |
| 
 | |
|         This does not consider alignment.
 | |
| 
 | |
|         Returns:
 | |
|             Contents of the entry along with any pad bytes before and
 | |
|             after it (bytes)
 | |
|         """
 | |
|         if data is None:
 | |
|             data = self.GetData()
 | |
|         return self.section.GetPaddedDataForEntry(self, data)
 | |
| 
 | |
|     def GetOffsets(self):
 | |
|         """Get the offsets for siblings
 | |
| 
 | |
|         Some entry types can contain information about the position or size of
 | |
|         other entries. An example of this is the Intel Flash Descriptor, which
 | |
|         knows where the Intel Management Engine section should go.
 | |
| 
 | |
|         If this entry knows about the position of other entries, it can specify
 | |
|         this by returning values here
 | |
| 
 | |
|         Returns:
 | |
|             Dict:
 | |
|                 key: Entry type
 | |
|                 value: List containing position and size of the given entry
 | |
|                     type. Either can be None if not known
 | |
|         """
 | |
|         return {}
 | |
| 
 | |
|     def SetOffsetSize(self, offset, size):
 | |
|         """Set the offset and/or size of an entry
 | |
| 
 | |
|         Args:
 | |
|             offset: New offset, or None to leave alone
 | |
|             size: New size, or None to leave alone
 | |
|         """
 | |
|         if offset is not None:
 | |
|             self.offset = offset
 | |
|         if size is not None:
 | |
|             self.size = size
 | |
| 
 | |
|     def SetImagePos(self, image_pos):
 | |
|         """Set the position in the image
 | |
| 
 | |
|         Args:
 | |
|             image_pos: Position of this entry in the image
 | |
|         """
 | |
|         self.image_pos = image_pos + self.offset
 | |
| 
 | |
|     def ProcessContents(self):
 | |
|         """Do any post-packing updates of entry contents
 | |
| 
 | |
|         This function should call ProcessContentsUpdate() to update the entry
 | |
|         contents, if necessary, returning its return value here.
 | |
| 
 | |
|         Args:
 | |
|             data: Data to set to the contents (bytes)
 | |
| 
 | |
|         Returns:
 | |
|             True if the new data size is OK, False if expansion is needed
 | |
| 
 | |
|         Raises:
 | |
|             ValueError if the new data size is not the same as the old and
 | |
|                 state.AllowEntryExpansion() is False
 | |
|         """
 | |
|         return True
 | |
| 
 | |
|     def WriteSymbols(self, section):
 | |
|         """Write symbol values into binary files for access at run time
 | |
| 
 | |
|         Args:
 | |
|           section: Section containing the entry
 | |
|         """
 | |
|         pass
 | |
| 
 | |
|     def CheckEntries(self):
 | |
|         """Check that the entry offsets are correct
 | |
| 
 | |
|         This is used for entries which have extra offset requirements (other
 | |
|         than having to be fully inside their section). Sub-classes can implement
 | |
|         this function and raise if there is a problem.
 | |
|         """
 | |
|         pass
 | |
| 
 | |
|     @staticmethod
 | |
|     def GetStr(value):
 | |
|         if value is None:
 | |
|             return '<none>  '
 | |
|         return '%08x' % value
 | |
| 
 | |
|     @staticmethod
 | |
|     def WriteMapLine(fd, indent, name, offset, size, image_pos):
 | |
|         print('%s  %s%s  %s  %s' % (Entry.GetStr(image_pos), ' ' * indent,
 | |
|                                     Entry.GetStr(offset), Entry.GetStr(size),
 | |
|                                     name), file=fd)
 | |
| 
 | |
|     def WriteMap(self, fd, indent):
 | |
|         """Write a map of the entry to a .map file
 | |
| 
 | |
|         Args:
 | |
|             fd: File to write the map to
 | |
|             indent: Curent indent level of map (0=none, 1=one level, etc.)
 | |
|         """
 | |
|         self.WriteMapLine(fd, indent, self.name, self.offset, self.size,
 | |
|                           self.image_pos)
 | |
| 
 | |
|     # pylint: disable=assignment-from-none
 | |
|     def GetEntries(self):
 | |
|         """Return a list of entries contained by this entry
 | |
| 
 | |
|         Returns:
 | |
|             List of entries, or None if none. A normal entry has no entries
 | |
|                 within it so will return None
 | |
|         """
 | |
|         return None
 | |
| 
 | |
|     def FindEntryByNode(self, find_node):
 | |
|         """Find a node in an entry, searching all subentries
 | |
| 
 | |
|         This does a recursive search.
 | |
| 
 | |
|         Args:
 | |
|             find_node (fdt.Node): Node to find
 | |
| 
 | |
|         Returns:
 | |
|             Entry: entry, if found, else None
 | |
|         """
 | |
|         entries = self.GetEntries()
 | |
|         if entries:
 | |
|             for entry in entries.values():
 | |
|                 if entry._node == find_node:
 | |
|                     return entry
 | |
|                 found = entry.FindEntryByNode(find_node)
 | |
|                 if found:
 | |
|                     return found
 | |
| 
 | |
|         return None
 | |
| 
 | |
|     def GetArg(self, name, datatype=str):
 | |
|         """Get the value of an entry argument or device-tree-node property
 | |
| 
 | |
|         Some node properties can be provided as arguments to binman. First check
 | |
|         the entry arguments, and fall back to the device tree if not found
 | |
| 
 | |
|         Args:
 | |
|             name: Argument name
 | |
|             datatype: Data type (str or int)
 | |
| 
 | |
|         Returns:
 | |
|             Value of argument as a string or int, or None if no value
 | |
| 
 | |
|         Raises:
 | |
|             ValueError if the argument cannot be converted to in
 | |
|         """
 | |
|         value = state.GetEntryArg(name)
 | |
|         if value is not None:
 | |
|             if datatype == int:
 | |
|                 try:
 | |
|                     value = int(value)
 | |
|                 except ValueError:
 | |
|                     self.Raise("Cannot convert entry arg '%s' (value '%s') to integer" %
 | |
|                                (name, value))
 | |
|             elif datatype == str:
 | |
|                 pass
 | |
|             else:
 | |
|                 raise ValueError("GetArg() internal error: Unknown data type '%s'" %
 | |
|                                  datatype)
 | |
|         else:
 | |
|             value = fdt_util.GetDatatype(self._node, name, datatype)
 | |
|         return value
 | |
| 
 | |
|     @staticmethod
 | |
|     def WriteDocs(modules, test_missing=None):
 | |
|         """Write out documentation about the various entry types to stdout
 | |
| 
 | |
|         Args:
 | |
|             modules: List of modules to include
 | |
|             test_missing: Used for testing. This is a module to report
 | |
|                 as missing
 | |
|         """
 | |
|         print('''Binman Entry Documentation
 | |
| ===========================
 | |
| 
 | |
| This file describes the entry types supported by binman. These entry types can
 | |
| be placed in an image one by one to build up a final firmware image. It is
 | |
| fairly easy to create new entry types. Just add a new file to the 'etype'
 | |
| directory. You can use the existing entries as examples.
 | |
| 
 | |
| Note that some entries are subclasses of others, using and extending their
 | |
| features to produce new behaviours.
 | |
| 
 | |
| 
 | |
| ''')
 | |
|         modules = sorted(modules)
 | |
| 
 | |
|         # Don't show the test entry
 | |
|         if '_testing' in modules:
 | |
|             modules.remove('_testing')
 | |
|         missing = []
 | |
|         for name in modules:
 | |
|             module = Entry.Lookup('WriteDocs', name, False)
 | |
|             docs = getattr(module, '__doc__')
 | |
|             if test_missing == name:
 | |
|                 docs = None
 | |
|             if docs:
 | |
|                 lines = docs.splitlines()
 | |
|                 first_line = lines[0]
 | |
|                 rest = [line[4:] for line in lines[1:]]
 | |
|                 hdr = 'Entry: %s: %s' % (name.replace('_', '-'), first_line)
 | |
| 
 | |
|                 # Create a reference for use by rST docs
 | |
|                 ref_name = f'etype_{module.__name__[6:]}'.lower()
 | |
|                 print('.. _%s:' % ref_name)
 | |
|                 print()
 | |
|                 print(hdr)
 | |
|                 print('-' * len(hdr))
 | |
|                 print('\n'.join(rest))
 | |
|                 print()
 | |
|                 print()
 | |
|             else:
 | |
|                 missing.append(name)
 | |
| 
 | |
|         if missing:
 | |
|             raise ValueError('Documentation is missing for modules: %s' %
 | |
|                              ', '.join(missing))
 | |
| 
 | |
|     def GetUniqueName(self):
 | |
|         """Get a unique name for a node
 | |
| 
 | |
|         Returns:
 | |
|             String containing a unique name for a node, consisting of the name
 | |
|             of all ancestors (starting from within the 'binman' node) separated
 | |
|             by a dot ('.'). This can be useful for generating unique filesnames
 | |
|             in the output directory.
 | |
|         """
 | |
|         name = self.name
 | |
|         node = self._node
 | |
|         while node.parent:
 | |
|             node = node.parent
 | |
|             if node.name in ('binman', '/'):
 | |
|                 break
 | |
|             name = '%s.%s' % (node.name, name)
 | |
|         return name
 | |
| 
 | |
|     def extend_to_limit(self, limit):
 | |
|         """Extend an entry so that it ends at the given offset limit"""
 | |
|         if self.offset + self.size < limit:
 | |
|             self.size = limit - self.offset
 | |
|             # Request the contents again, since changing the size requires that
 | |
|             # the data grows. This should not fail, but check it to be sure.
 | |
|             if not self.ObtainContents():
 | |
|                 self.Raise('Cannot obtain contents when expanding entry')
 | |
| 
 | |
|     def HasSibling(self, name):
 | |
|         """Check if there is a sibling of a given name
 | |
| 
 | |
|         Returns:
 | |
|             True if there is an entry with this name in the the same section,
 | |
|                 else False
 | |
|         """
 | |
|         return name in self.section.GetEntries()
 | |
| 
 | |
|     def GetSiblingImagePos(self, name):
 | |
|         """Return the image position of the given sibling
 | |
| 
 | |
|         Returns:
 | |
|             Image position of sibling, or None if the sibling has no position,
 | |
|                 or False if there is no such sibling
 | |
|         """
 | |
|         if not self.HasSibling(name):
 | |
|             return False
 | |
|         return self.section.GetEntries()[name].image_pos
 | |
| 
 | |
|     @staticmethod
 | |
|     def AddEntryInfo(entries, indent, name, etype, size, image_pos,
 | |
|                      uncomp_size, offset, entry):
 | |
|         """Add a new entry to the entries list
 | |
| 
 | |
|         Args:
 | |
|             entries: List (of EntryInfo objects) to add to
 | |
|             indent: Current indent level to add to list
 | |
|             name: Entry name (string)
 | |
|             etype: Entry type (string)
 | |
|             size: Entry size in bytes (int)
 | |
|             image_pos: Position within image in bytes (int)
 | |
|             uncomp_size: Uncompressed size if the entry uses compression, else
 | |
|                 None
 | |
|             offset: Entry offset within parent in bytes (int)
 | |
|             entry: Entry object
 | |
|         """
 | |
|         entries.append(EntryInfo(indent, name, etype, size, image_pos,
 | |
|                                  uncomp_size, offset, entry))
 | |
| 
 | |
|     def ListEntries(self, entries, indent):
 | |
|         """Add files in this entry to the list of entries
 | |
| 
 | |
|         This can be overridden by subclasses which need different behaviour.
 | |
| 
 | |
|         Args:
 | |
|             entries: List (of EntryInfo objects) to add to
 | |
|             indent: Current indent level to add to list
 | |
|         """
 | |
|         self.AddEntryInfo(entries, indent, self.name, self.etype, self.size,
 | |
|                           self.image_pos, self.uncomp_size, self.offset, self)
 | |
| 
 | |
|     def ReadData(self, decomp=True, alt_format=None):
 | |
|         """Read the data for an entry from the image
 | |
| 
 | |
|         This is used when the image has been read in and we want to extract the
 | |
|         data for a particular entry from that image.
 | |
| 
 | |
|         Args:
 | |
|             decomp: True to decompress any compressed data before returning it;
 | |
|                 False to return the raw, uncompressed data
 | |
| 
 | |
|         Returns:
 | |
|             Entry data (bytes)
 | |
|         """
 | |
|         # Use True here so that we get an uncompressed section to work from,
 | |
|         # although compressed sections are currently not supported
 | |
|         tout.debug("ReadChildData section '%s', entry '%s'" %
 | |
|                    (self.section.GetPath(), self.GetPath()))
 | |
|         data = self.section.ReadChildData(self, decomp, alt_format)
 | |
|         return data
 | |
| 
 | |
|     def ReadChildData(self, child, decomp=True, alt_format=None):
 | |
|         """Read the data for a particular child entry
 | |
| 
 | |
|         This reads data from the parent and extracts the piece that relates to
 | |
|         the given child.
 | |
| 
 | |
|         Args:
 | |
|             child (Entry): Child entry to read data for (must be valid)
 | |
|             decomp (bool): True to decompress any compressed data before
 | |
|                 returning it; False to return the raw, uncompressed data
 | |
|             alt_format (str): Alternative format to read in, or None
 | |
| 
 | |
|         Returns:
 | |
|             Data for the child (bytes)
 | |
|         """
 | |
|         pass
 | |
| 
 | |
|     def LoadData(self, decomp=True):
 | |
|         data = self.ReadData(decomp)
 | |
|         self.contents_size = len(data)
 | |
|         self.ProcessContentsUpdate(data)
 | |
|         self.Detail('Loaded data size %x' % len(data))
 | |
| 
 | |
|     def GetAltFormat(self, data, alt_format):
 | |
|         """Read the data for an extry in an alternative format
 | |
| 
 | |
|         Supported formats are list in the documentation for each entry. An
 | |
|         example is fdtmap which provides .
 | |
| 
 | |
|         Args:
 | |
|             data (bytes): Data to convert (this should have been produced by the
 | |
|                 entry)
 | |
|             alt_format (str): Format to use
 | |
| 
 | |
|         """
 | |
|         pass
 | |
| 
 | |
|     def GetImage(self):
 | |
|         """Get the image containing this entry
 | |
| 
 | |
|         Returns:
 | |
|             Image object containing this entry
 | |
|         """
 | |
|         return self.section.GetImage()
 | |
| 
 | |
|     def WriteData(self, data, decomp=True):
 | |
|         """Write the data to an entry in the image
 | |
| 
 | |
|         This is used when the image has been read in and we want to replace the
 | |
|         data for a particular entry in that image.
 | |
| 
 | |
|         The image must be re-packed and written out afterwards.
 | |
| 
 | |
|         Args:
 | |
|             data: Data to replace it with
 | |
|             decomp: True to compress the data if needed, False if data is
 | |
|                 already compressed so should be used as is
 | |
| 
 | |
|         Returns:
 | |
|             True if the data did not result in a resize of this entry, False if
 | |
|                  the entry must be resized
 | |
|         """
 | |
|         if self.size is not None:
 | |
|             self.contents_size = self.size
 | |
|         else:
 | |
|             self.contents_size = self.pre_reset_size
 | |
|         ok = self.ProcessContentsUpdate(data)
 | |
|         self.Detail('WriteData: size=%x, ok=%s' % (len(data), ok))
 | |
|         section_ok = self.section.WriteChildData(self)
 | |
|         return ok and section_ok
 | |
| 
 | |
|     def WriteChildData(self, child):
 | |
|         """Handle writing the data in a child entry
 | |
| 
 | |
|         This should be called on the child's parent section after the child's
 | |
|         data has been updated. It should update any data structures needed to
 | |
|         validate that the update is successful.
 | |
| 
 | |
|         This base-class implementation does nothing, since the base Entry object
 | |
|         does not have any children.
 | |
| 
 | |
|         Args:
 | |
|             child: Child Entry that was written
 | |
| 
 | |
|         Returns:
 | |
|             True if the section could be updated successfully, False if the
 | |
|                 data is such that the section could not update
 | |
|         """
 | |
|         return True
 | |
| 
 | |
|     def GetSiblingOrder(self):
 | |
|         """Get the relative order of an entry amoung its siblings
 | |
| 
 | |
|         Returns:
 | |
|             'start' if this entry is first among siblings, 'end' if last,
 | |
|                 otherwise None
 | |
|         """
 | |
|         entries = list(self.section.GetEntries().values())
 | |
|         if entries:
 | |
|             if self == entries[0]:
 | |
|                 return 'start'
 | |
|             elif self == entries[-1]:
 | |
|                 return 'end'
 | |
|         return 'middle'
 | |
| 
 | |
|     def SetAllowMissing(self, allow_missing):
 | |
|         """Set whether a section allows missing external blobs
 | |
| 
 | |
|         Args:
 | |
|             allow_missing: True if allowed, False if not allowed
 | |
|         """
 | |
|         # This is meaningless for anything other than sections
 | |
|         pass
 | |
| 
 | |
|     def SetAllowFakeBlob(self, allow_fake):
 | |
|         """Set whether a section allows to create a fake blob
 | |
| 
 | |
|         Args:
 | |
|             allow_fake: True if allowed, False if not allowed
 | |
|         """
 | |
|         self.allow_fake = allow_fake
 | |
| 
 | |
|     def CheckMissing(self, missing_list):
 | |
|         """Check if any entries in this section have missing external blobs
 | |
| 
 | |
|         If there are missing blobs, the entries are added to the list
 | |
| 
 | |
|         Args:
 | |
|             missing_list: List of Entry objects to be added to
 | |
|         """
 | |
|         if self.missing:
 | |
|             missing_list.append(self)
 | |
| 
 | |
|     def check_fake_fname(self, fname, size=0):
 | |
|         """If the file is missing and the entry allows fake blobs, fake it
 | |
| 
 | |
|         Sets self.faked to True if faked
 | |
| 
 | |
|         Args:
 | |
|             fname (str): Filename to check
 | |
|             size (int): Size of fake file to create
 | |
| 
 | |
|         Returns:
 | |
|             tuple:
 | |
|                 fname (str): Filename of faked file
 | |
|                 bool: True if the blob was faked, False if not
 | |
|         """
 | |
|         if self.allow_fake and not pathlib.Path(fname).is_file():
 | |
|             if not self.fake_fname:
 | |
|                 outfname = os.path.join(self.fake_dir, os.path.basename(fname))
 | |
|                 with open(outfname, "wb") as out:
 | |
|                     out.truncate(size)
 | |
|                 tout.info(f"Entry '{self._node.path}': Faked blob '{outfname}'")
 | |
|                 self.fake_fname = outfname
 | |
|             self.faked = True
 | |
|             return self.fake_fname, True
 | |
|         return fname, False
 | |
| 
 | |
|     def CheckFakedBlobs(self, faked_blobs_list):
 | |
|         """Check if any entries in this section have faked external blobs
 | |
| 
 | |
|         If there are faked blobs, the entries are added to the list
 | |
| 
 | |
|         Args:
 | |
|             fake_blobs_list: List of Entry objects to be added to
 | |
|         """
 | |
|         # This is meaningless for anything other than blobs
 | |
|         pass
 | |
| 
 | |
|     def GetAllowMissing(self):
 | |
|         """Get whether a section allows missing external blobs
 | |
| 
 | |
|         Returns:
 | |
|             True if allowed, False if not allowed
 | |
|         """
 | |
|         return self.allow_missing
 | |
| 
 | |
|     def record_missing_bintool(self, bintool):
 | |
|         """Record a missing bintool that was needed to produce this entry
 | |
| 
 | |
|         Args:
 | |
|             bintool (Bintool): Bintool that was missing
 | |
|         """
 | |
|         if bintool not in self.missing_bintools:
 | |
|             self.missing_bintools.append(bintool)
 | |
| 
 | |
|     def check_missing_bintools(self, missing_list):
 | |
|         """Check if any entries in this section have missing bintools
 | |
| 
 | |
|         If there are missing bintools, these are added to the list
 | |
| 
 | |
|         Args:
 | |
|             missing_list: List of Bintool objects to be added to
 | |
|         """
 | |
|         for bintool in self.missing_bintools:
 | |
|             if bintool not in missing_list:
 | |
|                 missing_list.append(bintool)
 | |
| 
 | |
| 
 | |
|     def GetHelpTags(self):
 | |
|         """Get the tags use for missing-blob help
 | |
| 
 | |
|         Returns:
 | |
|             list of possible tags, most desirable first
 | |
|         """
 | |
|         return list(filter(None, [self.missing_msg, self.name, self.etype]))
 | |
| 
 | |
|     def CompressData(self, indata):
 | |
|         """Compress data according to the entry's compression method
 | |
| 
 | |
|         Args:
 | |
|             indata: Data to compress
 | |
| 
 | |
|         Returns:
 | |
|             Compressed data
 | |
|         """
 | |
|         self.uncomp_data = indata
 | |
|         if self.compress != 'none':
 | |
|             self.uncomp_size = len(indata)
 | |
|             if self.comp_bintool.is_present():
 | |
|                 data = self.comp_bintool.compress(indata)
 | |
|             else:
 | |
|                 self.record_missing_bintool(self.comp_bintool)
 | |
|                 data = tools.get_bytes(0, 1024)
 | |
|         else:
 | |
|             data = indata
 | |
|         return data
 | |
| 
 | |
|     def DecompressData(self, indata):
 | |
|         """Decompress data according to the entry's compression method
 | |
| 
 | |
|         Args:
 | |
|             indata: Data to decompress
 | |
| 
 | |
|         Returns:
 | |
|             Decompressed data
 | |
|         """
 | |
|         if self.compress != 'none':
 | |
|             if self.comp_bintool.is_present():
 | |
|                 data = self.comp_bintool.decompress(indata)
 | |
|                 self.uncomp_size = len(data)
 | |
|             else:
 | |
|                 self.record_missing_bintool(self.comp_bintool)
 | |
|                 data = tools.get_bytes(0, 1024)
 | |
|         else:
 | |
|             data = indata
 | |
|         self.uncomp_data = data
 | |
|         return data
 | |
| 
 | |
|     @classmethod
 | |
|     def UseExpanded(cls, node, etype, new_etype):
 | |
|         """Check whether to use an expanded entry type
 | |
| 
 | |
|         This is called by Entry.Create() when it finds an expanded version of
 | |
|         an entry type (e.g. 'u-boot-expanded'). If this method returns True then
 | |
|         it will be used (e.g. in place of 'u-boot'). If it returns False, it is
 | |
|         ignored.
 | |
| 
 | |
|         Args:
 | |
|             node:     Node object containing information about the entry to
 | |
|                       create
 | |
|             etype:    Original entry type being used
 | |
|             new_etype: New entry type proposed
 | |
| 
 | |
|         Returns:
 | |
|             True to use this entry type, False to use the original one
 | |
|         """
 | |
|         tout.info("Node '%s': etype '%s': %s selected" %
 | |
|                   (node.path, etype, new_etype))
 | |
|         return True
 | |
| 
 | |
|     def CheckAltFormats(self, alt_formats):
 | |
|         """Add any alternative formats supported by this entry type
 | |
| 
 | |
|         Args:
 | |
|             alt_formats (dict): Dict to add alt_formats to:
 | |
|                 key: Name of alt format
 | |
|                 value: Help text
 | |
|         """
 | |
|         pass
 | |
| 
 | |
|     def AddBintools(self, btools):
 | |
|         """Add the bintools used by this entry type
 | |
| 
 | |
|         Args:
 | |
|             btools (dict of Bintool):
 | |
| 
 | |
|         Raise:
 | |
|             ValueError if compression algorithm is not supported
 | |
|         """
 | |
|         algo = self.compress
 | |
|         if algo != 'none':
 | |
|             algos = ['bzip2', 'lz4', 'lzma']
 | |
|             if algo not in algos:
 | |
|                 raise ValueError("Unknown algorithm '%s'" % algo)
 | |
|             names = {'lzma': 'lzma_alone'}
 | |
|             name = names.get(self.compress, self.compress)
 | |
|             self.comp_bintool = self.AddBintool(btools, name)
 | |
| 
 | |
|     @classmethod
 | |
|     def AddBintool(self, tools, name):
 | |
|         """Add a new bintool to the tools used by this etype
 | |
| 
 | |
|         Args:
 | |
|             name: Name of the tool
 | |
|         """
 | |
|         btool = bintool.Bintool.create(name)
 | |
|         tools[name] = btool
 | |
|         return btool
 | |
| 
 | |
|     def SetUpdateHash(self, update_hash):
 | |
|         """Set whether this entry's "hash" subnode should be updated
 | |
| 
 | |
|         Args:
 | |
|             update_hash: True if hash should be updated, False if not
 | |
|         """
 | |
|         self.update_hash = update_hash
 | |
| 
 | |
|     def collect_contents_to_file(self, entries, prefix, fake_size=0):
 | |
|         """Put the contents of a list of entries into a file
 | |
| 
 | |
|         Args:
 | |
|             entries (list of Entry): Entries to collect
 | |
|             prefix (str): Filename prefix of file to write to
 | |
|             fake_size (int): Size of fake file to create if needed
 | |
| 
 | |
|         If any entry does not have contents yet, this function returns False
 | |
|         for the data.
 | |
| 
 | |
|         Returns:
 | |
|             Tuple:
 | |
|                 bytes: Concatenated data from all the entries (or None)
 | |
|                 str: Filename of file written (or None if no data)
 | |
|                 str: Unique portion of filename (or None if no data)
 | |
|         """
 | |
|         data = b''
 | |
|         for entry in entries:
 | |
|             # First get the input data and put it in a file. If not available,
 | |
|             # try later.
 | |
|             if not entry.ObtainContents(fake_size=fake_size):
 | |
|                 return None, None, None
 | |
|             data += entry.GetData()
 | |
|         uniq = self.GetUniqueName()
 | |
|         fname = tools.get_output_filename(f'{prefix}.{uniq}')
 | |
|         tools.write_file(fname, data)
 | |
|         return data, fname, uniq
 | |
| 
 | |
|     @classmethod
 | |
|     def create_fake_dir(cls):
 | |
|         """Create the directory for fake files"""
 | |
|         cls.fake_dir = tools.get_output_filename('binman-fake')
 | |
|         if not os.path.exists(cls.fake_dir):
 | |
|             os.mkdir(cls.fake_dir)
 | |
|         tout.notice(f"Fake-blob dir is '{cls.fake_dir}'")
 | |
| 
 | |
|     def ensure_props(self):
 | |
|         """Raise an exception if properties are missing
 | |
| 
 | |
|         Args:
 | |
|             prop_list (list of str): List of properties to check for
 | |
| 
 | |
|         Raises:
 | |
|             ValueError: Any property is missing
 | |
|         """
 | |
|         not_present = []
 | |
|         for prop in self.required_props:
 | |
|             if not prop in self._node.props:
 | |
|                 not_present.append(prop)
 | |
|         if not_present:
 | |
|             self.Raise(f"'{self.etype}' entry is missing properties: {' '.join(not_present)}")
 |