|
"""
|
|
olefile (formerly OleFileIO_PL)
|
|
|
|
Module to read/write Microsoft OLE2 files (also called Structured Storage or
|
|
Microsoft Compound Document File Format), such as Microsoft Office 97-2003
|
|
documents, Image Composer and FlashPix files, Outlook messages, ...
|
|
This version is compatible with Python 2.7 and 3.4+
|
|
|
|
Project website: https://www.decalage.info/olefile
|
|
|
|
olefile is copyright (c) 2005-2018 Philippe Lagadec
|
|
(https://www.decalage.info)
|
|
|
|
olefile is based on the OleFileIO module from the PIL library v1.1.7
|
|
See: http://www.pythonware.com/products/pil/index.htm
|
|
and http://svn.effbot.org/public/tags/pil-1.1.7/PIL/OleFileIO.py
|
|
|
|
The Python Imaging Library (PIL) is
|
|
Copyright (c) 1997-2009 by Secret Labs AB
|
|
Copyright (c) 1995-2009 by Fredrik Lundh
|
|
|
|
See source code and LICENSE.txt for information on usage and redistribution.
|
|
"""
|
|
|
|
# Since OleFileIO_PL v0.45, only Python 2.7 and 3.4+ are supported
|
|
# This import enables print() as a function rather than a keyword
|
|
# (main requirement to be compatible with Python 3.x)
|
|
# The comment on the line below should be printed on Python 2.5 or older:
|
|
from __future__ import print_function # This version of olefile requires Python 2.7 or 3.4+.
|
|
|
|
|
|
#--- LICENSE ------------------------------------------------------------------
|
|
|
|
# olefile (formerly OleFileIO_PL) is copyright (c) 2005-2018 Philippe Lagadec
|
|
# (https://www.decalage.info)
|
|
#
|
|
# All rights reserved.
|
|
#
|
|
# Redistribution and use in source and binary forms, with or without modification,
|
|
# are permitted provided that the following conditions are met:
|
|
#
|
|
# * Redistributions of source code must retain the above copyright notice, this
|
|
# list of conditions and the following disclaimer.
|
|
# * Redistributions in binary form must reproduce the above copyright notice,
|
|
# this list of conditions and the following disclaimer in the documentation
|
|
# and/or other materials provided with the distribution.
|
|
#
|
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
|
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
# CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
# OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
# ----------
|
|
# PIL License:
|
|
#
|
|
# olefile is based on source code from the OleFileIO module of the Python
|
|
# Imaging Library (PIL) published by Fredrik Lundh under the following license:
|
|
|
|
# The Python Imaging Library (PIL) is
|
|
# Copyright (c) 1997-2009 by Secret Labs AB
|
|
# Copyright (c) 1995-2009 by Fredrik Lundh
|
|
#
|
|
# By obtaining, using, and/or copying this software and/or its associated
|
|
# documentation, you agree that you have read, understood, and will comply with
|
|
# the following terms and conditions:
|
|
#
|
|
# Permission to use, copy, modify, and distribute this software and its
|
|
# associated documentation for any purpose and without fee is hereby granted,
|
|
# provided that the above copyright notice appears in all copies, and that both
|
|
# that copyright notice and this permission notice appear in supporting
|
|
# documentation, and that the name of Secret Labs AB or the author(s) not be used
|
|
# in advertising or publicity pertaining to distribution of the software
|
|
# without specific, written prior permission.
|
|
#
|
|
# SECRET LABS AB AND THE AUTHORS DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS
|
|
# SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS.
|
|
# IN NO EVENT SHALL SECRET LABS AB OR THE AUTHORS BE LIABLE FOR ANY SPECIAL,
|
|
# INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM
|
|
# LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR
|
|
# OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR
|
|
# PERFORMANCE OF THIS SOFTWARE.
|
|
|
|
__date__ = "2018-09-09"
|
|
__version__ = '0.46'
|
|
__author__ = "Philippe Lagadec"
|
|
|
|
__all__ = ['isOleFile', 'OleFileIO', 'OleMetadata', 'enable_logging',
|
|
'MAGIC', 'STGTY_EMPTY', 'KEEP_UNICODE_NAMES',
|
|
'STGTY_STREAM', 'STGTY_STORAGE', 'STGTY_ROOT', 'STGTY_PROPERTY',
|
|
'STGTY_LOCKBYTES', 'MINIMAL_OLEFILE_SIZE',
|
|
'DEFECT_UNSURE', 'DEFECT_POTENTIAL', 'DEFECT_INCORRECT',
|
|
'DEFECT_FATAL', 'DEFAULT_PATH_ENCODING',
|
|
'MAXREGSECT', 'DIFSECT', 'FATSECT', 'ENDOFCHAIN', 'FREESECT',
|
|
'MAXREGSID', 'NOSTREAM', 'UNKNOWN_SIZE', 'WORD_CLSID'
|
|
]
|
|
|
|
import io
|
|
import sys
|
|
import struct, array, os.path, datetime, logging
|
|
|
|
#=== COMPATIBILITY WORKAROUNDS ================================================
|
|
|
|
# For Python 3.x, need to redefine long as int:
|
|
if str is not bytes:
|
|
long = int
|
|
|
|
# Need to make sure we use xrange both on Python 2 and 3.x:
|
|
try:
|
|
# on Python 2 we need xrange:
|
|
iterrange = xrange
|
|
except:
|
|
# no xrange, for Python 3 it was renamed as range:
|
|
iterrange = range
|
|
|
|
#[PL] workaround to fix an issue with array item size on 64 bits systems:
|
|
if array.array('L').itemsize == 4:
|
|
# on 32 bits platforms, long integers in an array are 32 bits:
|
|
UINT32 = 'L'
|
|
elif array.array('I').itemsize == 4:
|
|
# on 64 bits platforms, integers in an array are 32 bits:
|
|
UINT32 = 'I'
|
|
elif array.array('i').itemsize == 4:
|
|
# On 64 bit Jython, signed integers ('i') are the only way to store our 32
|
|
# bit values in an array in a *somewhat* reasonable way, as the otherwise
|
|
# perfectly suited 'H' (unsigned int, 32 bits) results in a completely
|
|
# unusable behaviour. This is most likely caused by the fact that Java
|
|
# doesn't have unsigned values, and thus Jython's "array" implementation,
|
|
# which is based on "jarray", doesn't have them either.
|
|
# NOTE: to trick Jython into converting the values it would normally
|
|
# interpret as "signed" into "unsigned", a binary-and operation with
|
|
# 0xFFFFFFFF can be used. This way it is possible to use the same comparing
|
|
# operations on all platforms / implementations. The corresponding code
|
|
# lines are flagged with a 'JYTHON-WORKAROUND' tag below.
|
|
UINT32 = 'i'
|
|
else:
|
|
raise ValueError('Need to fix a bug with 32 bit arrays, please contact author...')
|
|
|
|
|
|
#[PL] These workarounds were inspired from the Path module
|
|
# (see http://www.jorendorff.com/articles/python/path/)
|
|
try:
|
|
basestring
|
|
except NameError:
|
|
basestring = str
|
|
|
|
#[PL] Experimental setting: if True, OLE filenames will be kept in Unicode
|
|
# if False (default PIL behaviour), all filenames are converted to Latin-1.
|
|
KEEP_UNICODE_NAMES = True
|
|
|
|
if sys.version_info[0] < 3:
|
|
# On Python 2.x, the default encoding for path names is UTF-8:
|
|
DEFAULT_PATH_ENCODING = 'utf-8'
|
|
else:
|
|
# On Python 3.x, the default encoding for path names is Unicode (None):
|
|
DEFAULT_PATH_ENCODING = None
|
|
|
|
|
|
# === LOGGING =================================================================
|
|
|
|
def get_logger(name, level=logging.CRITICAL+1):
|
|
"""
|
|
Create a suitable logger object for this module.
|
|
The goal is not to change settings of the root logger, to avoid getting
|
|
other modules' logs on the screen.
|
|
If a logger exists with same name, reuse it. (Else it would have duplicate
|
|
handlers and messages would be doubled.)
|
|
The level is set to CRITICAL+1 by default, to avoid any logging.
|
|
"""
|
|
# First, test if there is already a logger with the same name, else it
|
|
# will generate duplicate messages (due to duplicate handlers):
|
|
if name in logging.Logger.manager.loggerDict:
|
|
#NOTE: another less intrusive but more "hackish" solution would be to
|
|
# use getLogger then test if its effective level is not default.
|
|
logger = logging.getLogger(name)
|
|
# make sure level is OK:
|
|
logger.setLevel(level)
|
|
return logger
|
|
# get a new logger:
|
|
logger = logging.getLogger(name)
|
|
# only add a NullHandler for this logger, it is up to the application
|
|
# to configure its own logging:
|
|
logger.addHandler(logging.NullHandler())
|
|
logger.setLevel(level)
|
|
return logger
|
|
|
|
|
|
# a global logger object used for debugging:
|
|
log = get_logger('olefile')
|
|
|
|
|
|
def enable_logging():
|
|
"""
|
|
Enable logging for this module (disabled by default).
|
|
This will set the module-specific logger level to NOTSET, which
|
|
means the main application controls the actual logging level.
|
|
"""
|
|
log.setLevel(logging.NOTSET)
|
|
|
|
|
|
#=== CONSTANTS ===============================================================
|
|
|
|
#: magic bytes that should be at the beginning of every OLE file:
|
|
MAGIC = b'\xD0\xCF\x11\xE0\xA1\xB1\x1A\xE1'
|
|
|
|
#[PL]: added constants for Sector IDs (from AAF specifications)
|
|
MAXREGSECT = 0xFFFFFFFA #: (-6) maximum SECT
|
|
DIFSECT = 0xFFFFFFFC #: (-4) denotes a DIFAT sector in a FAT
|
|
FATSECT = 0xFFFFFFFD #: (-3) denotes a FAT sector in a FAT
|
|
ENDOFCHAIN = 0xFFFFFFFE #: (-2) end of a virtual stream chain
|
|
FREESECT = 0xFFFFFFFF #: (-1) unallocated sector
|
|
|
|
#[PL]: added constants for Directory Entry IDs (from AAF specifications)
|
|
MAXREGSID = 0xFFFFFFFA #: (-6) maximum directory entry ID
|
|
NOSTREAM = 0xFFFFFFFF #: (-1) unallocated directory entry
|
|
|
|
#[PL] object types in storage (from AAF specifications)
|
|
STGTY_EMPTY = 0 #: empty directory entry
|
|
STGTY_STORAGE = 1 #: element is a storage object
|
|
STGTY_STREAM = 2 #: element is a stream object
|
|
STGTY_LOCKBYTES = 3 #: element is an ILockBytes object
|
|
STGTY_PROPERTY = 4 #: element is an IPropertyStorage object
|
|
STGTY_ROOT = 5 #: element is a root storage
|
|
|
|
# Unknown size for a stream (used by OleStream):
|
|
UNKNOWN_SIZE = 0x7FFFFFFF
|
|
|
|
#
|
|
# --------------------------------------------------------------------
|
|
# property types
|
|
|
|
VT_EMPTY=0; VT_NULL=1; VT_I2=2; VT_I4=3; VT_R4=4; VT_R8=5; VT_CY=6;
|
|
VT_DATE=7; VT_BSTR=8; VT_DISPATCH=9; VT_ERROR=10; VT_BOOL=11;
|
|
VT_VARIANT=12; VT_UNKNOWN=13; VT_DECIMAL=14; VT_I1=16; VT_UI1=17;
|
|
VT_UI2=18; VT_UI4=19; VT_I8=20; VT_UI8=21; VT_INT=22; VT_UINT=23;
|
|
VT_VOID=24; VT_HRESULT=25; VT_PTR=26; VT_SAFEARRAY=27; VT_CARRAY=28;
|
|
VT_USERDEFINED=29; VT_LPSTR=30; VT_LPWSTR=31; VT_FILETIME=64;
|
|
VT_BLOB=65; VT_STREAM=66; VT_STORAGE=67; VT_STREAMED_OBJECT=68;
|
|
VT_STORED_OBJECT=69; VT_BLOB_OBJECT=70; VT_CF=71; VT_CLSID=72;
|
|
VT_VECTOR=0x1000;
|
|
|
|
# map property id to name (for debugging purposes)
|
|
# VT = {}
|
|
# for keyword, var in list(vars().items()):
|
|
# if keyword[:3] == "VT_":
|
|
# VT[var] = keyword
|
|
|
|
#
|
|
# --------------------------------------------------------------------
|
|
# Some common document types (root.clsid fields)
|
|
|
|
WORD_CLSID = "00020900-0000-0000-C000-000000000046"
|
|
#TODO: check Excel, PPT, ...
|
|
|
|
#[PL]: Defect levels to classify parsing errors - see OleFileIO._raise_defect()
|
|
DEFECT_UNSURE = 10 # a case which looks weird, but not sure it's a defect
|
|
DEFECT_POTENTIAL = 20 # a potential defect
|
|
DEFECT_INCORRECT = 30 # an error according to specifications, but parsing
|
|
# can go on
|
|
DEFECT_FATAL = 40 # an error which cannot be ignored, parsing is
|
|
# impossible
|
|
|
|
# Minimal size of an empty OLE file, with 512-bytes sectors = 1536 bytes
|
|
# (this is used in isOleFile and OleFile.open)
|
|
MINIMAL_OLEFILE_SIZE = 1536
|
|
|
|
|
|
#=== FUNCTIONS ===============================================================
|
|
|
|
def isOleFile (filename):
|
|
"""
|
|
Test if a file is an OLE container (according to the magic bytes in its header).
|
|
|
|
.. note::
|
|
This function only checks the first 8 bytes of the file, not the
|
|
rest of the OLE structure.
|
|
|
|
.. versionadded:: 0.16
|
|
|
|
:param filename: filename, contents or file-like object of the OLE file (string-like or file-like object)
|
|
|
|
- if filename is a string smaller than 1536 bytes, it is the path
|
|
of the file to open. (bytes or unicode string)
|
|
- if filename is a string longer than 1535 bytes, it is parsed
|
|
as the content of an OLE file in memory. (bytes type only)
|
|
- if filename is a file-like object (with read and seek methods),
|
|
it is parsed as-is.
|
|
|
|
:type filename: bytes or str or unicode or file
|
|
:returns: True if OLE, False otherwise.
|
|
:rtype: bool
|
|
"""
|
|
# check if filename is a string-like or file-like object:
|
|
if hasattr(filename, 'read'):
|
|
# file-like object: use it directly
|
|
header = filename.read(len(MAGIC))
|
|
# just in case, seek back to start of file:
|
|
filename.seek(0)
|
|
elif isinstance(filename, bytes) and len(filename) >= MINIMAL_OLEFILE_SIZE:
|
|
# filename is a bytes string containing the OLE file to be parsed:
|
|
header = filename[:len(MAGIC)]
|
|
else:
|
|
# string-like object: filename of file on disk
|
|
with open(filename, 'rb') as fp:
|
|
header = fp.read(len(MAGIC))
|
|
if header == MAGIC:
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
|
|
if bytes is str:
|
|
# version for Python 2.x
|
|
def i8(c):
|
|
return ord(c)
|
|
else:
|
|
# version for Python 3.x
|
|
def i8(c):
|
|
return c if c.__class__ is int else c[0]
|
|
|
|
|
|
def i16(c, o = 0):
|
|
"""
|
|
Converts a 2-bytes (16 bits) string to an integer.
|
|
|
|
:param c: string containing bytes to convert
|
|
:param o: offset of bytes to convert in string
|
|
"""
|
|
return struct.unpack("<H", c[o:o+2])[0]
|
|
|
|
|
|
def i32(c, o = 0):
|
|
"""
|
|
Converts a 4-bytes (32 bits) string to an integer.
|
|
|
|
:param c: string containing bytes to convert
|
|
:param o: offset of bytes to convert in string
|
|
"""
|
|
return struct.unpack("<I", c[o:o+4])[0]
|
|
|
|
|
|
def _clsid(clsid):
|
|
"""
|
|
Converts a CLSID to a human-readable string.
|
|
|
|
:param clsid: string of length 16.
|
|
"""
|
|
assert len(clsid) == 16
|
|
# if clsid is only made of null bytes, return an empty string:
|
|
# (PL: why not simply return the string with zeroes?)
|
|
if not clsid.strip(b"\0"):
|
|
return ""
|
|
return (("%08X-%04X-%04X-%02X%02X-" + "%02X" * 6) %
|
|
((i32(clsid, 0), i16(clsid, 4), i16(clsid, 6)) +
|
|
tuple(map(i8, clsid[8:16]))))
|
|
|
|
|
|
|
|
def filetime2datetime(filetime):
|
|
"""
|
|
convert FILETIME (64 bits int) to Python datetime.datetime
|
|
"""
|
|
# TODO: manage exception when microseconds is too large
|
|
# inspired from https://code.activestate.com/recipes/511425-filetime-to-datetime/
|
|
_FILETIME_null_date = datetime.datetime(1601, 1, 1, 0, 0, 0)
|
|
#log.debug('timedelta days=%d' % (filetime//(10*1000000*3600*24)))
|
|
return _FILETIME_null_date + datetime.timedelta(microseconds=filetime//10)
|
|
|
|
|
|
|
|
#=== CLASSES ==================================================================
|
|
|
|
class OleMetadata:
|
|
"""
|
|
class to parse and store metadata from standard properties of OLE files.
|
|
|
|
Available attributes:
|
|
codepage, title, subject, author, keywords, comments, template,
|
|
last_saved_by, revision_number, total_edit_time, last_printed, create_time,
|
|
last_saved_time, num_pages, num_words, num_chars, thumbnail,
|
|
creating_application, security, codepage_doc, category, presentation_target,
|
|
bytes, lines, paragraphs, slides, notes, hidden_slides, mm_clips,
|
|
scale_crop, heading_pairs, titles_of_parts, manager, company, links_dirty,
|
|
chars_with_spaces, unused, shared_doc, link_base, hlinks, hlinks_changed,
|
|
version, dig_sig, content_type, content_status, language, doc_version
|
|
|
|
Note: an attribute is set to None when not present in the properties of the
|
|
OLE file.
|
|
|
|
References for SummaryInformation stream:
|
|
|
|
- https://msdn.microsoft.com/en-us/library/dd942545.aspx
|
|
- https://msdn.microsoft.com/en-us/library/dd925819%28v=office.12%29.aspx
|
|
- https://msdn.microsoft.com/en-us/library/windows/desktop/aa380376%28v=vs.85%29.aspx
|
|
- https://msdn.microsoft.com/en-us/library/aa372045.aspx
|
|
- http://sedna-soft.de/articles/summary-information-stream/
|
|
- https://poi.apache.org/apidocs/org/apache/poi/hpsf/SummaryInformation.html
|
|
|
|
References for DocumentSummaryInformation stream:
|
|
|
|
- https://msdn.microsoft.com/en-us/library/dd945671%28v=office.12%29.aspx
|
|
- https://msdn.microsoft.com/en-us/library/windows/desktop/aa380374%28v=vs.85%29.aspx
|
|
- https://poi.apache.org/apidocs/org/apache/poi/hpsf/DocumentSummaryInformation.html
|
|
|
|
new in version 0.25
|
|
"""
|
|
|
|
# attribute names for SummaryInformation stream properties:
|
|
# (ordered by property id, starting at 1)
|
|
SUMMARY_ATTRIBS = ['codepage', 'title', 'subject', 'author', 'keywords', 'comments',
|
|
'template', 'last_saved_by', 'revision_number', 'total_edit_time',
|
|
'last_printed', 'create_time', 'last_saved_time', 'num_pages',
|
|
'num_words', 'num_chars', 'thumbnail', 'creating_application',
|
|
'security']
|
|
|
|
# attribute names for DocumentSummaryInformation stream properties:
|
|
# (ordered by property id, starting at 1)
|
|
DOCSUM_ATTRIBS = ['codepage_doc', 'category', 'presentation_target', 'bytes', 'lines', 'paragraphs',
|
|
'slides', 'notes', 'hidden_slides', 'mm_clips',
|
|
'scale_crop', 'heading_pairs', 'titles_of_parts', 'manager',
|
|
'company', 'links_dirty', 'chars_with_spaces', 'unused', 'shared_doc',
|
|
'link_base', 'hlinks', 'hlinks_changed', 'version', 'dig_sig',
|
|
'content_type', 'content_status', 'language', 'doc_version']
|
|
|
|
def __init__(self):
|
|
"""
|
|
Constructor for OleMetadata
|
|
All attributes are set to None by default
|
|
"""
|
|
# properties from SummaryInformation stream
|
|
self.codepage = None
|
|
self.title = None
|
|
self.subject = None
|
|
self.author = None
|
|
self.keywords = None
|
|
self.comments = None
|
|
self.template = None
|
|
self.last_saved_by = None
|
|
self.revision_number = None
|
|
self.total_edit_time = None
|
|
self.last_printed = None
|
|
self.create_time = None
|
|
self.last_saved_time = None
|
|
self.num_pages = None
|
|
self.num_words = None
|
|
self.num_chars = None
|
|
self.thumbnail = None
|
|
self.creating_application = None
|
|
self.security = None
|
|
# properties from DocumentSummaryInformation stream
|
|
self.codepage_doc = None
|
|
self.category = None
|
|
self.presentation_target = None
|
|
self.bytes = None
|
|
self.lines = None
|
|
self.paragraphs = None
|
|
self.slides = None
|
|
self.notes = None
|
|
self.hidden_slides = None
|
|
self.mm_clips = None
|
|
self.scale_crop = None
|
|
self.heading_pairs = None
|
|
self.titles_of_parts = None
|
|
self.manager = None
|
|
self.company = None
|
|
self.links_dirty = None
|
|
self.chars_with_spaces = None
|
|
self.unused = None
|
|
self.shared_doc = None
|
|
self.link_base = None
|
|
self.hlinks = None
|
|
self.hlinks_changed = None
|
|
self.version = None
|
|
self.dig_sig = None
|
|
self.content_type = None
|
|
self.content_status = None
|
|
self.language = None
|
|
self.doc_version = None
|
|
|
|
|
|
def parse_properties(self, olefile):
|
|
"""
|
|
Parse standard properties of an OLE file, from the streams
|
|
``\\x05SummaryInformation`` and ``\\x05DocumentSummaryInformation``,
|
|
if present.
|
|
Properties are converted to strings, integers or python datetime objects.
|
|
If a property is not present, its value is set to None.
|
|
"""
|
|
# first set all attributes to None:
|
|
for attrib in (self.SUMMARY_ATTRIBS + self.DOCSUM_ATTRIBS):
|
|
setattr(self, attrib, None)
|
|
if olefile.exists("\x05SummaryInformation"):
|
|
# get properties from the stream:
|
|
# (converting timestamps to python datetime, except total_edit_time,
|
|
# which is property #10)
|
|
props = olefile.getproperties("\x05SummaryInformation",
|
|
convert_time=True, no_conversion=[10])
|
|
# store them into this object's attributes:
|
|
for i in range(len(self.SUMMARY_ATTRIBS)):
|
|
# ids for standards properties start at 0x01, until 0x13
|
|
value = props.get(i+1, None)
|
|
setattr(self, self.SUMMARY_ATTRIBS[i], value)
|
|
if olefile.exists("\x05DocumentSummaryInformation"):
|
|
# get properties from the stream:
|
|
props = olefile.getproperties("\x05DocumentSummaryInformation",
|
|
convert_time=True)
|
|
# store them into this object's attributes:
|
|
for i in range(len(self.DOCSUM_ATTRIBS)):
|
|
# ids for standards properties start at 0x01, until 0x13
|
|
value = props.get(i+1, None)
|
|
setattr(self, self.DOCSUM_ATTRIBS[i], value)
|
|
|
|
def dump(self):
|
|
"""
|
|
Dump all metadata, for debugging purposes.
|
|
"""
|
|
print('Properties from SummaryInformation stream:')
|
|
for prop in self.SUMMARY_ATTRIBS:
|
|
value = getattr(self, prop)
|
|
print('- %s: %s' % (prop, repr(value)))
|
|
print('Properties from DocumentSummaryInformation stream:')
|
|
for prop in self.DOCSUM_ATTRIBS:
|
|
value = getattr(self, prop)
|
|
print('- %s: %s' % (prop, repr(value)))
|
|
|
|
|
|
#--- OleStream ---------------------------------------------------------------
|
|
|
|
class OleStream(io.BytesIO):
|
|
"""
|
|
OLE2 Stream
|
|
|
|
Returns a read-only file object which can be used to read
|
|
the contents of a OLE stream (instance of the BytesIO class).
|
|
To open a stream, use the openstream method in the OleFile class.
|
|
|
|
This function can be used with either ordinary streams,
|
|
or ministreams, depending on the offset, sectorsize, and
|
|
fat table arguments.
|
|
|
|
Attributes:
|
|
|
|
- size: actual size of data stream, after it was opened.
|
|
"""
|
|
# FIXME: should store the list of sects obtained by following
|
|
# the fat chain, and load new sectors on demand instead of
|
|
# loading it all in one go.
|
|
|
|
def __init__(self, fp, sect, size, offset, sectorsize, fat, filesize, olefileio):
|
|
"""
|
|
Constructor for OleStream class.
|
|
|
|
:param fp: file object, the OLE container or the MiniFAT stream
|
|
:param sect: sector index of first sector in the stream
|
|
:param size: total size of the stream
|
|
:param offset: offset in bytes for the first FAT or MiniFAT sector
|
|
:param sectorsize: size of one sector
|
|
:param fat: array/list of sector indexes (FAT or MiniFAT)
|
|
:param filesize: size of OLE file (for debugging)
|
|
:param olefileio: OleFileIO object containing this stream
|
|
:returns: a BytesIO instance containing the OLE stream
|
|
"""
|
|
log.debug('OleStream.__init__:')
|
|
log.debug(' sect=%d (%X), size=%d, offset=%d, sectorsize=%d, len(fat)=%d, fp=%s'
|
|
%(sect,sect,size,offset,sectorsize,len(fat), repr(fp)))
|
|
self.ole = olefileio
|
|
# this check is necessary, otherwise when attempting to open a stream
|
|
# from a closed OleFile, a stream of size zero is returned without
|
|
# raising an exception. (see issue #81)
|
|
if self.ole.fp.closed:
|
|
raise OSError('Attempting to open a stream from a closed OLE File')
|
|
#[PL] To detect malformed documents with FAT loops, we compute the
|
|
# expected number of sectors in the stream:
|
|
unknown_size = False
|
|
if size == UNKNOWN_SIZE:
|
|
# this is the case when called from OleFileIO._open(), and stream
|
|
# size is not known in advance (for example when reading the
|
|
# Directory stream). Then we can only guess maximum size:
|
|
size = len(fat)*sectorsize
|
|
# and we keep a record that size was unknown:
|
|
unknown_size = True
|
|
log.debug(' stream with UNKNOWN SIZE')
|
|
nb_sectors = (size + (sectorsize-1)) // sectorsize
|
|
log.debug('nb_sectors = %d' % nb_sectors)
|
|
# This number should (at least) be less than the total number of
|
|
# sectors in the given FAT:
|
|
if nb_sectors > len(fat):
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'malformed OLE document, stream too large')
|
|
# optimization(?): data is first a list of strings, and join() is called
|
|
# at the end to concatenate all in one string.
|
|
# (this may not be really useful with recent Python versions)
|
|
data = []
|
|
# if size is zero, then first sector index should be ENDOFCHAIN:
|
|
if size == 0 and sect != ENDOFCHAIN:
|
|
log.debug('size == 0 and sect != ENDOFCHAIN:')
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'incorrect OLE sector index for empty stream')
|
|
#[PL] A fixed-length for loop is used instead of an undefined while
|
|
# loop to avoid DoS attacks:
|
|
for i in range(nb_sectors):
|
|
log.debug('Reading stream sector[%d] = %Xh' % (i, sect))
|
|
# Sector index may be ENDOFCHAIN, but only if size was unknown
|
|
if sect == ENDOFCHAIN:
|
|
if unknown_size:
|
|
log.debug('Reached ENDOFCHAIN sector for stream with unknown size')
|
|
break
|
|
else:
|
|
# else this means that the stream is smaller than declared:
|
|
log.debug('sect=ENDOFCHAIN before expected size')
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'incomplete OLE stream')
|
|
# sector index should be within FAT:
|
|
if sect<0 or sect>=len(fat):
|
|
log.debug('sect=%d (%X) / len(fat)=%d' % (sect, sect, len(fat)))
|
|
log.debug('i=%d / nb_sectors=%d' %(i, nb_sectors))
|
|
## tmp_data = b"".join(data)
|
|
## f = open('test_debug.bin', 'wb')
|
|
## f.write(tmp_data)
|
|
## f.close()
|
|
## log.debug('data read so far: %d bytes' % len(tmp_data))
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'incorrect OLE FAT, sector index out of range')
|
|
# stop reading here if the exception is ignored:
|
|
break
|
|
#TODO: merge this code with OleFileIO.getsect() ?
|
|
#TODO: check if this works with 4K sectors:
|
|
try:
|
|
fp.seek(offset + sectorsize * sect)
|
|
except:
|
|
log.debug('sect=%d, seek=%d, filesize=%d' %
|
|
(sect, offset+sectorsize*sect, filesize))
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'OLE sector index out of range')
|
|
# stop reading here if the exception is ignored:
|
|
break
|
|
sector_data = fp.read(sectorsize)
|
|
# [PL] check if there was enough data:
|
|
# Note: if sector is the last of the file, sometimes it is not a
|
|
# complete sector (of 512 or 4K), so we may read less than
|
|
# sectorsize.
|
|
if len(sector_data)!=sectorsize and sect!=(len(fat)-1):
|
|
log.debug('sect=%d / len(fat)=%d, seek=%d / filesize=%d, len read=%d' %
|
|
(sect, len(fat), offset+sectorsize*sect, filesize, len(sector_data)))
|
|
log.debug('seek+len(read)=%d' % (offset+sectorsize*sect+len(sector_data)))
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'incomplete OLE sector')
|
|
data.append(sector_data)
|
|
# jump to next sector in the FAT:
|
|
try:
|
|
sect = fat[sect] & 0xFFFFFFFF # JYTHON-WORKAROUND
|
|
except IndexError:
|
|
# [PL] if pointer is out of the FAT an exception is raised
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'incorrect OLE FAT, sector index out of range')
|
|
# stop reading here if the exception is ignored:
|
|
break
|
|
#[PL] Last sector should be a "end of chain" marker:
|
|
# if sect != ENDOFCHAIN:
|
|
# raise IOError('incorrect last sector index in OLE stream')
|
|
data = b"".join(data)
|
|
# Data is truncated to the actual stream size:
|
|
if len(data) >= size:
|
|
log.debug('Read data of length %d, truncated to stream size %d' % (len(data), size))
|
|
data = data[:size]
|
|
# actual stream size is stored for future use:
|
|
self.size = size
|
|
elif unknown_size:
|
|
# actual stream size was not known, now we know the size of read
|
|
# data:
|
|
log.debug('Read data of length %d, the stream size was unknown' % len(data))
|
|
self.size = len(data)
|
|
else:
|
|
# read data is less than expected:
|
|
log.debug('Read data of length %d, less than expected stream size %d' % (len(data), size))
|
|
# TODO: provide details in exception message
|
|
self.size = len(data)
|
|
self.ole._raise_defect(DEFECT_INCORRECT, 'OLE stream size is less than declared')
|
|
# when all data is read in memory, BytesIO constructor is called
|
|
io.BytesIO.__init__(self, data)
|
|
# Then the OleStream object can be used as a read-only file object.
|
|
|
|
|
|
#--- OleDirectoryEntry -------------------------------------------------------
|
|
|
|
class OleDirectoryEntry:
|
|
|
|
"""
|
|
OLE2 Directory Entry
|
|
"""
|
|
#[PL] parsing code moved from OleFileIO.loaddirectory
|
|
|
|
# struct to parse directory entries:
|
|
# <: little-endian byte order, standard sizes
|
|
# (note: this should guarantee that Q returns a 64 bits int)
|
|
# 64s: string containing entry name in unicode UTF-16 (max 31 chars) + null char = 64 bytes
|
|
# H: uint16, number of bytes used in name buffer, including null = (len+1)*2
|
|
# B: uint8, dir entry type (between 0 and 5)
|
|
# B: uint8, color: 0=black, 1=red
|
|
# I: uint32, index of left child node in the red-black tree, NOSTREAM if none
|
|
# I: uint32, index of right child node in the red-black tree, NOSTREAM if none
|
|
# I: uint32, index of child root node if it is a storage, else NOSTREAM
|
|
# 16s: CLSID, unique identifier (only used if it is a storage)
|
|
# I: uint32, user flags
|
|
# Q (was 8s): uint64, creation timestamp or zero
|
|
# Q (was 8s): uint64, modification timestamp or zero
|
|
# I: uint32, SID of first sector if stream or ministream, SID of 1st sector
|
|
# of stream containing ministreams if root entry, 0 otherwise
|
|
# I: uint32, total stream size in bytes if stream (low 32 bits), 0 otherwise
|
|
# I: uint32, total stream size in bytes if stream (high 32 bits), 0 otherwise
|
|
STRUCT_DIRENTRY = '<64sHBBIII16sIQQIII'
|
|
# size of a directory entry: 128 bytes
|
|
DIRENTRY_SIZE = 128
|
|
assert struct.calcsize(STRUCT_DIRENTRY) == DIRENTRY_SIZE
|
|
|
|
|
|
def __init__(self, entry, sid, olefile):
|
|
"""
|
|
Constructor for an OleDirectoryEntry object.
|
|
Parses a 128-bytes entry from the OLE Directory stream.
|
|
|
|
:param entry : string (must be 128 bytes long)
|
|
:param sid : index of this directory entry in the OLE file directory
|
|
:param olefile: OleFileIO containing this directory entry
|
|
"""
|
|
self.sid = sid
|
|
# ref to olefile is stored for future use
|
|
self.olefile = olefile
|
|
# kids is a list of children entries, if this entry is a storage:
|
|
# (list of OleDirectoryEntry objects)
|
|
self.kids = []
|
|
# kids_dict is a dictionary of children entries, indexed by their
|
|
# name in lowercase: used to quickly find an entry, and to detect
|
|
# duplicates
|
|
self.kids_dict = {}
|
|
# flag used to detect if the entry is referenced more than once in
|
|
# directory:
|
|
self.used = False
|
|
# decode DirEntry
|
|
(
|
|
self.name_raw, # 64s: string containing entry name in unicode UTF-16 (max 31 chars) + null char = 64 bytes
|
|
self.namelength, # H: uint16, number of bytes used in name buffer, including null = (len+1)*2
|
|
self.entry_type,
|
|
self.color,
|
|
self.sid_left,
|
|
self.sid_right,
|
|
self.sid_child,
|
|
clsid,
|
|
self.dwUserFlags,
|
|
self.createTime,
|
|
self.modifyTime,
|
|
self.isectStart,
|
|
self.sizeLow,
|
|
self.sizeHigh
|
|
) = struct.unpack(OleDirectoryEntry.STRUCT_DIRENTRY, entry)
|
|
if self.entry_type not in [STGTY_ROOT, STGTY_STORAGE, STGTY_STREAM, STGTY_EMPTY]:
|
|
olefile._raise_defect(DEFECT_INCORRECT, 'unhandled OLE storage type')
|
|
# only first directory entry can (and should) be root:
|
|
if self.entry_type == STGTY_ROOT and sid != 0:
|
|
olefile._raise_defect(DEFECT_INCORRECT, 'duplicate OLE root entry')
|
|
if sid == 0 and self.entry_type != STGTY_ROOT:
|
|
olefile._raise_defect(DEFECT_INCORRECT, 'incorrect OLE root entry')
|
|
#log.debug(struct.unpack(fmt_entry, entry[:len_entry]))
|
|
# name should be at most 31 unicode characters + null character,
|
|
# so 64 bytes in total (31*2 + 2):
|
|
if self.namelength>64:
|
|
olefile._raise_defect(DEFECT_INCORRECT, 'incorrect DirEntry name length >64 bytes')
|
|
# if exception not raised, namelength is set to the maximum value:
|
|
self.namelength = 64
|
|
# only characters without ending null char are kept:
|
|
self.name_utf16 = self.name_raw[:(self.namelength-2)]
|
|
#TODO: check if the name is actually followed by a null unicode character ([MS-CFB] 2.6.1)
|
|
#TODO: check if the name does not contain forbidden characters:
|
|
# [MS-CFB] 2.6.1: "The following characters are illegal and MUST NOT be part of the name: '/', '\', ':', '!'."
|
|
# name is converted from UTF-16LE to the path encoding specified in the OleFileIO:
|
|
self.name = olefile._decode_utf16_str(self.name_utf16)
|
|
|
|
log.debug('DirEntry SID=%d: %s' % (self.sid, repr(self.name)))
|
|
log.debug(' - type: %d' % self.entry_type)
|
|
log.debug(' - sect: %Xh' % self.isectStart)
|
|
log.debug(' - SID left: %d, right: %d, child: %d' % (self.sid_left,
|
|
self.sid_right, self.sid_child))
|
|
|
|
# sizeHigh is only used for 4K sectors, it should be zero for 512 bytes
|
|
# sectors, BUT apparently some implementations set it as 0xFFFFFFFF, 1
|
|
# or some other value so it cannot be raised as a defect in general:
|
|
if olefile.sectorsize == 512:
|
|
if self.sizeHigh != 0 and self.sizeHigh != 0xFFFFFFFF:
|
|
log.debug('sectorsize=%d, sizeLow=%d, sizeHigh=%d (%X)' %
|
|
(olefile.sectorsize, self.sizeLow, self.sizeHigh, self.sizeHigh))
|
|
olefile._raise_defect(DEFECT_UNSURE, 'incorrect OLE stream size')
|
|
self.size = self.sizeLow
|
|
else:
|
|
self.size = self.sizeLow + (long(self.sizeHigh)<<32)
|
|
log.debug(' - size: %d (sizeLow=%d, sizeHigh=%d)' % (self.size, self.sizeLow, self.sizeHigh))
|
|
|
|
self.clsid = _clsid(clsid)
|
|
# a storage should have a null size, BUT some implementations such as
|
|
# Word 8 for Mac seem to allow non-null values => Potential defect:
|
|
if self.entry_type == STGTY_STORAGE and self.size != 0:
|
|
olefile._raise_defect(DEFECT_POTENTIAL, 'OLE storage with size>0')
|
|
# check if stream is not already referenced elsewhere:
|
|
self.is_minifat = False
|
|
if self.entry_type in (STGTY_ROOT, STGTY_STREAM) and self.size>0:
|
|
if self.size < olefile.minisectorcutoff \
|
|
and self.entry_type==STGTY_STREAM: # only streams can be in MiniFAT
|
|
# ministream object
|
|
self.is_minifat = True
|
|
else:
|
|
self.is_minifat = False
|
|
olefile._check_duplicate_stream(self.isectStart, self.is_minifat)
|
|
self.sect_chain = None
|
|
|
|
|
|
def build_sect_chain(self, olefile):
|
|
if self.sect_chain:
|
|
return
|
|
if self.entry_type not in (STGTY_ROOT, STGTY_STREAM) or self.size == 0:
|
|
return
|
|
|
|
self.sect_chain = list()
|
|
|
|
if self.is_minifat and not olefile.minifat:
|
|
olefile.loadminifat()
|
|
|
|
next_sect = self.isectStart
|
|
while next_sect != ENDOFCHAIN:
|
|
self.sect_chain.append(next_sect)
|
|
if self.is_minifat:
|
|
next_sect = olefile.minifat[next_sect]
|
|
else:
|
|
next_sect = olefile.fat[next_sect]
|
|
|
|
def build_storage_tree(self):
|
|
"""
|
|
Read and build the red-black tree attached to this OleDirectoryEntry
|
|
object, if it is a storage.
|
|
Note that this method builds a tree of all subentries, so it should
|
|
only be called for the root object once.
|
|
"""
|
|
log.debug('build_storage_tree: SID=%d - %s - sid_child=%d'
|
|
% (self.sid, repr(self.name), self.sid_child))
|
|
if self.sid_child != NOSTREAM:
|
|
# if child SID is not NOSTREAM, then this entry is a storage.
|
|
# Let's walk through the tree of children to fill the kids list:
|
|
self.append_kids(self.sid_child)
|
|
|
|
# Note from OpenOffice documentation: the safest way is to
|
|
# recreate the tree because some implementations may store broken
|
|
# red-black trees...
|
|
|
|
# in the OLE file, entries are sorted on (length, name).
|
|
# for convenience, we sort them on name instead:
|
|
# (see rich comparison methods in this class)
|
|
self.kids.sort()
|
|
|
|
|
|
def append_kids(self, child_sid):
|
|
"""
|
|
Walk through red-black tree of children of this directory entry to add
|
|
all of them to the kids list. (recursive method)
|
|
|
|
:param child_sid: index of child directory entry to use, or None when called
|
|
first time for the root. (only used during recursion)
|
|
"""
|
|
log.debug('append_kids: child_sid=%d' % child_sid)
|
|
#[PL] this method was added to use simple recursion instead of a complex
|
|
# algorithm.
|
|
# if this is not a storage or a leaf of the tree, nothing to do:
|
|
if child_sid == NOSTREAM:
|
|
return
|
|
# check if child SID is in the proper range:
|
|
if child_sid<0 or child_sid>=len(self.olefile.direntries):
|
|
self.olefile._raise_defect(DEFECT_INCORRECT, 'OLE DirEntry index out of range')
|
|
else:
|
|
# get child direntry:
|
|
child = self.olefile._load_direntry(child_sid) #direntries[child_sid]
|
|
log.debug('append_kids: child_sid=%d - %s - sid_left=%d, sid_right=%d, sid_child=%d'
|
|
% (child.sid, repr(child.name), child.sid_left, child.sid_right, child.sid_child))
|
|
# Check if kid was not already referenced in a storage:
|
|
if child.used:
|
|
self.olefile._raise_defect(DEFECT_INCORRECT,
|
|
'OLE Entry referenced more than once')
|
|
return
|
|
child.used = True
|
|
# the directory entries are organized as a red-black tree.
|
|
# (cf. Wikipedia for details)
|
|
# First walk through left side of the tree:
|
|
self.append_kids(child.sid_left)
|
|
# Check if its name is not already used (case-insensitive):
|
|
name_lower = child.name.lower()
|
|
if name_lower in self.kids_dict:
|
|
self.olefile._raise_defect(DEFECT_INCORRECT,
|
|
"Duplicate filename in OLE storage")
|
|
# Then the child_sid OleDirectoryEntry object is appended to the
|
|
# kids list and dictionary:
|
|
self.kids.append(child)
|
|
self.kids_dict[name_lower] = child
|
|
# Finally walk through right side of the tree:
|
|
self.append_kids(child.sid_right)
|
|
# Afterwards build kid's own tree if it's also a storage:
|
|
child.build_storage_tree()
|
|
|
|
|
|
def __eq__(self, other):
|
|
"Compare entries by name"
|
|
return self.name == other.name
|
|
|
|
def __lt__(self, other):
|
|
"Compare entries by name"
|
|
return self.name < other.name
|
|
|
|
def __ne__(self, other):
|
|
return not self.__eq__(other)
|
|
|
|
def __le__(self, other):
|
|
return self.__eq__(other) or self.__lt__(other)
|
|
|
|
# Reflected __lt__() and __le__() will be used for __gt__() and __ge__()
|
|
|
|
#TODO: replace by the same function as MS implementation ?
|
|
# (order by name length first, then case-insensitive order)
|
|
|
|
|
|
def dump(self, tab = 0):
|
|
"Dump this entry, and all its subentries (for debug purposes only)"
|
|
TYPES = ["(invalid)", "(storage)", "(stream)", "(lockbytes)",
|
|
"(property)", "(root)"]
|
|
try:
|
|
type_name = TYPES[self.entry_type]
|
|
except IndexError:
|
|
type_name = '(UNKNOWN)'
|
|
print(" "*tab + repr(self.name), type_name, end=' ')
|
|
if self.entry_type in (STGTY_STREAM, STGTY_ROOT):
|
|
print(self.size, "bytes", end=' ')
|
|
print()
|
|
if self.entry_type in (STGTY_STORAGE, STGTY_ROOT) and self.clsid:
|
|
print(" "*tab + "{%s}" % self.clsid)
|
|
|
|
for kid in self.kids:
|
|
kid.dump(tab + 2)
|
|
|
|
|
|
def getmtime(self):
|
|
"""
|
|
Return modification time of a directory entry.
|
|
|
|
:returns: None if modification time is null, a python datetime object
|
|
otherwise (UTC timezone)
|
|
|
|
new in version 0.26
|
|
"""
|
|
if self.modifyTime == 0:
|
|
return None
|
|
return filetime2datetime(self.modifyTime)
|
|
|
|
|
|
def getctime(self):
|
|
"""
|
|
Return creation time of a directory entry.
|
|
|
|
:returns: None if modification time is null, a python datetime object
|
|
otherwise (UTC timezone)
|
|
|
|
new in version 0.26
|
|
"""
|
|
if self.createTime == 0:
|
|
return None
|
|
return filetime2datetime(self.createTime)
|
|
|
|
|
|
#--- OleFileIO ----------------------------------------------------------------
|
|
|
|
class OleFileIO:
|
|
"""
|
|
OLE container object
|
|
|
|
This class encapsulates the interface to an OLE 2 structured
|
|
storage file. Use the listdir and openstream methods to
|
|
access the contents of this file.
|
|
|
|
Object names are given as a list of strings, one for each subentry
|
|
level. The root entry should be omitted. For example, the following
|
|
code extracts all image streams from a Microsoft Image Composer file::
|
|
|
|
ole = OleFileIO("fan.mic")
|
|
|
|
for entry in ole.listdir():
|
|
if entry[1:2] == "Image":
|
|
fin = ole.openstream(entry)
|
|
fout = open(entry[0:1], "wb")
|
|
while True:
|
|
s = fin.read(8192)
|
|
if not s:
|
|
break
|
|
fout.write(s)
|
|
|
|
You can use the viewer application provided with the Python Imaging
|
|
Library to view the resulting files (which happens to be standard
|
|
TIFF files).
|
|
"""
|
|
|
|
def __init__(self, filename=None, raise_defects=DEFECT_FATAL,
|
|
write_mode=False, debug=False, path_encoding=DEFAULT_PATH_ENCODING):
|
|
"""
|
|
Constructor for the OleFileIO class.
|
|
|
|
:param filename: file to open.
|
|
|
|
- if filename is a string smaller than 1536 bytes, it is the path
|
|
of the file to open. (bytes or unicode string)
|
|
- if filename is a string longer than 1535 bytes, it is parsed
|
|
as the content of an OLE file in memory. (bytes type only)
|
|
- if filename is a file-like object (with read, seek and tell methods),
|
|
it is parsed as-is.
|
|
|
|
:param raise_defects: minimal level for defects to be raised as exceptions.
|
|
(use DEFECT_FATAL for a typical application, DEFECT_INCORRECT for a
|
|
security-oriented application, see source code for details)
|
|
|
|
:param write_mode: bool, if True the file is opened in read/write mode instead
|
|
of read-only by default.
|
|
|
|
:param debug: bool, set debug mode (deprecated, not used anymore)
|
|
|
|
:param path_encoding: None or str, name of the codec to use for path
|
|
names (streams and storages), or None for Unicode.
|
|
Unicode by default on Python 3+, UTF-8 on Python 2.x.
|
|
(new in olefile 0.42, was hardcoded to Latin-1 until olefile v0.41)
|
|
"""
|
|
# minimal level for defects to be raised as exceptions:
|
|
self._raise_defects_level = raise_defects
|
|
#: list of defects/issues not raised as exceptions:
|
|
#: tuples of (exception type, message)
|
|
self.parsing_issues = []
|
|
self.write_mode = write_mode
|
|
self.path_encoding = path_encoding
|
|
# initialize all attributes to default values:
|
|
self._filesize = None
|
|
self.ministream = None
|
|
self._used_streams_fat = []
|
|
self._used_streams_minifat = []
|
|
self.byte_order = None
|
|
self.directory_fp = None
|
|
self.direntries = None
|
|
self.dll_version = None
|
|
self.fat = None
|
|
self.first_difat_sector = None
|
|
self.first_dir_sector = None
|
|
self.first_mini_fat_sector = None
|
|
self.fp = None
|
|
self.header_clsid = None
|
|
self.header_signature = None
|
|
self.metadata = None
|
|
self.mini_sector_shift = None
|
|
self.mini_sector_size = None
|
|
self.mini_stream_cutoff_size = None
|
|
self.minifat = None
|
|
self.minifatsect = None
|
|
# TODO: duplicates?
|
|
self.minisectorcutoff = None
|
|
self.minisectorsize = None
|
|
self.ministream = None
|
|
self.minor_version = None
|
|
self.nb_sect = None
|
|
self.num_difat_sectors = None
|
|
self.num_dir_sectors = None
|
|
self.num_fat_sectors = None
|
|
self.num_mini_fat_sectors = None
|
|
self.reserved1 = None
|
|
self.reserved2 = None
|
|
self.root = None
|
|
self.sector_shift = None
|
|
self.sector_size = None
|
|
self.transaction_signature_number = None
|
|
if filename:
|
|
self.open(filename, write_mode=write_mode)
|
|
|
|
|
|
def __enter__(self):
|
|
return self
|
|
|
|
|
|
def __exit__(self, *args):
|
|
self.close()
|
|
|
|
|
|
def _raise_defect(self, defect_level, message, exception_type=IOError):
|
|
"""
|
|
This method should be called for any defect found during file parsing.
|
|
It may raise an IOError exception according to the minimal level chosen
|
|
for the OleFileIO object.
|
|
|
|
:param defect_level: defect level, possible values are:
|
|
|
|
- DEFECT_UNSURE : a case which looks weird, but not sure it's a defect
|
|
- DEFECT_POTENTIAL : a potential defect
|
|
- DEFECT_INCORRECT : an error according to specifications, but parsing can go on
|
|
- DEFECT_FATAL : an error which cannot be ignored, parsing is impossible
|
|
|
|
:param message: string describing the defect, used with raised exception.
|
|
:param exception_type: exception class to be raised, IOError by default
|
|
"""
|
|
# added by [PL]
|
|
if defect_level >= self._raise_defects_level:
|
|
log.error(message)
|
|
raise exception_type(message)
|
|
else:
|
|
# just record the issue, no exception raised:
|
|
self.parsing_issues.append((exception_type, message))
|
|
log.warning(message)
|
|
|
|
|
|
def _decode_utf16_str(self, utf16_str, errors='replace'):
|
|
"""
|
|
Decode a string encoded in UTF-16 LE format, as found in the OLE
|
|
directory or in property streams. Return a string encoded
|
|
according to the path_encoding specified for the OleFileIO object.
|
|
|
|
:param utf16_str: bytes string encoded in UTF-16 LE format
|
|
:param errors: str, see python documentation for str.decode()
|
|
:return: str, encoded according to path_encoding
|
|
"""
|
|
unicode_str = utf16_str.decode('UTF-16LE', errors)
|
|
if self.path_encoding:
|
|
# an encoding has been specified for path names:
|
|
return unicode_str.encode(self.path_encoding, errors)
|
|
else:
|
|
# path_encoding=None, return the Unicode string as-is:
|
|
return unicode_str
|
|
|
|
|
|
def open(self, filename, write_mode=False):
|
|
"""
|
|
Open an OLE2 file in read-only or read/write mode.
|
|
Read and parse the header, FAT and directory.
|
|
|
|
:param filename: string-like or file-like object, OLE file to parse
|
|
|
|
- if filename is a string smaller than 1536 bytes, it is the path
|
|
of the file to open. (bytes or unicode string)
|
|
- if filename is a string longer than 1535 bytes, it is parsed
|
|
as the content of an OLE file in memory. (bytes type only)
|
|
- if filename is a file-like object (with read, seek and tell methods),
|
|
it is parsed as-is.
|
|
|
|
:param write_mode: bool, if True the file is opened in read/write mode instead
|
|
of read-only by default. (ignored if filename is not a path)
|
|
"""
|
|
self.write_mode = write_mode
|
|
#[PL] check if filename is a string-like or file-like object:
|
|
# (it is better to check for a read() method)
|
|
if hasattr(filename, 'read'):
|
|
#TODO: also check seek and tell methods?
|
|
# file-like object: use it directly
|
|
self.fp = filename
|
|
elif isinstance(filename, bytes) and len(filename) >= MINIMAL_OLEFILE_SIZE:
|
|
# filename is a bytes string containing the OLE file to be parsed:
|
|
# convert it to BytesIO
|
|
self.fp = io.BytesIO(filename)
|
|
else:
|
|
# string-like object: filename of file on disk
|
|
if self.write_mode:
|
|
# open file in mode 'read with update, binary'
|
|
# According to https://docs.python.org/2/library/functions.html#open
|
|
# 'w' would truncate the file, 'a' may only append on some Unixes
|
|
mode = 'r+b'
|
|
else:
|
|
# read-only mode by default
|
|
mode = 'rb'
|
|
self.fp = open(filename, mode)
|
|
# obtain the filesize by using seek and tell, which should work on most
|
|
# file-like objects:
|
|
#TODO: do it above, using getsize with filename when possible?
|
|
#TODO: fix code to fail with clear exception when filesize cannot be obtained
|
|
filesize=0
|
|
self.fp.seek(0, os.SEEK_END)
|
|
try:
|
|
filesize = self.fp.tell()
|
|
finally:
|
|
self.fp.seek(0)
|
|
self._filesize = filesize
|
|
log.debug('File size: %d bytes (%Xh)' % (self._filesize, self._filesize))
|
|
|
|
# lists of streams in FAT and MiniFAT, to detect duplicate references
|
|
# (list of indexes of first sectors of each stream)
|
|
self._used_streams_fat = []
|
|
self._used_streams_minifat = []
|
|
|
|
header = self.fp.read(512)
|
|
|
|
if len(header) != 512 or header[:8] != MAGIC:
|
|
log.debug('Magic = %r instead of %r' % (header[:8], MAGIC))
|
|
self._raise_defect(DEFECT_FATAL, "not an OLE2 structured storage file")
|
|
|
|
# [PL] header structure according to AAF specifications:
|
|
##Header
|
|
##struct StructuredStorageHeader { // [offset from start (bytes), length (bytes)]
|
|
##BYTE _abSig[8]; // [00H,08] {0xd0, 0xcf, 0x11, 0xe0, 0xa1, 0xb1,
|
|
## // 0x1a, 0xe1} for current version
|
|
##CLSID _clsid; // [08H,16] reserved must be zero (WriteClassStg/
|
|
## // GetClassFile uses root directory class id)
|
|
##USHORT _uMinorVersion; // [18H,02] minor version of the format: 33 is
|
|
## // written by reference implementation
|
|
##USHORT _uDllVersion; // [1AH,02] major version of the dll/format: 3 for
|
|
## // 512-byte sectors, 4 for 4 KB sectors
|
|
##USHORT _uByteOrder; // [1CH,02] 0xFFFE: indicates Intel byte-ordering
|
|
##USHORT _uSectorShift; // [1EH,02] size of sectors in power-of-two;
|
|
## // typically 9 indicating 512-byte sectors
|
|
##USHORT _uMiniSectorShift; // [20H,02] size of mini-sectors in power-of-two;
|
|
## // typically 6 indicating 64-byte mini-sectors
|
|
##USHORT _usReserved; // [22H,02] reserved, must be zero
|
|
##ULONG _ulReserved1; // [24H,04] reserved, must be zero
|
|
##FSINDEX _csectDir; // [28H,04] must be zero for 512-byte sectors,
|
|
## // number of SECTs in directory chain for 4 KB
|
|
## // sectors
|
|
##FSINDEX _csectFat; // [2CH,04] number of SECTs in the FAT chain
|
|
##SECT _sectDirStart; // [30H,04] first SECT in the directory chain
|
|
##DFSIGNATURE _signature; // [34H,04] signature used for transactions; must
|
|
## // be zero. The reference implementation
|
|
## // does not support transactions
|
|
##ULONG _ulMiniSectorCutoff; // [38H,04] maximum size for a mini stream;
|
|
## // typically 4096 bytes
|
|
##SECT _sectMiniFatStart; // [3CH,04] first SECT in the MiniFAT chain
|
|
##FSINDEX _csectMiniFat; // [40H,04] number of SECTs in the MiniFAT chain
|
|
##SECT _sectDifStart; // [44H,04] first SECT in the DIFAT chain
|
|
##FSINDEX _csectDif; // [48H,04] number of SECTs in the DIFAT chain
|
|
##SECT _sectFat[109]; // [4CH,436] the SECTs of first 109 FAT sectors
|
|
##};
|
|
|
|
# [PL] header decoding:
|
|
# '<' indicates little-endian byte ordering for Intel (cf. struct module help)
|
|
fmt_header = '<8s16sHHHHHHLLLLLLLLLL'
|
|
header_size = struct.calcsize(fmt_header)
|
|
log.debug( "fmt_header size = %d, +FAT = %d" % (header_size, header_size + 109*4) )
|
|
header1 = header[:header_size]
|
|
(
|
|
self.header_signature,
|
|
self.header_clsid,
|
|
self.minor_version,
|
|
self.dll_version,
|
|
self.byte_order,
|
|
self.sector_shift,
|
|
self.mini_sector_shift,
|
|
self.reserved1,
|
|
self.reserved2,
|
|
self.num_dir_sectors,
|
|
self.num_fat_sectors,
|
|
self.first_dir_sector,
|
|
self.transaction_signature_number,
|
|
self.mini_stream_cutoff_size,
|
|
self.first_mini_fat_sector,
|
|
self.num_mini_fat_sectors,
|
|
self.first_difat_sector,
|
|
self.num_difat_sectors
|
|
) = struct.unpack(fmt_header, header1)
|
|
log.debug( struct.unpack(fmt_header, header1))
|
|
|
|
if self.header_signature != MAGIC:
|
|
# OLE signature should always be present
|
|
self._raise_defect(DEFECT_FATAL, "incorrect OLE signature")
|
|
if self.header_clsid != bytearray(16):
|
|
# according to AAF specs, CLSID should always be zero
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect CLSID in OLE header")
|
|
log.debug( "Minor Version = %d" % self.minor_version )
|
|
# TODO: according to MS-CFB, minor version should be 0x003E
|
|
log.debug( "DLL Version = %d (expected: 3 or 4)" % self.dll_version )
|
|
if self.dll_version not in [3, 4]:
|
|
# version 3: usual format, 512 bytes per sector
|
|
# version 4: large format, 4K per sector
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect DllVersion in OLE header")
|
|
log.debug( "Byte Order = %X (expected: FFFE)" % self.byte_order )
|
|
if self.byte_order != 0xFFFE:
|
|
# For now only common little-endian documents are handled correctly
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect ByteOrder in OLE header")
|
|
# TODO: add big-endian support for documents created on Mac ?
|
|
# But according to [MS-CFB] ? v20140502, ByteOrder MUST be 0xFFFE.
|
|
self.sector_size = 2**self.sector_shift
|
|
log.debug( "Sector Size = %d bytes (expected: 512 or 4096)" % self.sector_size )
|
|
if self.sector_size not in [512, 4096]:
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect sector_size in OLE header")
|
|
if (self.dll_version==3 and self.sector_size!=512) \
|
|
or (self.dll_version==4 and self.sector_size!=4096):
|
|
self._raise_defect(DEFECT_INCORRECT, "sector_size does not match DllVersion in OLE header")
|
|
self.mini_sector_size = 2**self.mini_sector_shift
|
|
log.debug( "MiniFAT Sector Size = %d bytes (expected: 64)" % self.mini_sector_size )
|
|
if self.mini_sector_size not in [64]:
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect mini_sector_size in OLE header")
|
|
if self.reserved1 != 0 or self.reserved2 != 0:
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect OLE header (non-null reserved bytes)")
|
|
log.debug( "Number of Directory sectors = %d" % self.num_dir_sectors )
|
|
# Number of directory sectors (only allowed if DllVersion != 3)
|
|
if self.sector_size==512 and self.num_dir_sectors!=0:
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect number of directory sectors in OLE header")
|
|
log.debug( "Number of FAT sectors = %d" % self.num_fat_sectors )
|
|
# num_fat_sectors = number of FAT sectors in the file
|
|
log.debug( "First Directory sector = %Xh" % self.first_dir_sector )
|
|
# first_dir_sector = 1st sector containing the directory
|
|
log.debug( "Transaction Signature Number = %d" % self.transaction_signature_number )
|
|
# Signature should be zero, BUT some implementations do not follow this
|
|
# rule => only a potential defect:
|
|
# (according to MS-CFB, may be != 0 for applications supporting file
|
|
# transactions)
|
|
if self.transaction_signature_number != 0:
|
|
self._raise_defect(DEFECT_POTENTIAL, "incorrect OLE header (transaction_signature_number>0)")
|
|
log.debug( "Mini Stream cutoff size = %Xh (expected: 1000h)" % self.mini_stream_cutoff_size )
|
|
# MS-CFB: This integer field MUST be set to 0x00001000. This field
|
|
# specifies the maximum size of a user-defined data stream allocated
|
|
# from the mini FAT and mini stream, and that cutoff is 4096 bytes.
|
|
# Any user-defined data stream larger than or equal to this cutoff size
|
|
# must be allocated as normal sectors from the FAT.
|
|
if self.mini_stream_cutoff_size != 0x1000:
|
|
self._raise_defect(DEFECT_INCORRECT, "incorrect mini_stream_cutoff_size in OLE header")
|
|
# if no exception is raised, the cutoff size is fixed to 0x1000
|
|
log.warning('Fixing the mini_stream_cutoff_size to 4096 (mandatory value) instead of %d' %
|
|
self.mini_stream_cutoff_size)
|
|
self.mini_stream_cutoff_size = 0x1000
|
|
# TODO: check if these values are OK
|
|
log.debug( "First MiniFAT sector = %Xh" % self.first_mini_fat_sector )
|
|
log.debug( "Number of MiniFAT sectors = %d" % self.num_mini_fat_sectors )
|
|
log.debug( "First DIFAT sector = %Xh" % self.first_difat_sector )
|
|
log.debug( "Number of DIFAT sectors = %d" % self.num_difat_sectors )
|
|
|
|
# calculate the number of sectors in the file
|
|
# (-1 because header doesn't count)
|
|
self.nb_sect = ( (filesize + self.sector_size-1) // self.sector_size) - 1
|
|
log.debug( "Maximum number of sectors in the file: %d (%Xh)" % (self.nb_sect, self.nb_sect))
|
|
#TODO: change this test, because an OLE file MAY contain other data
|
|
# after the last sector.
|
|
|
|
# file clsid
|
|
self.header_clsid = _clsid(header[8:24])
|
|
|
|
#TODO: remove redundant attributes, and fix the code which uses them?
|
|
self.sectorsize = self.sector_size #1 << i16(header, 30)
|
|
self.minisectorsize = self.mini_sector_size #1 << i16(header, 32)
|
|
self.minisectorcutoff = self.mini_stream_cutoff_size # i32(header, 56)
|
|
|
|
# check known streams for duplicate references (these are always in FAT,
|
|
# never in MiniFAT):
|
|
self._check_duplicate_stream(self.first_dir_sector)
|
|
# check MiniFAT only if it is not empty:
|
|
if self.num_mini_fat_sectors:
|
|
self._check_duplicate_stream(self.first_mini_fat_sector)
|
|
# check DIFAT only if it is not empty:
|
|
if self.num_difat_sectors:
|
|
self._check_duplicate_stream(self.first_difat_sector)
|
|
|
|
# Load file allocation tables
|
|
self.loadfat(header)
|
|
# Load directory. This sets both the direntries list (ordered by sid)
|
|
# and the root (ordered by hierarchy) members.
|
|
self.loaddirectory(self.first_dir_sector)
|
|
self.minifatsect = self.first_mini_fat_sector
|
|
|
|
|
|
def close(self):
|
|
"""
|
|
close the OLE file, to release the file object
|
|
"""
|
|
self.fp.close()
|
|
|
|
|
|
def _check_duplicate_stream(self, first_sect, minifat=False):
|
|
"""
|
|
Checks if a stream has not been already referenced elsewhere.
|
|
This method should only be called once for each known stream, and only
|
|
if stream size is not null.
|
|
|
|
:param first_sect: int, index of first sector of the stream in FAT
|
|
:param minifat: bool, if True, stream is located in the MiniFAT, else in the FAT
|
|
"""
|
|
if minifat:
|
|
log.debug('_check_duplicate_stream: sect=%Xh in MiniFAT' % first_sect)
|
|
used_streams = self._used_streams_minifat
|
|
else:
|
|
log.debug('_check_duplicate_stream: sect=%Xh in FAT' % first_sect)
|
|
# some values can be safely ignored (not a real stream):
|
|
if first_sect in (DIFSECT,FATSECT,ENDOFCHAIN,FREESECT):
|
|
return
|
|
used_streams = self._used_streams_fat
|
|
#TODO: would it be more efficient using a dict or hash values, instead
|
|
# of a list of long ?
|
|
if first_sect in used_streams:
|
|
self._raise_defect(DEFECT_INCORRECT, 'Stream referenced twice')
|
|
else:
|
|
used_streams.append(first_sect)
|
|
|
|
|
|
def dumpfat(self, fat, firstindex=0):
|
|
"""
|
|
Display a part of FAT in human-readable form for debugging purposes
|
|
"""
|
|
# dictionary to convert special FAT values in human-readable strings
|
|
VPL = 8 # values per line (8+1 * 8+1 = 81)
|
|
fatnames = {
|
|
FREESECT: "..free..",
|
|
ENDOFCHAIN: "[ END. ]",
|
|
FATSECT: "FATSECT ",
|
|
DIFSECT: "DIFSECT "
|
|
}
|
|
nbsect = len(fat)
|
|
nlines = (nbsect+VPL-1)//VPL
|
|
print("index", end=" ")
|
|
for i in range(VPL):
|
|
print("%8X" % i, end=" ")
|
|
print()
|
|
for l in range(nlines):
|
|
index = l*VPL
|
|
print("%6X:" % (firstindex+index), end=" ")
|
|
for i in range(index, index+VPL):
|
|
if i>=nbsect:
|
|
break
|
|
sect = fat[i]
|
|
aux = sect & 0xFFFFFFFF # JYTHON-WORKAROUND
|
|
if aux in fatnames:
|
|
name = fatnames[aux]
|
|
else:
|
|
if sect == i+1:
|
|
name = " --->"
|
|
else:
|
|
name = "%8X" % sect
|
|
print(name, end=" ")
|
|
print()
|
|
|
|
|
|
def dumpsect(self, sector, firstindex=0):
|
|
"""
|
|
Display a sector in a human-readable form, for debugging purposes
|
|
"""
|
|
VPL=8 # number of values per line (8+1 * 8+1 = 81)
|
|
tab = array.array(UINT32, sector)
|
|
if sys.byteorder == 'big':
|
|
tab.byteswap()
|
|
nbsect = len(tab)
|
|
nlines = (nbsect+VPL-1)//VPL
|
|
print("index", end=" ")
|
|
for i in range(VPL):
|
|
print("%8X" % i, end=" ")
|
|
print()
|
|
for l in range(nlines):
|
|
index = l*VPL
|
|
print("%6X:" % (firstindex+index), end=" ")
|
|
for i in range(index, index+VPL):
|
|
if i>=nbsect:
|
|
break
|
|
sect = tab[i]
|
|
name = "%8X" % sect
|
|
print(name, end=" ")
|
|
print()
|
|
|
|
def sect2array(self, sect):
|
|
"""
|
|
convert a sector to an array of 32 bits unsigned integers,
|
|
swapping bytes on big endian CPUs such as PowerPC (old Macs)
|
|
"""
|
|
a = array.array(UINT32, sect)
|
|
# if CPU is big endian, swap bytes:
|
|
if sys.byteorder == 'big':
|
|
a.byteswap()
|
|
return a
|
|
|
|
|
|
def loadfat_sect(self, sect):
|
|
"""
|
|
Adds the indexes of the given sector to the FAT
|
|
|
|
:param sect: string containing the first FAT sector, or array of long integers
|
|
:returns: index of last FAT sector.
|
|
"""
|
|
# a FAT sector is an array of ulong integers.
|
|
if isinstance(sect, array.array):
|
|
# if sect is already an array it is directly used
|
|
fat1 = sect
|
|
else:
|
|
# if it's a raw sector, it is parsed in an array
|
|
fat1 = self.sect2array(sect)
|
|
# Display the sector contents only if the logging level is debug:
|
|
if log.isEnabledFor(logging.DEBUG):
|
|
self.dumpsect(sect)
|
|
# The FAT is a sector chain starting at the first index of itself.
|
|
# initialize isect, just in case:
|
|
isect = None
|
|
for isect in fat1:
|
|
isect = isect & 0xFFFFFFFF # JYTHON-WORKAROUND
|
|
log.debug("isect = %X" % isect)
|
|
if isect == ENDOFCHAIN or isect == FREESECT:
|
|
# the end of the sector chain has been reached
|
|
log.debug("found end of sector chain")
|
|
break
|
|
# read the FAT sector
|
|
s = self.getsect(isect)
|
|
# parse it as an array of 32 bits integers, and add it to the
|
|
# global FAT array
|
|
nextfat = self.sect2array(s)
|
|
self.fat = self.fat + nextfat
|
|
return isect
|
|
|
|
|
|
def loadfat(self, header):
|
|
"""
|
|
Load the FAT table.
|
|
"""
|
|
# The 1st sector of the file contains sector numbers for the first 109
|
|
# FAT sectors, right after the header which is 76 bytes long.
|
|
# (always 109, whatever the sector size: 512 bytes = 76+4*109)
|
|
# Additional sectors are described by DIF blocks
|
|
|
|
log.debug('Loading the FAT table, starting with the 1st sector after the header')
|
|
sect = header[76:512]
|
|
log.debug( "len(sect)=%d, so %d integers" % (len(sect), len(sect)//4) )
|
|
#fat = []
|
|
# [PL] FAT is an array of 32 bits unsigned ints, it's more effective
|
|
# to use an array than a list in Python.
|
|
# It's initialized as empty first:
|
|
self.fat = array.array(UINT32)
|
|
self.loadfat_sect(sect)
|
|
#self.dumpfat(self.fat)
|
|
## for i in range(0, len(sect), 4):
|
|
## ix = i32(sect, i)
|
|
## #[PL] if ix == -2 or ix == -1: # ix == 0xFFFFFFFE or ix == 0xFFFFFFFF:
|
|
## if ix == 0xFFFFFFFE or ix == 0xFFFFFFFF:
|
|
## break
|
|
## s = self.getsect(ix)
|
|
## #fat = fat + [i32(s, i) for i in range(0, len(s), 4)]
|
|
## fat = fat + array.array(UINT32, s)
|
|
if self.num_difat_sectors != 0:
|
|
log.debug('DIFAT is used, because file size > 6.8MB.')
|
|
# [PL] There's a DIFAT because file is larger than 6.8MB
|
|
# some checks just in case:
|
|
if self.num_fat_sectors <= 109:
|
|
# there must be at least 109 blocks in header and the rest in
|
|
# DIFAT, so number of sectors must be >109.
|
|
self._raise_defect(DEFECT_INCORRECT, 'incorrect DIFAT, not enough sectors')
|
|
if self.first_difat_sector >= self.nb_sect:
|
|
# initial DIFAT block index must be valid
|
|
self._raise_defect(DEFECT_FATAL, 'incorrect DIFAT, first index out of range')
|
|
log.debug( "DIFAT analysis..." )
|
|
# We compute the necessary number of DIFAT sectors :
|
|
# Number of pointers per DIFAT sector = (sectorsize/4)-1
|
|
# (-1 because the last pointer is the next DIFAT sector number)
|
|
nb_difat_sectors = (self.sectorsize//4)-1
|
|
# (if 512 bytes: each DIFAT sector = 127 pointers + 1 towards next DIFAT sector)
|
|
nb_difat = (self.num_fat_sectors-109 + nb_difat_sectors-1)//nb_difat_sectors
|
|
log.debug( "nb_difat = %d" % nb_difat )
|
|
if self.num_difat_sectors != nb_difat:
|
|
raise IOError('incorrect DIFAT')
|
|
isect_difat = self.first_difat_sector
|
|
for i in iterrange(nb_difat):
|
|
log.debug( "DIFAT block %d, sector %X" % (i, isect_difat) )
|
|
#TODO: check if corresponding FAT SID = DIFSECT
|
|
sector_difat = self.getsect(isect_difat)
|
|
difat = self.sect2array(sector_difat)
|
|
# Display the sector contents only if the logging level is debug:
|
|
if log.isEnabledFor(logging.DEBUG):
|
|
self.dumpsect(sector_difat)
|
|
self.loadfat_sect(difat[:nb_difat_sectors])
|
|
# last DIFAT pointer is next DIFAT sector:
|
|
isect_difat = difat[nb_difat_sectors]
|
|
log.debug( "next DIFAT sector: %X" % isect_difat )
|
|
# checks:
|
|
if isect_difat not in [ENDOFCHAIN, FREESECT]:
|
|
# last DIFAT pointer value must be ENDOFCHAIN or FREESECT
|
|
raise IOError('incorrect end of DIFAT')
|
|
## if len(self.fat) != self.num_fat_sectors:
|
|
## # FAT should contain num_fat_sectors blocks
|
|
## print("FAT length: %d instead of %d" % (len(self.fat), self.num_fat_sectors))
|
|
## raise IOError('incorrect DIFAT')
|
|
else:
|
|
log.debug('No DIFAT, because file size < 6.8MB.')
|
|
# since FAT is read from fixed-size sectors, it may contain more values
|
|
# than the actual number of sectors in the file.
|
|
# Keep only the relevant sector indexes:
|
|
if len(self.fat) > self.nb_sect:
|
|
log.debug('len(fat)=%d, shrunk to nb_sect=%d' % (len(self.fat), self.nb_sect))
|
|
self.fat = self.fat[:self.nb_sect]
|
|
log.debug('FAT references %d sectors / Maximum %d sectors in file' % (len(self.fat), self.nb_sect))
|
|
# Display the FAT contents only if the logging level is debug:
|
|
if log.isEnabledFor(logging.DEBUG):
|
|
log.debug('\nFAT:')
|
|
self.dumpfat(self.fat)
|
|
|
|
|
|
def loadminifat(self):
|
|
"""
|
|
Load the MiniFAT table.
|
|
"""
|
|
# MiniFAT is stored in a standard sub-stream, pointed to by a header
|
|
# field.
|
|
# NOTE: there are two sizes to take into account for this stream:
|
|
# 1) Stream size is calculated according to the number of sectors
|
|
# declared in the OLE header. This allocated stream may be more than
|
|
# needed to store the actual sector indexes.
|
|
# (self.num_mini_fat_sectors is the number of sectors of size self.sector_size)
|
|
stream_size = self.num_mini_fat_sectors * self.sector_size
|
|
# 2) Actually used size is calculated by dividing the MiniStream size
|
|
# (given by root entry size) by the size of mini sectors, *4 for
|
|
# 32 bits indexes:
|
|
nb_minisectors = (self.root.size + self.mini_sector_size-1) // self.mini_sector_size
|
|
used_size = nb_minisectors * 4
|
|
log.debug('loadminifat(): minifatsect=%d, nb FAT sectors=%d, used_size=%d, stream_size=%d, nb MiniSectors=%d' %
|
|
(self.minifatsect, self.num_mini_fat_sectors, used_size, stream_size, nb_minisectors))
|
|
if used_size > stream_size:
|
|
# This is not really a problem, but may indicate a wrong implementation:
|
|
self._raise_defect(DEFECT_INCORRECT, 'OLE MiniStream is larger than MiniFAT')
|
|
# In any case, first read stream_size:
|
|
s = self._open(self.minifatsect, stream_size, force_FAT=True).read()
|
|
#[PL] Old code replaced by an array:
|
|
#self.minifat = [i32(s, i) for i in range(0, len(s), 4)]
|
|
self.minifat = self.sect2array(s)
|
|
# Then shrink the array to used size, to avoid indexes out of MiniStream:
|
|
log.debug('MiniFAT shrunk from %d to %d sectors' % (len(self.minifat), nb_minisectors))
|
|
self.minifat = self.minifat[:nb_minisectors]
|
|
log.debug('loadminifat(): len=%d' % len(self.minifat))
|
|
# Display the FAT contents only if the logging level is debug:
|
|
if log.isEnabledFor(logging.DEBUG):
|
|
log.debug('\nMiniFAT:')
|
|
self.dumpfat(self.minifat)
|
|
|
|
def getsect(self, sect):
|
|
"""
|
|
Read given sector from file on disk.
|
|
|
|
:param sect: int, sector index
|
|
:returns: a string containing the sector data.
|
|
"""
|
|
# From [MS-CFB]: A sector number can be converted into a byte offset
|
|
# into the file by using the following formula:
|
|
# (sector number + 1) x Sector Size.
|
|
# This implies that sector #0 of the file begins at byte offset Sector
|
|
# Size, not at 0.
|
|
|
|
# [PL] the original code in PIL was wrong when sectors are 4KB instead of
|
|
# 512 bytes:
|
|
#self.fp.seek(512 + self.sectorsize * sect)
|
|
#[PL]: added safety checks:
|
|
#print("getsect(%X)" % sect)
|
|
try:
|
|
self.fp.seek(self.sectorsize * (sect+1))
|
|
except:
|
|
log.debug('getsect(): sect=%X, seek=%d, filesize=%d' %
|
|
(sect, self.sectorsize*(sect+1), self._filesize))
|
|
self._raise_defect(DEFECT_FATAL, 'OLE sector index out of range')
|
|
sector = self.fp.read(self.sectorsize)
|
|
if len(sector) != self.sectorsize:
|
|
log.debug('getsect(): sect=%X, read=%d, sectorsize=%d' %
|
|
(sect, len(sector), self.sectorsize))
|
|
self._raise_defect(DEFECT_FATAL, 'incomplete OLE sector')
|
|
return sector
|
|
|
|
|
|
def write_sect(self, sect, data, padding=b'\x00'):
|
|
"""
|
|
Write given sector to file on disk.
|
|
|
|
:param sect: int, sector index
|
|
:param data: bytes, sector data
|
|
:param padding: single byte, padding character if data < sector size
|
|
"""
|
|
if not isinstance(data, bytes):
|
|
raise TypeError("write_sect: data must be a bytes string")
|
|
if not isinstance(padding, bytes) or len(padding)!=1:
|
|
raise TypeError("write_sect: padding must be a bytes string of 1 char")
|
|
#TODO: we could allow padding=None for no padding at all
|
|
try:
|
|
self.fp.seek(self.sectorsize * (sect+1))
|
|
except:
|
|
log.debug('write_sect(): sect=%X, seek=%d, filesize=%d' %
|
|
(sect, self.sectorsize*(sect+1), self._filesize))
|
|
self._raise_defect(DEFECT_FATAL, 'OLE sector index out of range')
|
|
if len(data) < self.sectorsize:
|
|
# add padding
|
|
data += padding * (self.sectorsize - len(data))
|
|
elif len(data) < self.sectorsize:
|
|
raise ValueError("Data is larger than sector size")
|
|
self.fp.write(data)
|
|
|
|
def _write_mini_sect(self, fp_pos, data, padding = b'\x00'):
|
|
"""
|
|
Write given sector to file on disk.
|
|
|
|
:param fp_pos: int, file position
|
|
:param data: bytes, sector data
|
|
:param padding: single byte, padding character if data < sector size
|
|
"""
|
|
if not isinstance(data, bytes):
|
|
raise TypeError("write_mini_sect: data must be a bytes string")
|
|
if not isinstance(padding, bytes) or len(padding) != 1:
|
|
raise TypeError("write_mini_sect: padding must be a bytes string of 1 char")
|
|
|
|
try:
|
|
self.fp.seek(fp_pos)
|
|
except:
|
|
log.debug('write_mini_sect(): fp_pos=%d, filesize=%d' %
|
|
(fp_pos, self._filesize))
|
|
self._raise_defect(DEFECT_FATAL, 'OLE sector index out of range')
|
|
len_data = len(data)
|
|
if len_data < self.mini_sector_size:
|
|
data += padding * (self.mini_sector_size - len_data)
|
|
if self.mini_sector_size < len_data:
|
|
raise ValueError("Data is larger than sector size")
|
|
self.fp.write(data)
|
|
|
|
def loaddirectory(self, sect):
|
|
"""
|
|
Load the directory.
|
|
|
|
:param sect: sector index of directory stream.
|
|
"""
|
|
log.debug('Loading the Directory:')
|
|
# The directory is stored in a standard
|
|
# substream, independent of its size.
|
|
|
|
# open directory stream as a read-only file:
|
|
# (stream size is not known in advance)
|
|
self.directory_fp = self._open(sect, force_FAT=True)
|
|
|
|
#[PL] to detect malformed documents and avoid DoS attacks, the maximum
|
|
# number of directory entries can be calculated:
|
|
max_entries = self.directory_fp.size // 128
|
|
log.debug('loaddirectory: size=%d, max_entries=%d' %
|
|
(self.directory_fp.size, max_entries))
|
|
|
|
# Create list of directory entries
|
|
#self.direntries = []
|
|
# We start with a list of "None" object
|
|
self.direntries = [None] * max_entries
|
|
## for sid in iterrange(max_entries):
|
|
## entry = fp.read(128)
|
|
## if not entry:
|
|
## break
|
|
## self.direntries.append(OleDirectoryEntry(entry, sid, self))
|
|
# load root entry:
|
|
root_entry = self._load_direntry(0)
|
|
# Root entry is the first entry:
|
|
self.root = self.direntries[0]
|
|
# TODO: read ALL directory entries (ignore bad entries?)
|
|
# TODO: adapt build_storage_tree to avoid duplicate reads
|
|
# for i in range(1, max_entries):
|
|
# self._load_direntry(i)
|
|
# read and build all storage trees, starting from the root:
|
|
self.root.build_storage_tree()
|
|
|
|
|
|
def _load_direntry (self, sid):
|
|
"""
|
|
Load a directory entry from the directory.
|
|
This method should only be called once for each storage/stream when
|
|
loading the directory.
|
|
|
|
:param sid: index of storage/stream in the directory.
|
|
:returns: a OleDirectoryEntry object
|
|
|
|
:exception IOError: if the entry has always been referenced.
|
|
"""
|
|
# check if SID is OK:
|
|
if sid<0 or sid>=len(self.direntries):
|
|
self._raise_defect(DEFECT_FATAL, "OLE directory index out of range")
|
|
# check if entry was already referenced:
|
|
if self.direntries[sid] is not None:
|
|
self._raise_defect(DEFECT_INCORRECT,
|
|
"double reference for OLE stream/storage")
|
|
# if exception not raised, return the object
|
|
return self.direntries[sid]
|
|
self.directory_fp.seek(sid * 128)
|
|
entry = self.directory_fp.read(128)
|
|
self.direntries[sid] = OleDirectoryEntry(entry, sid, self)
|
|
return self.direntries[sid]
|
|
|
|
|
|
def dumpdirectory(self):
|
|
"""
|
|
Dump directory (for debugging only)
|
|
"""
|
|
self.root.dump()
|
|
|
|
|
|
def _open(self, start, size = UNKNOWN_SIZE, force_FAT=False):
|
|
"""
|
|
Open a stream, either in FAT or MiniFAT according to its size.
|
|
(openstream helper)
|
|
|
|
:param start: index of first sector
|
|
:param size: size of stream (or nothing if size is unknown)
|
|
:param force_FAT: if False (default), stream will be opened in FAT or MiniFAT
|
|
according to size. If True, it will always be opened in FAT.
|
|
"""
|
|
log.debug('OleFileIO.open(): sect=%Xh, size=%d, force_FAT=%s' %
|
|
(start, size, str(force_FAT)))
|
|
# stream size is compared to the mini_stream_cutoff_size threshold:
|
|
if size < self.minisectorcutoff and not force_FAT:
|
|
# ministream object
|
|
if not self.ministream:
|
|
# load MiniFAT if it wasn't already done:
|
|
self.loadminifat()
|
|
# The first sector index of the miniFAT stream is stored in the
|
|
# root directory entry:
|
|
size_ministream = self.root.size
|
|
log.debug('Opening MiniStream: sect=%Xh, size=%d' %
|
|
(self.root.isectStart, size_ministream))
|
|
self.ministream = self._open(self.root.isectStart,
|
|
size_ministream, force_FAT=True)
|
|
return OleStream(fp=self.ministream, sect=start, size=size,
|
|
offset=0, sectorsize=self.minisectorsize,
|
|
fat=self.minifat, filesize=self.ministream.size,
|
|
olefileio=self)
|
|
else:
|
|
# standard stream
|
|
return OleStream(fp=self.fp, sect=start, size=size,
|
|
offset=self.sectorsize,
|
|
sectorsize=self.sectorsize, fat=self.fat,
|
|
filesize=self._filesize,
|
|
olefileio=self)
|
|
|
|
|
|
def _list(self, files, prefix, node, streams=True, storages=False):
|
|
"""
|
|
listdir helper
|
|
|
|
:param files: list of files to fill in
|
|
:param prefix: current location in storage tree (list of names)
|
|
:param node: current node (OleDirectoryEntry object)
|
|
:param streams: bool, include streams if True (True by default) - new in v0.26
|
|
:param storages: bool, include storages if True (False by default) - new in v0.26
|
|
(note: the root storage is never included)
|
|
"""
|
|
prefix = prefix + [node.name]
|
|
for entry in node.kids:
|
|
if entry.entry_type == STGTY_STORAGE:
|
|
# this is a storage
|
|
if storages:
|
|
# add it to the list
|
|
files.append(prefix[1:] + [entry.name])
|
|
# check its kids
|
|
self._list(files, prefix, entry, streams, storages)
|
|
elif entry.entry_type == STGTY_STREAM:
|
|
# this is a stream
|
|
if streams:
|
|
# add it to the list
|
|
files.append(prefix[1:] + [entry.name])
|
|
else:
|
|
self._raise_defect(DEFECT_INCORRECT, 'The directory tree contains an entry which is not a stream nor a storage.')
|
|
|
|
|
|
def listdir(self, streams=True, storages=False):
|
|
"""
|
|
Return a list of streams and/or storages stored in this file
|
|
|
|
:param streams: bool, include streams if True (True by default) - new in v0.26
|
|
:param storages: bool, include storages if True (False by default) - new in v0.26
|
|
(note: the root storage is never included)
|
|
:returns: list of stream and/or storage paths
|
|
"""
|
|
files = []
|
|
self._list(files, [], self.root, streams, storages)
|
|
return files
|
|
|
|
|
|
def _find(self, filename):
|
|
"""
|
|
Returns directory entry of given filename. (openstream helper)
|
|
Note: this method is case-insensitive.
|
|
|
|
:param filename: path of stream in storage tree (except root entry), either:
|
|
|
|
- a string using Unix path syntax, for example:
|
|
'storage_1/storage_1.2/stream'
|
|
- or a list of storage filenames, path to the desired stream/storage.
|
|
Example: ['storage_1', 'storage_1.2', 'stream']
|
|
|
|
:returns: sid of requested filename
|
|
:exception IOError: if file not found
|
|
"""
|
|
|
|
# if filename is a string instead of a list, split it on slashes to
|
|
# convert to a list:
|
|
if isinstance(filename, basestring):
|
|
filename = filename.split('/')
|
|
# walk across storage tree, following given path:
|
|
node = self.root
|
|
for name in filename:
|
|
for kid in node.kids:
|
|
if kid.name.lower() == name.lower():
|
|
break
|
|
else:
|
|
raise IOError("file not found")
|
|
node = kid
|
|
return node.sid
|
|
|
|
|
|
def openstream(self, filename):
|
|
"""
|
|
Open a stream as a read-only file object (BytesIO).
|
|
Note: filename is case-insensitive.
|
|
|
|
:param filename: path of stream in storage tree (except root entry), either:
|
|
|
|
- a string using Unix path syntax, for example:
|
|
'storage_1/storage_1.2/stream'
|
|
- or a list of storage filenames, path to the desired stream/storage.
|
|
Example: ['storage_1', 'storage_1.2', 'stream']
|
|
|
|
:returns: file object (read-only)
|
|
:exception IOError: if filename not found, or if this is not a stream.
|
|
"""
|
|
sid = self._find(filename)
|
|
entry = self.direntries[sid]
|
|
if entry.entry_type != STGTY_STREAM:
|
|
raise IOError("this file is not a stream")
|
|
return self._open(entry.isectStart, entry.size)
|
|
|
|
def _write_mini_stream(self, entry, data_to_write):
|
|
if not entry.sect_chain:
|
|
entry.build_sect_chain(self)
|
|
nb_sectors = len(entry.sect_chain)
|
|
|
|
if not self.root.sect_chain:
|
|
self.root.build_sect_chain(self)
|
|
block_size = self.sector_size // self.mini_sector_size
|
|
for idx, sect in enumerate(entry.sect_chain):
|
|
sect_base = sect // block_size
|
|
sect_offset = sect % block_size
|
|
fp_pos = (self.root.sect_chain[sect_base] + 1)*self.sector_size + sect_offset*self.mini_sector_size
|
|
if idx < (nb_sectors - 1):
|
|
data_per_sector = data_to_write[idx * self.mini_sector_size: (idx + 1) * self.mini_sector_size]
|
|
else:
|
|
data_per_sector = data_to_write[idx * self.mini_sector_size:]
|
|
self._write_mini_sect(fp_pos, data_per_sector)
|
|
|
|
def write_stream(self, stream_name, data):
|
|
"""
|
|
Write a stream to disk. For now, it is only possible to replace an
|
|
existing stream by data of the same size.
|
|
|
|
:param stream_name: path of stream in storage tree (except root entry), either:
|
|
|
|
- a string using Unix path syntax, for example:
|
|
'storage_1/storage_1.2/stream'
|
|
- or a list of storage filenames, path to the desired stream/storage.
|
|
Example: ['storage_1', 'storage_1.2', 'stream']
|
|
|
|
:param data: bytes, data to be written, must be the same size as the original
|
|
stream.
|
|
"""
|
|
if not isinstance(data, bytes):
|
|
raise TypeError("write_stream: data must be a bytes string")
|
|
sid = self._find(stream_name)
|
|
entry = self.direntries[sid]
|
|
if entry.entry_type != STGTY_STREAM:
|
|
raise IOError("this is not a stream")
|
|
size = entry.size
|
|
if size != len(data):
|
|
raise ValueError("write_stream: data must be the same size as the existing stream")
|
|
if size < self.minisectorcutoff and entry.entry_type != STGTY_ROOT:
|
|
return self._write_mini_stream(entry = entry, data_to_write = data)
|
|
|
|
sect = entry.isectStart
|
|
# number of sectors to write
|
|
nb_sectors = (size + (self.sectorsize-1)) // self.sectorsize
|
|
log.debug('nb_sectors = %d' % nb_sectors)
|
|
for i in range(nb_sectors):
|
|
## try:
|
|
## self.fp.seek(offset + self.sectorsize * sect)
|
|
## except:
|
|
## log.debug('sect=%d, seek=%d' %
|
|
## (sect, offset+self.sectorsize*sect))
|
|
## raise IOError('OLE sector index out of range')
|
|
# extract one sector from data, the last one being smaller:
|
|
if i<(nb_sectors-1):
|
|
data_sector = data [i*self.sectorsize : (i+1)*self.sectorsize]
|
|
#TODO: comment this if it works
|
|
assert(len(data_sector)==self.sectorsize)
|
|
else:
|
|
data_sector = data [i*self.sectorsize:]
|
|
#TODO: comment this if it works
|
|
log.debug('write_stream: size=%d sectorsize=%d data_sector=%Xh size%%sectorsize=%d'
|
|
% (size, self.sectorsize, len(data_sector), size % self.sectorsize))
|
|
assert(len(data_sector) % self.sectorsize==size % self.sectorsize)
|
|
self.write_sect(sect, data_sector)
|
|
## self.fp.write(data_sector)
|
|
# jump to next sector in the FAT:
|
|
try:
|
|
sect = self.fat[sect]
|
|
except IndexError:
|
|
# [PL] if pointer is out of the FAT an exception is raised
|
|
raise IOError('incorrect OLE FAT, sector index out of range')
|
|
#[PL] Last sector should be a "end of chain" marker:
|
|
if sect != ENDOFCHAIN:
|
|
raise IOError('incorrect last sector index in OLE stream')
|
|
|
|
|
|
def get_type(self, filename):
|
|
"""
|
|
Test if given filename exists as a stream or a storage in the OLE
|
|
container, and return its type.
|
|
|
|
:param filename: path of stream in storage tree. (see openstream for syntax)
|
|
:returns: False if object does not exist, its entry type (>0) otherwise:
|
|
|
|
- STGTY_STREAM: a stream
|
|
- STGTY_STORAGE: a storage
|
|
- STGTY_ROOT: the root entry
|
|
"""
|
|
try:
|
|
sid = self._find(filename)
|
|
entry = self.direntries[sid]
|
|
return entry.entry_type
|
|
except:
|
|
return False
|
|
|
|
|
|
def getclsid(self, filename):
|
|
"""
|
|
Return clsid of a stream/storage.
|
|
|
|
:param filename: path of stream/storage in storage tree. (see openstream for
|
|
syntax)
|
|
:returns: Empty string if clsid is null, a printable representation of the clsid otherwise
|
|
|
|
new in version 0.44
|
|
"""
|
|
sid = self._find(filename)
|
|
entry = self.direntries[sid]
|
|
return entry.clsid
|
|
|
|
|
|
def getmtime(self, filename):
|
|
"""
|
|
Return modification time of a stream/storage.
|
|
|
|
:param filename: path of stream/storage in storage tree. (see openstream for
|
|
syntax)
|
|
:returns: None if modification time is null, a python datetime object
|
|
otherwise (UTC timezone)
|
|
|
|
new in version 0.26
|
|
"""
|
|
sid = self._find(filename)
|
|
entry = self.direntries[sid]
|
|
return entry.getmtime()
|
|
|
|
|
|
def getctime(self, filename):
|
|
"""
|
|
Return creation time of a stream/storage.
|
|
|
|
:param filename: path of stream/storage in storage tree. (see openstream for
|
|
syntax)
|
|
:returns: None if creation time is null, a python datetime object
|
|
otherwise (UTC timezone)
|
|
|
|
new in version 0.26
|
|
"""
|
|
sid = self._find(filename)
|
|
entry = self.direntries[sid]
|
|
return entry.getctime()
|
|
|
|
|
|
def exists(self, filename):
|
|
"""
|
|
Test if given filename exists as a stream or a storage in the OLE
|
|
container.
|
|
Note: filename is case-insensitive.
|
|
|
|
:param filename: path of stream in storage tree. (see openstream for syntax)
|
|
:returns: True if object exist, else False.
|
|
"""
|
|
try:
|
|
sid = self._find(filename)
|
|
return True
|
|
except:
|
|
return False
|
|
|
|
|
|
def get_size(self, filename):
|
|
"""
|
|
Return size of a stream in the OLE container, in bytes.
|
|
|
|
:param filename: path of stream in storage tree (see openstream for syntax)
|
|
:returns: size in bytes (long integer)
|
|
:exception IOError: if file not found
|
|
:exception TypeError: if this is not a stream.
|
|
"""
|
|
sid = self._find(filename)
|
|
entry = self.direntries[sid]
|
|
if entry.entry_type != STGTY_STREAM:
|
|
#TODO: Should it return zero instead of raising an exception ?
|
|
raise TypeError('object is not an OLE stream')
|
|
return entry.size
|
|
|
|
|
|
def get_rootentry_name(self):
|
|
"""
|
|
Return root entry name. Should usually be 'Root Entry' or 'R' in most
|
|
implementations.
|
|
"""
|
|
return self.root.name
|
|
|
|
|
|
def getproperties(self, filename, convert_time=False, no_conversion=None):
|
|
"""
|
|
Return properties described in substream.
|
|
|
|
:param filename: path of stream in storage tree (see openstream for syntax)
|
|
:param convert_time: bool, if True timestamps will be converted to Python datetime
|
|
:param no_conversion: None or list of int, timestamps not to be converted
|
|
(for example total editing time is not a real timestamp)
|
|
|
|
:returns: a dictionary of values indexed by id (integer)
|
|
"""
|
|
#REFERENCE: [MS-OLEPS] https://msdn.microsoft.com/en-us/library/dd942421.aspx
|
|
# make sure no_conversion is a list, just to simplify code below:
|
|
if no_conversion == None:
|
|
no_conversion = []
|
|
# stream path as a string to report exceptions:
|
|
streampath = filename
|
|
if not isinstance(streampath, str):
|
|
streampath = '/'.join(streampath)
|
|
|
|
fp = self.openstream(filename)
|
|
|
|
data = {}
|
|
|
|
try:
|
|
# header
|
|
s = fp.read(28)
|
|
clsid = _clsid(s[8:24])
|
|
|
|
# format id
|
|
s = fp.read(20)
|
|
fmtid = _clsid(s[:16])
|
|
fp.seek(i32(s, 16))
|
|
|
|
# get section
|
|
s = b"****" + fp.read(i32(fp.read(4))-4)
|
|
# number of properties:
|
|
num_props = i32(s, 4)
|
|
except BaseException as exc:
|
|
# catch exception while parsing property header, and only raise
|
|
# a DEFECT_INCORRECT then return an empty dict, because this is not
|
|
# a fatal error when parsing the whole file
|
|
msg = 'Error while parsing properties header in stream %s: %s' % (
|
|
repr(streampath), exc)
|
|
self._raise_defect(DEFECT_INCORRECT, msg, type(exc))
|
|
return data
|
|
|
|
# clamp num_props based on the data length
|
|
num_props = min(num_props, int(len(s) / 8))
|
|
|
|
for i in iterrange(num_props):
|
|
property_id = 0 # just in case of an exception
|
|
try:
|
|
property_id = i32(s, 8+i*8)
|
|
offset = i32(s, 12+i*8)
|
|
property_type = i32(s, offset)
|
|
|
|
log.debug('property id=%d: type=%d offset=%X' % (property_id, property_type, offset))
|
|
|
|
# test for common types first (should perhaps use
|
|
# a dictionary instead?)
|
|
|
|
if property_type == VT_I2: # 16-bit signed integer
|
|
value = i16(s, offset+4)
|
|
if value >= 32768:
|
|
value = value - 65536
|
|
elif property_type == VT_UI2: # 2-byte unsigned integer
|
|
value = i16(s, offset+4)
|
|
elif property_type in (VT_I4, VT_INT, VT_ERROR):
|
|
# VT_I4: 32-bit signed integer
|
|
# VT_ERROR: HRESULT, similar to 32-bit signed integer,
|
|
# see https://msdn.microsoft.com/en-us/library/cc230330.aspx
|
|
value = i32(s, offset+4)
|
|
elif property_type in (VT_UI4, VT_UINT): # 4-byte unsigned integer
|
|
value = i32(s, offset+4) # FIXME
|
|
elif property_type in (VT_BSTR, VT_LPSTR):
|
|
# CodePageString, see https://msdn.microsoft.com/en-us/library/dd942354.aspx
|
|
# size is a 32 bits integer, including the null terminator, and
|
|
# possibly trailing or embedded null chars
|
|
#TODO: if codepage is unicode, the string should be converted as such
|
|
count = i32(s, offset+4)
|
|
value = s[offset+8:offset+8+count-1]
|
|
# remove all null chars:
|
|
value = value.replace(b'\x00', b'')
|
|
elif property_type == VT_BLOB:
|
|
# binary large object (BLOB)
|
|
# see https://msdn.microsoft.com/en-us/library/dd942282.aspx
|
|
count = i32(s, offset+4)
|
|
value = s[offset+8:offset+8+count]
|
|
elif property_type == VT_LPWSTR:
|
|
# UnicodeString
|
|
# see https://msdn.microsoft.com/en-us/library/dd942313.aspx
|
|
# "the string should NOT contain embedded or additional trailing
|
|
# null characters."
|
|
count = i32(s, offset+4)
|
|
value = self._decode_utf16_str(s[offset+8:offset+8+count*2])
|
|
elif property_type == VT_FILETIME:
|
|
value = long(i32(s, offset+4)) + (long(i32(s, offset+8))<<32)
|
|
# FILETIME is a 64-bit int: "number of 100ns periods
|
|
# since Jan 1,1601".
|
|
if convert_time and property_id not in no_conversion:
|
|
log.debug('Converting property #%d to python datetime, value=%d=%fs'
|
|
%(property_id, value, float(value)/10000000))
|
|
# convert FILETIME to Python datetime.datetime
|
|
# inspired from https://code.activestate.com/recipes/511425-filetime-to-datetime/
|
|
_FILETIME_null_date = datetime.datetime(1601, 1, 1, 0, 0, 0)
|
|
log.debug('timedelta days=%d' % (value//(10*1000000*3600*24)))
|
|
value = _FILETIME_null_date + datetime.timedelta(microseconds=value//10)
|
|
else:
|
|
# legacy code kept for backward compatibility: returns a
|
|
# number of seconds since Jan 1,1601
|
|
value = value // 10000000 # seconds
|
|
elif property_type == VT_UI1: # 1-byte unsigned integer
|
|
value = i8(s[offset+4])
|
|
elif property_type == VT_CLSID:
|
|
value = _clsid(s[offset+4:offset+20])
|
|
elif property_type == VT_CF:
|
|
# PropertyIdentifier or ClipboardData??
|
|
# see https://msdn.microsoft.com/en-us/library/dd941945.aspx
|
|
count = i32(s, offset+4)
|
|
value = s[offset+8:offset+8+count]
|
|
elif property_type == VT_BOOL:
|
|
# VARIANT_BOOL, 16 bits bool, 0x0000=Fals, 0xFFFF=True
|
|
# see https://msdn.microsoft.com/en-us/library/cc237864.aspx
|
|
value = bool(i16(s, offset+4))
|
|
else:
|
|
value = None # everything else yields "None"
|
|
log.debug('property id=%d: type=%d not implemented in parser yet' % (property_id, property_type))
|
|
|
|
# missing: VT_EMPTY, VT_NULL, VT_R4, VT_R8, VT_CY, VT_DATE,
|
|
# VT_DECIMAL, VT_I1, VT_I8, VT_UI8,
|
|
# see https://msdn.microsoft.com/en-us/library/dd942033.aspx
|
|
|
|
# FIXME: add support for VT_VECTOR
|
|
# VT_VECTOR is a 32 uint giving the number of items, followed by
|
|
# the items in sequence. The VT_VECTOR value is combined with the
|
|
# type of items, e.g. VT_VECTOR|VT_BSTR
|
|
# see https://msdn.microsoft.com/en-us/library/dd942011.aspx
|
|
|
|
#print("%08x" % property_id, repr(value), end=" ")
|
|
#print("(%s)" % VT[i32(s, offset) & 0xFFF])
|
|
|
|
data[property_id] = value
|
|
except BaseException as exc:
|
|
# catch exception while parsing each property, and only raise
|
|
# a DEFECT_INCORRECT, because parsing can go on
|
|
msg = 'Error while parsing property id %d in stream %s: %s' % (
|
|
property_id, repr(streampath), exc)
|
|
self._raise_defect(DEFECT_INCORRECT, msg, type(exc))
|
|
|
|
return data
|
|
|
|
def get_metadata(self):
|
|
"""
|
|
Parse standard properties streams, return an OleMetadata object
|
|
containing all the available metadata.
|
|
(also stored in the metadata attribute of the OleFileIO object)
|
|
|
|
new in version 0.25
|
|
"""
|
|
self.metadata = OleMetadata()
|
|
self.metadata.parse_properties(self)
|
|
return self.metadata
|
|
|
|
#
|
|
# --------------------------------------------------------------------
|
|
# This script can be used to dump the directory of any OLE2 structured
|
|
# storage file.
|
|
|
|
def main():
|
|
"""
|
|
Main function when olefile is runs as a script from the command line.
|
|
This will open an OLE2 file and display its structure and properties
|
|
:return: nothing
|
|
"""
|
|
import sys, optparse
|
|
|
|
DEFAULT_LOG_LEVEL = "warning" # Default log level
|
|
LOG_LEVELS = {
|
|
'debug': logging.DEBUG,
|
|
'info': logging.INFO,
|
|
'warning': logging.WARNING,
|
|
'error': logging.ERROR,
|
|
'critical': logging.CRITICAL
|
|
}
|
|
|
|
usage = 'usage: %prog [options] <filename> [filename2 ...]'
|
|
parser = optparse.OptionParser(usage=usage)
|
|
parser.add_option("-c", action="store_true", dest="check_streams",
|
|
help='check all streams (for debugging purposes)')
|
|
parser.add_option("-d", action="store_true", dest="debug_mode",
|
|
help='debug mode, shortcut for -l debug (displays a lot of debug information, for developers only)')
|
|
parser.add_option('-l', '--loglevel', dest="loglevel", action="store", default=DEFAULT_LOG_LEVEL,
|
|
help="logging level debug/info/warning/error/critical (default=%default)")
|
|
|
|
(options, args) = parser.parse_args()
|
|
|
|
print('olefile version %s %s - https://www.decalage.info/en/olefile\n' % (__version__, __date__))
|
|
|
|
# Print help if no arguments are passed
|
|
if len(args) == 0:
|
|
print(__doc__)
|
|
parser.print_help()
|
|
sys.exit()
|
|
|
|
if options.debug_mode:
|
|
options.loglevel = 'debug'
|
|
|
|
# setup logging to the console
|
|
logging.basicConfig(level=LOG_LEVELS[options.loglevel], format='%(levelname)-8s %(message)s')
|
|
|
|
# also enable the module's logger:
|
|
enable_logging()
|
|
|
|
for filename in args:
|
|
try:
|
|
ole = OleFileIO(filename)#, raise_defects=DEFECT_INCORRECT)
|
|
print("-" * 68)
|
|
print(filename)
|
|
print("-" * 68)
|
|
ole.dumpdirectory()
|
|
for streamname in ole.listdir():
|
|
if streamname[-1][0] == "\005":
|
|
print("%r: properties" % streamname)
|
|
try:
|
|
props = ole.getproperties(streamname, convert_time=True)
|
|
props = sorted(props.items())
|
|
for k, v in props:
|
|
#[PL]: avoid to display too large or binary values:
|
|
if isinstance(v, (basestring, bytes)):
|
|
if len(v) > 50:
|
|
v = v[:50]
|
|
if isinstance(v, bytes):
|
|
# quick and dirty binary check:
|
|
for c in (1,2,3,4,5,6,7,11,12,14,15,16,17,18,19,20,
|
|
21,22,23,24,25,26,27,28,29,30,31):
|
|
if c in bytearray(v):
|
|
v = '(binary data)'
|
|
break
|
|
print(" ", k, v)
|
|
except:
|
|
log.exception('Error while parsing property stream %r' % streamname)
|
|
|
|
if options.check_streams:
|
|
# Read all streams to check if there are errors:
|
|
print('\nChecking streams...')
|
|
for streamname in ole.listdir():
|
|
# print name using repr() to convert binary chars to \xNN:
|
|
print('-', repr('/'.join(streamname)),'-', end=' ')
|
|
st_type = ole.get_type(streamname)
|
|
if st_type == STGTY_STREAM:
|
|
print('size %d' % ole.get_size(streamname))
|
|
# just try to read stream in memory:
|
|
ole.openstream(streamname)
|
|
else:
|
|
print('NOT a stream : type=%d' % st_type)
|
|
print()
|
|
|
|
## for streamname in ole.listdir():
|
|
## # print name using repr() to convert binary chars to \xNN:
|
|
## print('-', repr('/'.join(streamname)),'-', end=' ')
|
|
## print(ole.getmtime(streamname))
|
|
## print()
|
|
|
|
print('Modification/Creation times of all directory entries:')
|
|
for entry in ole.direntries:
|
|
if entry is not None:
|
|
print('- %s: mtime=%s ctime=%s' % (entry.name,
|
|
entry.getmtime(), entry.getctime()))
|
|
print()
|
|
|
|
# parse and display metadata:
|
|
try:
|
|
meta = ole.get_metadata()
|
|
meta.dump()
|
|
except:
|
|
log.exception('Error while parsing metadata')
|
|
print()
|
|
#[PL] Test a few new methods:
|
|
root = ole.get_rootentry_name()
|
|
print('Root entry name: "%s"' % root)
|
|
if ole.exists('worddocument'):
|
|
print("This is a Word document.")
|
|
print("type of stream 'WordDocument':", ole.get_type('worddocument'))
|
|
print("size :", ole.get_size('worddocument'))
|
|
if ole.exists('macros/vba'):
|
|
print("This document may contain VBA macros.")
|
|
|
|
# print parsing issues:
|
|
print('\nNon-fatal issues raised during parsing:')
|
|
if ole.parsing_issues:
|
|
for exctype, msg in ole.parsing_issues:
|
|
print('- %s: %s' % (exctype.__name__, msg))
|
|
else:
|
|
print('None')
|
|
except:
|
|
log.exception('Error while parsing file %r' % filename)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|
|
|
|
# this code was developed while listening to The Wedding Present "Sea Monsters"
|