mirror of
https://github.com/noDRM/DeDRM_tools
synced 2024-12-28 22:24:27 +01:00
topazscripts 1.6
This commit is contained in:
parent
a1fec0b54d
commit
58e9c973ab
4 changed files with 598 additions and 90 deletions
|
@ -1,3 +1,8 @@
|
||||||
|
Changes in version 1.6
|
||||||
|
- support for books whose paragraphs have no styles
|
||||||
|
- support to run cmbtc_dump on Linux and Mac OSX provided you know your PID of your ipod or standalone Kindle
|
||||||
|
(contributed by DiapDealer)
|
||||||
|
|
||||||
Changes in version 1.5
|
Changes in version 1.5
|
||||||
- completely reworked generation of styles to use actual page heights and widths
|
- completely reworked generation of styles to use actual page heights and widths
|
||||||
- added new script getpagedim.py to support the above
|
- added new script getpagedim.py to support the above
|
||||||
|
|
504
Topaz_Tools/lib/cmbtc_dump_mac_linux.py
Normal file
504
Topaz_Tools/lib/cmbtc_dump_mac_linux.py
Normal file
|
@ -0,0 +1,504 @@
|
||||||
|
#! /usr/bin/python
|
||||||
|
|
||||||
|
from __future__ import with_statement
|
||||||
|
|
||||||
|
import csv
|
||||||
|
import sys
|
||||||
|
import os
|
||||||
|
import getopt
|
||||||
|
import zlib
|
||||||
|
from struct import pack
|
||||||
|
from struct import unpack
|
||||||
|
|
||||||
|
MAX_PATH = 255
|
||||||
|
|
||||||
|
# Put the first 8 characters of your Kindle PID here
|
||||||
|
# or supply it with the -p option in the command line
|
||||||
|
####################################################
|
||||||
|
kindlePID = "12345678"
|
||||||
|
####################################################
|
||||||
|
|
||||||
|
global bookFile
|
||||||
|
global bookPayloadOffset
|
||||||
|
global bookHeaderRecords
|
||||||
|
global bookMetadata
|
||||||
|
global bookKey
|
||||||
|
global command
|
||||||
|
|
||||||
|
#
|
||||||
|
# Exceptions for all the problems that might happen during the script
|
||||||
|
#
|
||||||
|
|
||||||
|
class CMBDTCError(Exception):
|
||||||
|
pass
|
||||||
|
|
||||||
|
class CMBDTCFatal(Exception):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# Open the book file at path
|
||||||
|
#
|
||||||
|
|
||||||
|
def openBook(path):
|
||||||
|
try:
|
||||||
|
return open(path,'rb')
|
||||||
|
except:
|
||||||
|
raise CMBDTCFatal("Could not open book file: " + path)
|
||||||
|
|
||||||
|
#
|
||||||
|
# Get a 7 bit encoded number from the book file
|
||||||
|
#
|
||||||
|
|
||||||
|
def bookReadEncodedNumber():
|
||||||
|
flag = False
|
||||||
|
data = ord(bookFile.read(1))
|
||||||
|
|
||||||
|
if data == 0xFF:
|
||||||
|
flag = True
|
||||||
|
data = ord(bookFile.read(1))
|
||||||
|
|
||||||
|
if data >= 0x80:
|
||||||
|
datax = (data & 0x7F)
|
||||||
|
while data >= 0x80 :
|
||||||
|
data = ord(bookFile.read(1))
|
||||||
|
datax = (datax <<7) + (data & 0x7F)
|
||||||
|
data = datax
|
||||||
|
|
||||||
|
if flag:
|
||||||
|
data = -data
|
||||||
|
return data
|
||||||
|
|
||||||
|
#
|
||||||
|
# Encode a number in 7 bit format
|
||||||
|
#
|
||||||
|
|
||||||
|
def encodeNumber(number):
|
||||||
|
result = ""
|
||||||
|
negative = False
|
||||||
|
flag = 0
|
||||||
|
print("Using encodeNumber routine")
|
||||||
|
|
||||||
|
if number < 0 :
|
||||||
|
number = -number + 1
|
||||||
|
negative = True
|
||||||
|
|
||||||
|
while True:
|
||||||
|
byte = number & 0x7F
|
||||||
|
number = number >> 7
|
||||||
|
byte += flag
|
||||||
|
result += chr(byte)
|
||||||
|
flag = 0x80
|
||||||
|
if number == 0 : break
|
||||||
|
|
||||||
|
if negative:
|
||||||
|
result += chr(0xFF)
|
||||||
|
|
||||||
|
return result[::-1]
|
||||||
|
|
||||||
|
#
|
||||||
|
# Get a length prefixed string from the file
|
||||||
|
#
|
||||||
|
|
||||||
|
def bookReadString():
|
||||||
|
stringLength = bookReadEncodedNumber()
|
||||||
|
return unpack(str(stringLength)+"s",bookFile.read(stringLength))[0]
|
||||||
|
|
||||||
|
#
|
||||||
|
# Returns a length prefixed string
|
||||||
|
#
|
||||||
|
|
||||||
|
def lengthPrefixString(data):
|
||||||
|
return encodeNumber(len(data))+data
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# Read and return the data of one header record at the current book file position [[offset,decompressedLength,compressedLength],...]
|
||||||
|
#
|
||||||
|
|
||||||
|
def bookReadHeaderRecordData():
|
||||||
|
nbValues = bookReadEncodedNumber()
|
||||||
|
values = []
|
||||||
|
for i in range (0,nbValues):
|
||||||
|
values.append([bookReadEncodedNumber(),bookReadEncodedNumber(),bookReadEncodedNumber()])
|
||||||
|
return values
|
||||||
|
|
||||||
|
#
|
||||||
|
# Read and parse one header record at the current book file position and return the associated data [[offset,decompressedLength,compressedLength],...]
|
||||||
|
#
|
||||||
|
|
||||||
|
def parseTopazHeaderRecord():
|
||||||
|
if ord(bookFile.read(1)) != 0x63:
|
||||||
|
raise CMBDTCFatal("Parse Error : Invalid Header")
|
||||||
|
|
||||||
|
tag = bookReadString()
|
||||||
|
record = bookReadHeaderRecordData()
|
||||||
|
return [tag,record]
|
||||||
|
|
||||||
|
#
|
||||||
|
# Parse the header of a Topaz file, get all the header records and the offset for the payload
|
||||||
|
#
|
||||||
|
|
||||||
|
def parseTopazHeader():
|
||||||
|
global bookHeaderRecords
|
||||||
|
global bookPayloadOffset
|
||||||
|
magic = unpack("4s",bookFile.read(4))[0]
|
||||||
|
|
||||||
|
if magic != 'TPZ0':
|
||||||
|
raise CMBDTCFatal("Parse Error : Invalid Header, not a Topaz file")
|
||||||
|
|
||||||
|
nbRecords = bookReadEncodedNumber()
|
||||||
|
bookHeaderRecords = {}
|
||||||
|
|
||||||
|
for i in range (0,nbRecords):
|
||||||
|
result = parseTopazHeaderRecord()
|
||||||
|
print result[0], result[1]
|
||||||
|
bookHeaderRecords[result[0]] = result[1]
|
||||||
|
|
||||||
|
if ord(bookFile.read(1)) != 0x64 :
|
||||||
|
raise CMBDTCFatal("Parse Error : Invalid Header")
|
||||||
|
|
||||||
|
bookPayloadOffset = bookFile.tell()
|
||||||
|
|
||||||
|
#
|
||||||
|
# Get a record in the book payload, given its name and index. If necessary the record is decrypted. The record is not decompressed
|
||||||
|
# Correction, the record is correctly decompressed too
|
||||||
|
#
|
||||||
|
|
||||||
|
def getBookPayloadRecord(name, index):
|
||||||
|
encrypted = False
|
||||||
|
compressed = False
|
||||||
|
|
||||||
|
try:
|
||||||
|
recordOffset = bookHeaderRecords[name][index][0]
|
||||||
|
except:
|
||||||
|
raise CMBDTCFatal("Parse Error : Invalid Record, record not found")
|
||||||
|
|
||||||
|
bookFile.seek(bookPayloadOffset + recordOffset)
|
||||||
|
|
||||||
|
tag = bookReadString()
|
||||||
|
if tag != name :
|
||||||
|
raise CMBDTCFatal("Parse Error : Invalid Record, record name doesn't match")
|
||||||
|
|
||||||
|
recordIndex = bookReadEncodedNumber()
|
||||||
|
|
||||||
|
if recordIndex < 0 :
|
||||||
|
encrypted = True
|
||||||
|
recordIndex = -recordIndex -1
|
||||||
|
|
||||||
|
if recordIndex != index :
|
||||||
|
raise CMBDTCFatal("Parse Error : Invalid Record, index doesn't match")
|
||||||
|
|
||||||
|
if (bookHeaderRecords[name][index][2] > 0):
|
||||||
|
compressed = True
|
||||||
|
record = bookFile.read(bookHeaderRecords[name][index][2])
|
||||||
|
else:
|
||||||
|
record = bookFile.read(bookHeaderRecords[name][index][1])
|
||||||
|
|
||||||
|
if encrypted:
|
||||||
|
ctx = topazCryptoInit(bookKey)
|
||||||
|
record = topazCryptoDecrypt(record,ctx)
|
||||||
|
|
||||||
|
if compressed:
|
||||||
|
record = zlib.decompress(record)
|
||||||
|
|
||||||
|
return record
|
||||||
|
|
||||||
|
#
|
||||||
|
# Extract, decrypt and decompress a book record indicated by name and index and print it or save it in "filename"
|
||||||
|
#
|
||||||
|
|
||||||
|
def extractBookPayloadRecord(name, index, filename):
|
||||||
|
compressed = False
|
||||||
|
|
||||||
|
try:
|
||||||
|
compressed = bookHeaderRecords[name][index][2] != 0
|
||||||
|
record = getBookPayloadRecord(name,index)
|
||||||
|
except:
|
||||||
|
print("Could not find record")
|
||||||
|
|
||||||
|
# if compressed:
|
||||||
|
# try:
|
||||||
|
# record = zlib.decompress(record)
|
||||||
|
# except:
|
||||||
|
# raise CMBDTCFatal("Could not decompress record")
|
||||||
|
|
||||||
|
if filename != "":
|
||||||
|
try:
|
||||||
|
file = open(filename,"wb")
|
||||||
|
file.write(record)
|
||||||
|
file.close()
|
||||||
|
except:
|
||||||
|
raise CMBDTCFatal("Could not write to destination file")
|
||||||
|
else:
|
||||||
|
print(record)
|
||||||
|
|
||||||
|
#
|
||||||
|
# return next record [key,value] from the book metadata from the current book position
|
||||||
|
#
|
||||||
|
|
||||||
|
def readMetadataRecord():
|
||||||
|
return [bookReadString(),bookReadString()]
|
||||||
|
|
||||||
|
#
|
||||||
|
# Parse the metadata record from the book payload and return a list of [key,values]
|
||||||
|
#
|
||||||
|
|
||||||
|
def parseMetadata():
|
||||||
|
global bookHeaderRecords
|
||||||
|
global bookPayloadAddress
|
||||||
|
global bookMetadata
|
||||||
|
bookMetadata = {}
|
||||||
|
bookFile.seek(bookPayloadOffset + bookHeaderRecords["metadata"][0][0])
|
||||||
|
tag = bookReadString()
|
||||||
|
if tag != "metadata" :
|
||||||
|
raise CMBDTCFatal("Parse Error : Record Names Don't Match")
|
||||||
|
|
||||||
|
flags = ord(bookFile.read(1))
|
||||||
|
nbRecords = ord(bookFile.read(1))
|
||||||
|
|
||||||
|
for i in range (0,nbRecords) :
|
||||||
|
record =readMetadataRecord()
|
||||||
|
bookMetadata[record[0]] = record[1]
|
||||||
|
|
||||||
|
#
|
||||||
|
# Context initialisation for the Topaz Crypto
|
||||||
|
#
|
||||||
|
|
||||||
|
def topazCryptoInit(key):
|
||||||
|
ctx1 = 0x0CAFFE19E
|
||||||
|
|
||||||
|
for keyChar in key:
|
||||||
|
keyByte = ord(keyChar)
|
||||||
|
ctx2 = ctx1
|
||||||
|
ctx1 = ((((ctx1 >>2) * (ctx1 >>7))&0xFFFFFFFF) ^ (keyByte * keyByte * 0x0F902007)& 0xFFFFFFFF )
|
||||||
|
return [ctx1,ctx2]
|
||||||
|
|
||||||
|
#
|
||||||
|
# decrypt data with the context prepared by topazCryptoInit()
|
||||||
|
#
|
||||||
|
|
||||||
|
def topazCryptoDecrypt(data, ctx):
|
||||||
|
ctx1 = ctx[0]
|
||||||
|
ctx2 = ctx[1]
|
||||||
|
|
||||||
|
plainText = ""
|
||||||
|
|
||||||
|
for dataChar in data:
|
||||||
|
dataByte = ord(dataChar)
|
||||||
|
m = (dataByte ^ ((ctx1 >> 3) &0xFF) ^ ((ctx2<<3) & 0xFF)) &0xFF
|
||||||
|
ctx2 = ctx1
|
||||||
|
ctx1 = (((ctx1 >> 2) * (ctx1 >> 7)) &0xFFFFFFFF) ^((m * m * 0x0F902007) &0xFFFFFFFF)
|
||||||
|
plainText += chr(m)
|
||||||
|
|
||||||
|
return plainText
|
||||||
|
|
||||||
|
#
|
||||||
|
# Decrypt a payload record with the PID
|
||||||
|
#
|
||||||
|
|
||||||
|
def decryptRecord(data,PID):
|
||||||
|
ctx = topazCryptoInit(PID)
|
||||||
|
return topazCryptoDecrypt(data, ctx)
|
||||||
|
|
||||||
|
#
|
||||||
|
# Try to decrypt a dkey record (contains the book PID)
|
||||||
|
#
|
||||||
|
|
||||||
|
def decryptDkeyRecord(data,PID):
|
||||||
|
record = decryptRecord(data,PID)
|
||||||
|
fields = unpack("3sB8sB8s3s",record)
|
||||||
|
|
||||||
|
if fields[0] != "PID" or fields[5] != "pid" :
|
||||||
|
raise CMBDTCError("Didn't find PID magic numbers in record")
|
||||||
|
elif fields[1] != 8 or fields[3] != 8 :
|
||||||
|
raise CMBDTCError("Record didn't contain correct length fields")
|
||||||
|
elif fields[2] != PID :
|
||||||
|
raise CMBDTCError("Record didn't contain PID")
|
||||||
|
|
||||||
|
return fields[4]
|
||||||
|
|
||||||
|
#
|
||||||
|
# Decrypt all the book's dkey records (contain the book PID)
|
||||||
|
#
|
||||||
|
|
||||||
|
def decryptDkeyRecords(data,PID):
|
||||||
|
nbKeyRecords = ord(data[0])
|
||||||
|
records = []
|
||||||
|
data = data[1:]
|
||||||
|
for i in range (0,nbKeyRecords):
|
||||||
|
length = ord(data[0])
|
||||||
|
try:
|
||||||
|
key = decryptDkeyRecord(data[1:length+1],PID)
|
||||||
|
records.append(key)
|
||||||
|
except CMBDTCError:
|
||||||
|
pass
|
||||||
|
data = data[1+length:]
|
||||||
|
|
||||||
|
return records
|
||||||
|
|
||||||
|
#
|
||||||
|
# Create decrypted book payload
|
||||||
|
#
|
||||||
|
|
||||||
|
def createDecryptedPayload(payload):
|
||||||
|
for headerRecord in bookHeaderRecords:
|
||||||
|
name = headerRecord
|
||||||
|
if name != "dkey" :
|
||||||
|
ext = '.dat'
|
||||||
|
if name == 'img' : ext = '.jpg'
|
||||||
|
for index in range (0,len(bookHeaderRecords[name])) :
|
||||||
|
fnum = "%04d" % index
|
||||||
|
fname = name + fnum + ext
|
||||||
|
destdir = payload
|
||||||
|
if name == 'img':
|
||||||
|
destdir = os.path.join(payload,'img')
|
||||||
|
if name == 'page':
|
||||||
|
destdir = os.path.join(payload,'page')
|
||||||
|
if name == 'glyphs':
|
||||||
|
destdir = os.path.join(payload,'glyphs')
|
||||||
|
outputFile = os.path.join(destdir,fname)
|
||||||
|
file(outputFile, 'wb').write(getBookPayloadRecord(name, index))
|
||||||
|
|
||||||
|
|
||||||
|
# Create decrypted book
|
||||||
|
#
|
||||||
|
|
||||||
|
def createDecryptedBook(outdir):
|
||||||
|
if not os.path.exists(outdir):
|
||||||
|
os.makedirs(outdir)
|
||||||
|
|
||||||
|
destdir = os.path.join(outdir,'img')
|
||||||
|
if not os.path.exists(destdir):
|
||||||
|
os.makedirs(destdir)
|
||||||
|
|
||||||
|
destdir = os.path.join(outdir,'page')
|
||||||
|
if not os.path.exists(destdir):
|
||||||
|
os.makedirs(destdir)
|
||||||
|
|
||||||
|
destdir = os.path.join(outdir,'glyphs')
|
||||||
|
if not os.path.exists(destdir):
|
||||||
|
os.makedirs(destdir)
|
||||||
|
|
||||||
|
createDecryptedPayload(outdir)
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# Set the command to execute by the programm according to cmdLine parameters
|
||||||
|
#
|
||||||
|
|
||||||
|
def setCommand(name) :
|
||||||
|
global command
|
||||||
|
if command != "" :
|
||||||
|
raise CMBDTCFatal("Invalid command line parameters")
|
||||||
|
else :
|
||||||
|
command = name
|
||||||
|
|
||||||
|
#
|
||||||
|
# Program usage
|
||||||
|
#
|
||||||
|
|
||||||
|
def usage():
|
||||||
|
print("\nUsage:")
|
||||||
|
print("\ncmbtc_dump_linux.py [options] bookFileName\n")
|
||||||
|
print("-p Adds a PID to the list of PIDs that are tried to decrypt the book key (can be used several times)")
|
||||||
|
print("-d Dumps the unencrypted book as files to outdir")
|
||||||
|
print("-o Output directory to save book files to")
|
||||||
|
print("-v Verbose (can be used several times)")
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# Main
|
||||||
|
#
|
||||||
|
|
||||||
|
def main(argv=sys.argv):
|
||||||
|
global bookMetadata
|
||||||
|
global bookKey
|
||||||
|
global bookFile
|
||||||
|
global command
|
||||||
|
|
||||||
|
progname = os.path.basename(argv[0])
|
||||||
|
|
||||||
|
verbose = 0
|
||||||
|
recordName = ""
|
||||||
|
recordIndex = 0
|
||||||
|
outdir = ""
|
||||||
|
PIDs = []
|
||||||
|
command = ""
|
||||||
|
|
||||||
|
# Preloads your Kindle pid from the top of the program.
|
||||||
|
PIDs.append(kindlePID)
|
||||||
|
|
||||||
|
try:
|
||||||
|
opts, args = getopt.getopt(sys.argv[1:], "vo:p:d")
|
||||||
|
except getopt.GetoptError, err:
|
||||||
|
# print help information and exit:
|
||||||
|
print str(err) # will print something like "option -a not recognized"
|
||||||
|
usage()
|
||||||
|
sys.exit(2)
|
||||||
|
|
||||||
|
if len(opts) == 0 and len(args) == 0 :
|
||||||
|
usage()
|
||||||
|
sys.exit(2)
|
||||||
|
|
||||||
|
for o, a in opts:
|
||||||
|
if o == "-v":
|
||||||
|
verbose+=1
|
||||||
|
if o =="-o":
|
||||||
|
if a == None :
|
||||||
|
raise CMBDTCFatal("Invalid parameter for -o")
|
||||||
|
outdir = a
|
||||||
|
if o =="-p":
|
||||||
|
PIDs.append(a)
|
||||||
|
if o =="-d":
|
||||||
|
setCommand("doit")
|
||||||
|
|
||||||
|
if command == "" :
|
||||||
|
raise CMBDTCFatal("No action supplied on command line")
|
||||||
|
|
||||||
|
#
|
||||||
|
# Open book and parse metadata
|
||||||
|
#
|
||||||
|
|
||||||
|
if len(args) == 1:
|
||||||
|
|
||||||
|
bookFile = openBook(args[0])
|
||||||
|
parseTopazHeader()
|
||||||
|
parseMetadata()
|
||||||
|
|
||||||
|
#
|
||||||
|
# Decrypt book key
|
||||||
|
#
|
||||||
|
|
||||||
|
dkey = getBookPayloadRecord('dkey', 0)
|
||||||
|
|
||||||
|
bookKeys = []
|
||||||
|
for PID in PIDs :
|
||||||
|
bookKeys+=decryptDkeyRecords(dkey,PID)
|
||||||
|
|
||||||
|
if len(bookKeys) == 0 :
|
||||||
|
if verbose > 0 :
|
||||||
|
print ("Book key could not be found. Maybe this book is not registered with this device.")
|
||||||
|
else :
|
||||||
|
bookKey = bookKeys[0]
|
||||||
|
if verbose > 0:
|
||||||
|
print("Book key: " + bookKey.encode('hex'))
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
if command == "printRecord" :
|
||||||
|
extractBookPayloadRecord(recordName,int(recordIndex),outputFile)
|
||||||
|
if outputFile != "" and verbose>0 :
|
||||||
|
print("Wrote record to file: "+outputFile)
|
||||||
|
elif command == "doit" :
|
||||||
|
if outdir != "" :
|
||||||
|
createDecryptedBook(outdir)
|
||||||
|
if verbose >0 :
|
||||||
|
print ("Decrypted book saved. Don't pirate!")
|
||||||
|
elif verbose > 0:
|
||||||
|
print("Output directory name was not supplied.")
|
||||||
|
|
||||||
|
return 0
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
sys.exit(main())
|
|
@ -13,7 +13,8 @@ from struct import unpack
|
||||||
class DocParser(object):
|
class DocParser(object):
|
||||||
def __init__(self, flatxml, classlst, fileid):
|
def __init__(self, flatxml, classlst, fileid):
|
||||||
self.id = os.path.basename(fileid).replace('.dat','')
|
self.id = os.path.basename(fileid).replace('.dat','')
|
||||||
self.flatdoc = flatxml.split('\n')
|
self.docList = flatxml.split('\n')
|
||||||
|
self.docSize = len(self.docList)
|
||||||
self.classList = {}
|
self.classList = {}
|
||||||
tmpList = classlst.split('\n')
|
tmpList = classlst.split('\n')
|
||||||
for pclass in tmpList:
|
for pclass in tmpList:
|
||||||
|
@ -29,12 +30,10 @@ class DocParser(object):
|
||||||
self.paracont_stemid = []
|
self.paracont_stemid = []
|
||||||
self.parastems_stemid = []
|
self.parastems_stemid = []
|
||||||
|
|
||||||
# find tag if within pos to end inclusive
|
# return tag at line pos in document
|
||||||
def lineinDoc(self, pos) :
|
def lineinDoc(self, pos) :
|
||||||
docList = self.flatdoc
|
if (pos >= 0) and (pos < self.docSize) :
|
||||||
cnt = len(docList)
|
item = self.docList[pos]
|
||||||
if (pos >= 0) and (pos < cnt) :
|
|
||||||
item = docList[pos]
|
|
||||||
if item.find('=') >= 0:
|
if item.find('=') >= 0:
|
||||||
(name, argres) = item.split('=',1)
|
(name, argres) = item.split('=',1)
|
||||||
else :
|
else :
|
||||||
|
@ -43,20 +42,18 @@ class DocParser(object):
|
||||||
return name, argres
|
return name, argres
|
||||||
|
|
||||||
|
|
||||||
# find tag if within pos to end inclusive
|
# find tag in doc if within pos to end inclusive
|
||||||
def findinDoc(self, tagpath, pos, end) :
|
def findinDoc(self, tagpath, pos, end) :
|
||||||
result = None
|
result = None
|
||||||
docList = self.flatdoc
|
|
||||||
cnt = len(docList)
|
|
||||||
if end == -1 :
|
if end == -1 :
|
||||||
end = cnt
|
end = self.docSize
|
||||||
else:
|
else:
|
||||||
end = min(cnt,end)
|
end = min(self.docSize, end)
|
||||||
foundat = -1
|
foundat = -1
|
||||||
for j in xrange(pos, end):
|
for j in xrange(pos, end):
|
||||||
item = docList[j]
|
item = self.docList[j]
|
||||||
if item.find('=') >= 0:
|
if item.find('=') >= 0:
|
||||||
(name, argres) = item.split('=')
|
(name, argres) = item.split('=',1)
|
||||||
else :
|
else :
|
||||||
name = item
|
name = item
|
||||||
argres = ''
|
argres = ''
|
||||||
|
@ -85,7 +82,7 @@ class DocParser(object):
|
||||||
|
|
||||||
result = []
|
result = []
|
||||||
|
|
||||||
# normal paragraph
|
# paragraph
|
||||||
(pos, pclass) = self.findinDoc('paragraph.class',start,end)
|
(pos, pclass) = self.findinDoc('paragraph.class',start,end)
|
||||||
|
|
||||||
# class names are an issue given topaz may start them with numerals (not allowed),
|
# class names are an issue given topaz may start them with numerals (not allowed),
|
||||||
|
@ -94,19 +91,20 @@ class DocParser(object):
|
||||||
# from a base class (but then not actually provide all of these _reclustereed
|
# from a base class (but then not actually provide all of these _reclustereed
|
||||||
# classes in the stylesheet!
|
# classes in the stylesheet!
|
||||||
|
|
||||||
# so we clean this up by lowercasing, prepend 'cl_', and getting any baseclass
|
# so we clean this up by lowercasing, prepend 'cl-', and getting any baseclass
|
||||||
# that exists in the stylesheet first, and then adding this specific class
|
# that exists in the stylesheet first, and then adding this specific class
|
||||||
# after
|
# after
|
||||||
classres = ''
|
if pclass != None :
|
||||||
pclass = pclass.lower()
|
classres = ''
|
||||||
pclass = 'cl-' + pclass
|
pclass = pclass.lower()
|
||||||
p = pclass.find('_')
|
pclass = 'cl-' + pclass
|
||||||
if p > 0 :
|
p = pclass.find('_')
|
||||||
baseclass = pclass[0:p]
|
if p > 0 :
|
||||||
if baseclass in self.classList:
|
baseclass = pclass[0:p]
|
||||||
classres += baseclass + ' '
|
if baseclass in self.classList:
|
||||||
classres += pclass
|
classres += baseclass + ' '
|
||||||
pclass = classres
|
classres += pclass
|
||||||
|
pclass = classres
|
||||||
|
|
||||||
# build up a description of the paragraph in result and return it
|
# build up a description of the paragraph in result and return it
|
||||||
# first check for the basic - all words paragraph
|
# first check for the basic - all words paragraph
|
||||||
|
@ -128,9 +126,7 @@ class DocParser(object):
|
||||||
|
|
||||||
# if end is -1 then we must search to end of document
|
# if end is -1 then we must search to end of document
|
||||||
if end == -1 :
|
if end == -1 :
|
||||||
docList = self.flatdoc
|
end = self.docSize
|
||||||
cnt = len(docList)
|
|
||||||
end = cnt
|
|
||||||
|
|
||||||
while (line < end) :
|
while (line < end) :
|
||||||
|
|
||||||
|
@ -171,20 +167,20 @@ class DocParser(object):
|
||||||
return pclass, result
|
return pclass, result
|
||||||
|
|
||||||
|
|
||||||
def buildParagraph(self, cname, pdesc, type, regtype) :
|
def buildParagraph(self, pclass, pdesc, type, regtype) :
|
||||||
parares = ''
|
parares = ''
|
||||||
sep =''
|
sep =''
|
||||||
|
|
||||||
br_lb = False
|
classres = ''
|
||||||
if (regtype == 'fixed') or (regtype == 'chapterheading'):
|
if pclass :
|
||||||
br_lb = True
|
classres = ' class="' + pclass + '"'
|
||||||
|
|
||||||
handle_links = False
|
br_lb = (regtype == 'fixed') or (regtype == 'chapterheading')
|
||||||
if len(self.link_id) > 0:
|
|
||||||
handle_links = True
|
handle_links = len(self.link_id) > 0
|
||||||
|
|
||||||
if (type == 'full') or (type == 'begin') :
|
if (type == 'full') or (type == 'begin') :
|
||||||
parares += '<p class="' + cname + '">'
|
parares += '<p' + classres + '>'
|
||||||
|
|
||||||
if (type == 'end'):
|
if (type == 'end'):
|
||||||
parares += ' '
|
parares += ' '
|
||||||
|
@ -218,10 +214,7 @@ class DocParser(object):
|
||||||
if word == '_link_' : word = ''
|
if word == '_link_' : word = ''
|
||||||
|
|
||||||
if word == '_lb_':
|
if word == '_lb_':
|
||||||
if (num-1) in self.dehyphen_rootid :
|
if ((num-1) in self.dehyphen_rootid ) or handle_links:
|
||||||
word = ''
|
|
||||||
sep = ''
|
|
||||||
elif handle_links :
|
|
||||||
word = ''
|
word = ''
|
||||||
sep = ''
|
sep = ''
|
||||||
elif br_lb :
|
elif br_lb :
|
||||||
|
@ -261,43 +254,51 @@ class DocParser(object):
|
||||||
|
|
||||||
htmlpage = ''
|
htmlpage = ''
|
||||||
|
|
||||||
# first collect information from the xml doc that describes this page
|
# get the ocr text
|
||||||
(pos, argres) = self.findinDoc('info.word.ocrText',0,-1)
|
(pos, argres) = self.findinDoc('info.word.ocrText',0,-1)
|
||||||
if argres : self.ocrtext = argres.split('|')
|
if argres : self.ocrtext = argres.split('|')
|
||||||
|
|
||||||
|
# get information to dehyphenate the text
|
||||||
(pos, argres) = self.findinDoc('info.dehyphen.rootID',0,-1)
|
(pos, argres) = self.findinDoc('info.dehyphen.rootID',0,-1)
|
||||||
if argres:
|
if argres:
|
||||||
argList = argres.split('|')
|
argList = argres.split('|')
|
||||||
self.dehyphen_rootid = [ int(strval) for strval in argList]
|
self.dehyphen_rootid = [ int(strval) for strval in argList]
|
||||||
|
|
||||||
|
# determine if first paragraph is continued from previous page
|
||||||
(pos, self.parastems_stemid) = self.findinDoc('info.paraStems.stemID',0,-1)
|
(pos, self.parastems_stemid) = self.findinDoc('info.paraStems.stemID',0,-1)
|
||||||
if self.parastems_stemid == None : self.parastems_stemid = []
|
first_para_continued = (self.parastems_stemid != None)
|
||||||
|
|
||||||
|
# determine if last paragraph is continued onto the next page
|
||||||
(pos, self.paracont_stemid) = self.findinDoc('info.paraCont.stemID',0,-1)
|
(pos, self.paracont_stemid) = self.findinDoc('info.paraCont.stemID',0,-1)
|
||||||
if self.paracont_stemid == None : self.paracont_stemid = []
|
last_para_continued = (self.paracont_stemid != None)
|
||||||
|
|
||||||
|
|
||||||
|
# collect link ids
|
||||||
(pos, argres) = self.findinDoc('info.word.link_id',0,-1)
|
(pos, argres) = self.findinDoc('info.word.link_id',0,-1)
|
||||||
if argres:
|
if argres:
|
||||||
argList = argres.split('|')
|
argList = argres.split('|')
|
||||||
self.link_id = [ int(strval) for strval in argList]
|
self.link_id = [ int(strval) for strval in argList]
|
||||||
|
|
||||||
|
# collect link destination page numbers
|
||||||
(pos, argres) = self.findinDoc('info.links.page',0,-1)
|
(pos, argres) = self.findinDoc('info.links.page',0,-1)
|
||||||
if argres :
|
if argres :
|
||||||
argList = argres.split('|')
|
argList = argres.split('|')
|
||||||
self.link_page = [ int(strval) for strval in argList]
|
self.link_page = [ int(strval) for strval in argList]
|
||||||
|
|
||||||
|
# collect link titles
|
||||||
(pos, argres) = self.findinDoc('info.links.title',0,-1)
|
(pos, argres) = self.findinDoc('info.links.title',0,-1)
|
||||||
if argres :
|
if argres :
|
||||||
self.link_title = argres.split('|')
|
self.link_title = argres.split('|')
|
||||||
else:
|
else:
|
||||||
self.link_title.append('')
|
self.link_title.append('')
|
||||||
|
|
||||||
|
|
||||||
|
# get page type
|
||||||
(pos, pagetype) = self.findinDoc('page.type',0,-1)
|
(pos, pagetype) = self.findinDoc('page.type',0,-1)
|
||||||
|
|
||||||
|
|
||||||
# generate a list of each region starting point
|
# generate a list of each region starting point
|
||||||
# each region has one paragraph,, or one image, or one chapterheading
|
# each region has one paragraph,, or one image, or one chapterheading
|
||||||
|
|
||||||
regionList= self.posinDoc('region')
|
regionList= self.posinDoc('region')
|
||||||
regcnt = len(regionList)
|
regcnt = len(regionList)
|
||||||
regionList.append(-1)
|
regionList.append(-1)
|
||||||
|
@ -308,47 +309,48 @@ class DocParser(object):
|
||||||
# process each region tag and convert what you can to html
|
# process each region tag and convert what you can to html
|
||||||
|
|
||||||
for j in xrange(regcnt):
|
for j in xrange(regcnt):
|
||||||
|
|
||||||
start = regionList[j]
|
start = regionList[j]
|
||||||
end = regionList[j+1]
|
end = regionList[j+1]
|
||||||
|
|
||||||
(pos, regtype) = self.findinDoc('region.type',start,end)
|
(pos, regtype) = self.findinDoc('region.type',start,end)
|
||||||
|
|
||||||
|
# set anchor for link target on this page
|
||||||
|
if not anchorSet and not first_para_continued:
|
||||||
|
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
||||||
|
anchorSet = True
|
||||||
|
|
||||||
if regtype == 'graphic' :
|
if regtype == 'graphic' :
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
||||||
if simgsrc:
|
if simgsrc:
|
||||||
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
||||||
|
|
||||||
|
|
||||||
elif regtype == 'chapterheading' :
|
elif regtype == 'chapterheading' :
|
||||||
(pclass, pdesc) = self.getParaDescription(start,end)
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
||||||
if not breakSet:
|
if not breakSet:
|
||||||
htmlpage += '<div style="page-break-after: always;"> </div>\n'
|
htmlpage += '<div style="page-break-after: always;"> </div>\n'
|
||||||
breakSet = True
|
breakSet = True
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
tag = 'h1'
|
tag = 'h1'
|
||||||
if pclass[3:7] == 'ch1-' : tag = 'h1'
|
if pclass and (len(pclass) >= 7):
|
||||||
if pclass[3:7] == 'ch2-' : tag = 'h2'
|
if pclass[3:7] == 'ch1-' : tag = 'h1'
|
||||||
if pclass[3:7] == 'ch3-' : tag = 'h3'
|
if pclass[3:7] == 'ch2-' : tag = 'h2'
|
||||||
htmlpage += '<' + tag + ' class="' + pclass + '">'
|
if pclass[3:7] == 'ch3-' : tag = 'h3'
|
||||||
|
htmlpage += '<' + tag + ' class="' + pclass + '">'
|
||||||
|
else:
|
||||||
|
htmlpage += '<' + tag + '>'
|
||||||
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
|
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
|
||||||
htmlpage += '</' + tag + '>'
|
htmlpage += '</' + tag + '>'
|
||||||
|
|
||||||
|
|
||||||
elif (regtype == 'text') or (regtype == 'fixed') or (regtype == 'insert') or (regtype == 'listitem') :
|
elif (regtype == 'text') or (regtype == 'fixed') or (regtype == 'insert') or (regtype == 'listitem') :
|
||||||
ptype = 'full'
|
ptype = 'full'
|
||||||
# check to see if this is a continution from the previous page
|
# check to see if this is a continution from the previous page
|
||||||
if (len(self.parastems_stemid) > 0):
|
if first_para_continued :
|
||||||
ptype = 'end'
|
ptype = 'end'
|
||||||
self.parastems_stemid=[]
|
first_para_continued = False
|
||||||
else:
|
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
(pclass, pdesc) = self.getParaDescription(start,end)
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
||||||
if ptype == 'full' :
|
if pclass and (len(pclass) >= 6) and (ptype == 'full'):
|
||||||
tag = 'p'
|
tag = 'p'
|
||||||
if pclass[3:6] == 'h1-' : tag = 'h4'
|
if pclass[3:6] == 'h1-' : tag = 'h4'
|
||||||
if pclass[3:6] == 'h2-' : tag = 'h5'
|
if pclass[3:6] == 'h2-' : tag = 'h5'
|
||||||
|
@ -359,28 +361,22 @@ class DocParser(object):
|
||||||
else :
|
else :
|
||||||
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
||||||
|
|
||||||
|
|
||||||
elif (regtype == 'tocentry') :
|
elif (regtype == 'tocentry') :
|
||||||
ptype = 'full'
|
ptype = 'full'
|
||||||
# check to see if this is a continution from the previous page
|
if first_para_continued :
|
||||||
if (len(self.parastems_stemid) > 0) and (j == 0):
|
|
||||||
# process the first paragraph as a continuation from the last page
|
|
||||||
ptype = 'end'
|
ptype = 'end'
|
||||||
self.parastems_stemid = []
|
first_para_continued = False
|
||||||
else:
|
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
(pclass, pdesc) = self.getParaDescription(start,end)
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
||||||
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
||||||
|
|
||||||
|
|
||||||
elif (regtype == 'synth_fcvr.center') or (regtype == 'synth_text.center'):
|
elif (regtype == 'synth_fcvr.center') or (regtype == 'synth_text.center'):
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
||||||
if simgsrc:
|
if simgsrc:
|
||||||
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
||||||
|
|
||||||
|
|
||||||
else :
|
else :
|
||||||
print 'Warning: Unknown region type', regtype
|
print 'Warning: Unknown region type', regtype
|
||||||
(pos, temp) = self.findinDoc('paragraph',start,end)
|
(pos, temp) = self.findinDoc('paragraph',start,end)
|
||||||
|
@ -389,15 +385,11 @@ class DocParser(object):
|
||||||
regtype = 'fixed'
|
regtype = 'fixed'
|
||||||
ptype = 'full'
|
ptype = 'full'
|
||||||
# check to see if this is a continution from the previous page
|
# check to see if this is a continution from the previous page
|
||||||
if (len(self.parastems_stemid) > 0):
|
if first_para_continued :
|
||||||
ptype = 'end'
|
ptype = 'end'
|
||||||
self.parastems_stemid=[]
|
first_para_continued = False
|
||||||
else:
|
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
(pclass, pdesc) = self.getParaDescription(start,end)
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
||||||
if ptype == 'full' :
|
if pclass and (ptype == 'full') and (len(pclass) >= 6):
|
||||||
tag = 'p'
|
tag = 'p'
|
||||||
if pclass[3:6] == 'h1-' : tag = 'h4'
|
if pclass[3:6] == 'h1-' : tag = 'h4'
|
||||||
if pclass[3:6] == 'h2-' : tag = 'h5'
|
if pclass[3:6] == 'h2-' : tag = 'h5'
|
||||||
|
@ -408,24 +400,20 @@ class DocParser(object):
|
||||||
else :
|
else :
|
||||||
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
||||||
else :
|
else :
|
||||||
print 'Treating this like a "image" region'
|
print 'Treating this like a "graphic" region'
|
||||||
if not anchorSet:
|
|
||||||
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
||||||
anchorSet = True
|
|
||||||
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
||||||
if simgsrc:
|
if simgsrc:
|
||||||
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
||||||
|
|
||||||
if len(self.paracont_stemid) > 0 :
|
|
||||||
|
if last_para_continued :
|
||||||
if htmlpage[-4:] == '</p>':
|
if htmlpage[-4:] == '</p>':
|
||||||
htmlpage = htmlpage[0:-4]
|
htmlpage = htmlpage[0:-4]
|
||||||
|
last_para_continued = False
|
||||||
|
|
||||||
return htmlpage
|
return htmlpage
|
||||||
|
|
||||||
|
|
||||||
return self.convert2HTML()
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def convert2HTML(flatxml, classlst, fileid):
|
def convert2HTML(flatxml, classlst, fileid):
|
||||||
|
|
||||||
|
|
|
@ -3,7 +3,7 @@ Contributors:
|
||||||
clarknova - for all of the svg and glyph generation and many other bug fixes and improvements
|
clarknova - for all of the svg and glyph generation and many other bug fixes and improvements
|
||||||
skindle - for figuing out the general case for the mode loops
|
skindle - for figuing out the general case for the mode loops
|
||||||
some updates - for conversion to xml, basic html
|
some updates - for conversion to xml, basic html
|
||||||
DiapDealer - for extensive testing and feedback
|
DiapDealer - for extensive testing and feedback, and standalone linux/macosx version of cmbtc_dump
|
||||||
stewball - for extensive testing and feedback
|
stewball - for extensive testing and feedback
|
||||||
|
|
||||||
and others for posting, feedback and testing
|
and others for posting, feedback and testing
|
||||||
|
@ -29,6 +29,17 @@ genxml.py - main program to convert everything to xml
|
||||||
genhtml.py - main program to generate "book.html"
|
genhtml.py - main program to generate "book.html"
|
||||||
gensvg.py - (author: clarknova) main program to create an svg grpahic of each page
|
gensvg.py - (author: clarknova) main program to create an svg grpahic of each page
|
||||||
|
|
||||||
|
|
||||||
|
In addition there is now a new file:
|
||||||
|
|
||||||
|
cmbtc_dump_mac_linux.py
|
||||||
|
|
||||||
|
If you know the pid of your ipod and/or your standalone Kindle and your book
|
||||||
|
was meant for that device, you can use this program to dump the proper sections
|
||||||
|
on Mac OSX and Linux (and even Windows if you do not have Kindle4PC installed).
|
||||||
|
Thank DiapDealer for creating it!
|
||||||
|
|
||||||
|
|
||||||
Please note, gensvg.py, genhtml.py, and genxml.py import and use
|
Please note, gensvg.py, genhtml.py, and genxml.py import and use
|
||||||
decode_meta.py, convert2xml.py, flatxml2html.py, getpagedim.py and stylexml2css.py
|
decode_meta.py, convert2xml.py, flatxml2html.py, getpagedim.py and stylexml2css.py
|
||||||
so please keep all of these python scripts together in the same place.
|
so please keep all of these python scripts together in the same place.
|
||||||
|
|
Loading…
Reference in a new issue