| #!/usr/bin/env python3 |
| # |
| # Copyright (C) 2012 The Android Open Source Project |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| """ |
| Usage: generate-notice-files --text-output [plain text output file] \ |
| --html-output [html output file] \ |
| --xml-output [xml output file] \ |
| -t [file title] -s [directory of notices] |
| |
| Generate the Android notice files, including both text and html files. |
| |
| -h to display this usage message and exit. |
| """ |
| from collections import defaultdict |
| import argparse |
| import hashlib |
| import itertools |
| import os |
| import os.path |
| import re |
| import struct |
| import sys |
| |
| MD5_BLOCKSIZE = 1024 * 1024 |
| HTML_ESCAPE_TABLE = { |
| b"&": b"&", |
| b'"': b""", |
| b"'": b"'", |
| b">": b">", |
| b"<": b"<", |
| } |
| |
| def md5sum(filename): |
| """Calculate an MD5 of the file given by FILENAME, |
| and return hex digest as a string. |
| Output should be compatible with md5sum command""" |
| |
| f = open(filename, "rb") |
| sum = hashlib.md5() |
| while 1: |
| block = f.read(MD5_BLOCKSIZE) |
| if not block: |
| break |
| sum.update(block) |
| f.close() |
| return sum.hexdigest() |
| |
| |
| def html_escape(text): |
| """Produce entities within text.""" |
| # Using for i in text doesn't work since i will be an int, not a byte. |
| # There are multiple ways to solve this, but the most performant way |
| # to iterate over a byte array is to use unpack. Using the |
| # for i in range(len(text)) and using that to get a byte using array |
| # slices is twice as slow as this method. |
| return b"".join(HTML_ESCAPE_TABLE.get(i,i) for i in struct.unpack(str(len(text)) + 'c', text)) |
| |
| HTML_OUTPUT_CSS=b""" |
| <style type="text/css"> |
| body { padding: 0; font-family: sans-serif; } |
| .same-license { background-color: #eeeeee; border-top: 20px solid white; padding: 10px; } |
| .label { font-weight: bold; } |
| .file-list { margin-left: 1em; color: blue; } |
| </style> |
| |
| """ |
| |
| def combine_notice_files_html(file_hash, input_dir, output_filename): |
| """Combine notice files in FILE_HASH and output a HTML version to OUTPUT_FILENAME.""" |
| |
| SRC_DIR_STRIP_RE = re.compile(input_dir + "(/.*).txt") |
| |
| # Set up a filename to row id table (anchors inside tables don't work in |
| # most browsers, but href's to table row ids do) |
| id_table = {} |
| id_count = 0 |
| for value in file_hash: |
| for filename in value: |
| id_table[filename] = id_count |
| id_count += 1 |
| |
| # Open the output file, and output the header pieces |
| output_file = open(output_filename, "wb") |
| |
| output_file.write(b"<html><head>\n") |
| output_file.write(HTML_OUTPUT_CSS) |
| output_file.write(b'</head><body topmargin="0" leftmargin="0" rightmargin="0" bottommargin="0">\n') |
| |
| # Output our table of contents |
| output_file.write(b'<div class="toc">\n') |
| output_file.write(b"<ul>\n") |
| |
| # Flatten the list of lists into a single list of filenames |
| sorted_filenames = sorted(itertools.chain.from_iterable(file_hash)) |
| |
| # Print out a nice table of contents |
| for filename in sorted_filenames: |
| stripped_filename = SRC_DIR_STRIP_RE.sub(r"\1", filename) |
| output_file.write(('<li><a href="#id%d">%s</a></li>\n' % (id_table.get(filename), stripped_filename)).encode()) |
| |
| output_file.write(b"</ul>\n") |
| output_file.write(b"</div><!-- table of contents -->\n") |
| # Output the individual notice file lists |
| output_file.write(b'<table cellpadding="0" cellspacing="0" border="0">\n') |
| for value in file_hash: |
| output_file.write(('<tr id="id%d"><td class="same-license">\n' % id_table.get(value[0])).encode()) |
| output_file.write(b'<div class="label">Notices for file(s):</div>\n') |
| output_file.write(b'<div class="file-list">\n') |
| for filename in value: |
| output_file.write(("%s <br/>\n" % (SRC_DIR_STRIP_RE.sub(r"\1", filename))).encode()) |
| output_file.write(b"</div><!-- file-list -->\n\n") |
| output_file.write(b'<pre class="license-text">\n') |
| with open(value[0], "rb") as notice_file: |
| output_file.write(html_escape(notice_file.read())) |
| output_file.write(b"\n</pre><!-- license-text -->\n") |
| output_file.write(b"</td></tr><!-- same-license -->\n\n\n\n") |
| |
| # Finish off the file output |
| output_file.write(b"</table>\n") |
| output_file.write(b"</body></html>\n") |
| output_file.close() |
| |
| def combine_notice_files_text(file_hash, input_dir, output_filename, file_title): |
| """Combine notice files in FILE_HASH and output a text version to OUTPUT_FILENAME.""" |
| |
| SRC_DIR_STRIP_RE = re.compile(input_dir + "(/.*).txt") |
| output_file = open(output_filename, "wb") |
| output_file.write(file_title.encode()) |
| output_file.write(b"\n") |
| for value in file_hash: |
| output_file.write(b"============================================================\n") |
| output_file.write(b"Notices for file(s):\n") |
| for filename in value: |
| output_file.write(SRC_DIR_STRIP_RE.sub(r"\1", filename).encode()) |
| output_file.write(b"\n") |
| output_file.write(b"------------------------------------------------------------\n") |
| with open(value[0], "rb") as notice_file: |
| output_file.write(notice_file.read()) |
| output_file.write(b"\n") |
| output_file.close() |
| |
| def combine_notice_files_xml(files_with_same_hash, input_dir, output_filename): |
| """Combine notice files in FILE_HASH and output a XML version to OUTPUT_FILENAME.""" |
| |
| SRC_DIR_STRIP_RE = re.compile(input_dir + "(/.*).txt") |
| |
| # Set up a filename to row id table (anchors inside tables don't work in |
| # most browsers, but href's to table row ids do) |
| id_table = {} |
| for file_key, files in files_with_same_hash.items(): |
| for filename in files: |
| id_table[filename] = file_key |
| |
| # Open the output file, and output the header pieces |
| output_file = open(output_filename, "wb") |
| |
| output_file.write(b'<?xml version="1.0" encoding="utf-8"?>\n') |
| output_file.write(b"<licenses>\n") |
| |
| # Flatten the list of lists into a single list of filenames |
| sorted_filenames = sorted(list(id_table)) |
| |
| # Print out a nice table of contents |
| for filename in sorted_filenames: |
| stripped_filename = SRC_DIR_STRIP_RE.sub(r"\1", filename) |
| output_file.write(('<file-name contentId="%s">%s</file-name>\n' % (id_table.get(filename), stripped_filename)).encode()) |
| output_file.write(b"\n\n") |
| |
| processed_file_keys = [] |
| # Output the individual notice file lists |
| for filename in sorted_filenames: |
| file_key = id_table.get(filename) |
| if file_key in processed_file_keys: |
| continue |
| processed_file_keys.append(file_key) |
| |
| output_file.write(('<file-content contentId="%s"><![CDATA[' % file_key).encode()) |
| with open(filename, "rb") as notice_file: |
| output_file.write(html_escape(notice_file.read())) |
| output_file.write(b"]]></file-content>\n\n") |
| |
| # Finish off the file output |
| output_file.write(b"</licenses>\n") |
| output_file.close() |
| |
| def get_args(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument( |
| '--text-output', required=True, |
| help='The text output file path.') |
| parser.add_argument( |
| '--html-output', |
| help='The html output file path.') |
| parser.add_argument( |
| '--xml-output', |
| help='The xml output file path.') |
| parser.add_argument( |
| '-t', '--title', required=True, |
| help='The file title.') |
| parser.add_argument( |
| '-s', '--source-dir', required=True, |
| help='The directory containing notices.') |
| parser.add_argument( |
| '-i', '--included-subdirs', action='append', |
| help='The sub directories which should be included.') |
| parser.add_argument( |
| '-e', '--excluded-subdirs', action='append', |
| help='The sub directories which should be excluded.') |
| return parser.parse_args() |
| |
| def main(argv): |
| args = get_args() |
| |
| txt_output_file = args.text_output |
| html_output_file = args.html_output |
| xml_output_file = args.xml_output |
| file_title = args.title |
| included_subdirs = [] |
| excluded_subdirs = [] |
| if args.included_subdirs is not None: |
| included_subdirs = args.included_subdirs |
| if args.excluded_subdirs is not None: |
| excluded_subdirs = args.excluded_subdirs |
| |
| # Find all the notice files and md5 them |
| input_dir = os.path.normpath(args.source_dir) |
| files_with_same_hash = defaultdict(list) |
| for root, dir, files in os.walk(input_dir): |
| for file in files: |
| matched = True |
| if len(included_subdirs) > 0: |
| matched = False |
| for subdir in included_subdirs: |
| if (root == (input_dir + '/' + subdir) or |
| root.startswith(input_dir + '/' + subdir + '/')): |
| matched = True |
| break |
| elif len(excluded_subdirs) > 0: |
| for subdir in excluded_subdirs: |
| if (root == (input_dir + '/' + subdir) or |
| root.startswith(input_dir + '/' + subdir + '/')): |
| matched = False |
| break |
| if matched and file.endswith(".txt"): |
| filename = os.path.join(root, file) |
| file_md5sum = md5sum(filename) |
| files_with_same_hash[file_md5sum].append(filename) |
| |
| filesets = [sorted(files_with_same_hash[md5]) for md5 in sorted(list(files_with_same_hash))] |
| |
| combine_notice_files_text(filesets, input_dir, txt_output_file, file_title) |
| |
| if html_output_file is not None: |
| combine_notice_files_html(filesets, input_dir, html_output_file) |
| |
| if xml_output_file is not None: |
| combine_notice_files_xml(files_with_same_hash, input_dir, xml_output_file) |
| |
| if __name__ == "__main__": |
| main(sys.argv) |