idf_size.py 8.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231
  1. #!/usr/bin/env python
  2. #
  3. # esp-idf alternative to "size" to print ELF file sizes, also analyzes
  4. # the linker map file to dump higher resolution details.
  5. #
  6. # Includes information which is not shown in "xtensa-esp32-elf-size",
  7. # or easy to parse from "xtensa-esp32-elf-objdump" or raw map files.
  8. #
  9. # Copyright 2017 Espressif Systems (Shanghai) PTE LTD
  10. #
  11. # Licensed under the Apache License, Version 2.0 (the "License");
  12. # you may not use this file except in compliance with the License.
  13. # You may obtain a copy of the License at
  14. #
  15. # http://www.apache.org/licenses/LICENSE-2.0
  16. #
  17. # Unless required by applicable law or agreed to in writing, software
  18. # distributed under the License is distributed on an "AS IS" BASIS,
  19. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  20. # See the License for the specific language governing permissions and
  21. # limitations under the License.
  22. #
  23. import argparse, sys, subprocess, re
  24. import os.path
  25. import pprint
  26. DEFAULT_TOOLCHAIN_PREFIX = "xtensa-esp32-elf-"
  27. CHIP_SIZES = {
  28. "esp32" : {
  29. "total_iram" : 0x20000,
  30. "total_irom" : 0x330000,
  31. "total_drom" : 0x800000,
  32. # total dram is determined from objdump output
  33. }
  34. }
  35. def scan_to_header(f, header_line):
  36. """ Scan forward in a file until you reach 'header_line', then return """
  37. for line in f:
  38. if line.strip() == header_line:
  39. return
  40. raise RuntimeError("Didn't find line '%s' in file" % header_line)
  41. def load_map_data(map_file):
  42. memory_config = load_memory_config(map_file)
  43. sections = load_sections(map_file)
  44. return memory_config, sections
  45. def output_section_for_address(memory_config, address):
  46. for m in memory_config.values():
  47. if m["origin"] <= address and m["origin"] + m["length"] > address:
  48. return m["name"]
  49. return None
  50. def load_memory_config(map_file):
  51. """ Memory Configuration section is the total size of each output section """
  52. result = {}
  53. scan_to_header(map_file, "Memory Configuration")
  54. RE_MEMORY_SECTION = r"(?P<name>[^ ]+) +0x(?P<origin>[\da-f]+) +0x(?P<length>[\da-f]+)"
  55. for line in map_file:
  56. m = re.match(RE_MEMORY_SECTION, line)
  57. if m is None:
  58. if len(result) == 0:
  59. continue # whitespace or a header, before the content we want
  60. else:
  61. return result # we're at the end of the Memory Configuration
  62. section = {
  63. "name" : m.group("name"),
  64. "origin" : int(m.group("origin"), 16),
  65. "length" : int(m.group("length"), 16),
  66. }
  67. if section["name"] != "*default*":
  68. result[section["name"]] = section
  69. raise RuntimeError("End of file while scanning memory configuration?")
  70. def load_sections(map_file):
  71. """ Load section size information from the MAP file.
  72. Returns a dict of 'sections', where each key is a section name and the value
  73. is a dict with details about this section, including a "sources" key which holds a list of source file line information for each symbol linked into the section.
  74. """
  75. scan_to_header(map_file, "Linker script and memory map")
  76. scan_to_header(map_file, "END GROUP")
  77. sections = {}
  78. section = None
  79. for line in map_file:
  80. # output section header, ie '.iram0.text 0x0000000040080400 0x129a5'
  81. RE_SECTION_HEADER = r"(?P<name>[^ ]+) +0x(?P<address>[\da-f]+) +0x(?P<size>[\da-f]+)$"
  82. m = re.match(RE_SECTION_HEADER, line)
  83. if m is not None: # start of a new section
  84. section = {
  85. "name" : m.group("name"),
  86. "address" : int(m.group("address"), 16),
  87. "size" : int(m.group("size"), 16),
  88. "sources" : [],
  89. }
  90. sections[section["name"]] = section
  91. continue
  92. # source file line, ie
  93. # 0x0000000040080400 0xa4 /home/gus/esp/32/idf/examples/get-started/hello_world/build/esp32/libesp32.a(cpu_start.o)
  94. RE_SOURCE_LINE = r".*? +0x(?P<address>[\da-f]+) +0x(?P<size>[\da-f]+) (?P<archive>.+\.a)\((?P<object_file>.+\.o)\)"
  95. m = re.match(RE_SOURCE_LINE, line)
  96. if section is not None and m is not None: # input source file details
  97. source = {
  98. "size" : int(m.group("size"), 16),
  99. "address" : int(m.group("address"), 16),
  100. "archive" : os.path.basename(m.group("archive")),
  101. "object_file" : m.group("object_file"),
  102. }
  103. source["file"] = "%s:%s" % (source["archive"], source["object_file"])
  104. section["sources"] += [ source ]
  105. return sections
  106. def sizes_by_key(sections, key):
  107. """ Takes a dict of sections (from load_sections) and returns
  108. a dict keyed by 'key' with aggregate output size information.
  109. Key can be either "archive" (for per-archive data) or "file" (for per-file data) in the result.
  110. """
  111. result = {}
  112. for section in sections.values():
  113. for s in section["sources"]:
  114. if not s[key] in result:
  115. result[s[key]] = {}
  116. archive = result[s[key]]
  117. if not section["name"] in archive:
  118. archive[section["name"]] = 0
  119. archive[section["name"]] += s["size"]
  120. return result
  121. def main():
  122. parser = argparse.ArgumentParser("idf_size - a tool to print IDF elf file sizes")
  123. parser.add_argument(
  124. '--toolchain-prefix',
  125. help="Triplet prefix to add before objdump executable",
  126. default=DEFAULT_TOOLCHAIN_PREFIX)
  127. parser.add_argument(
  128. 'map_file', help='MAP file produced by linker',
  129. type=argparse.FileType('r'))
  130. parser.add_argument(
  131. '--archives', help='Print per-archive sizes', action='store_true')
  132. parser.add_argument(
  133. '--files', help='Print per-file sizes', action='store_true')
  134. args = parser.parse_args()
  135. memory_config, sections = load_map_data(args.map_file)
  136. print_summary(memory_config, sections)
  137. if args.archives:
  138. print("Per-archive contributions to ELF file:")
  139. print_detailed_sizes(sections, "archive", "Archive File")
  140. if args.files:
  141. print("Per-file contributions to ELF file:")
  142. print_detailed_sizes(sections, "file", "Object File")
  143. def print_summary(memory_config, sections):
  144. def get_size(section):
  145. try:
  146. return sections[section]["size"]
  147. except KeyError:
  148. return 0
  149. # if linker script changes, these need to change
  150. total_iram = memory_config["iram0_0_seg"]["length"]
  151. total_dram = memory_config["dram0_0_seg"]["length"]
  152. used_data = get_size(".dram0.data")
  153. used_bss = get_size(".dram0.bss")
  154. used_dram = used_data + used_bss
  155. used_iram = sum( get_size(s) for s in sections.keys() if s.startswith(".iram0") )
  156. flash_code = get_size(".flash.text")
  157. flash_rodata = get_size(".flash.rodata")
  158. total_size = used_data + used_iram + flash_code + flash_rodata
  159. print("Total sizes:")
  160. print(" DRAM .data size: %7d bytes" % used_data)
  161. print(" DRAM .bss size: %7d bytes" % used_bss)
  162. print("Used static DRAM: %7d bytes (%7d available, %.1f%% used)" %
  163. (used_dram, total_dram - used_dram,
  164. 100.0 * used_dram / total_dram))
  165. print("Used static IRAM: %7d bytes (%7d available, %.1f%% used)" %
  166. (used_iram, total_iram - used_iram,
  167. 100.0 * used_iram / total_iram))
  168. print(" Flash code: %7d bytes" % flash_code)
  169. print(" Flash rodata: %7d bytes" % flash_rodata)
  170. print("Total image size:~%7d bytes (.bin may be padded larger)" % (total_size))
  171. def print_detailed_sizes(sections, key, header):
  172. sizes = sizes_by_key(sections, key)
  173. sub_heading = None
  174. headings = (header,
  175. "DRAM .data",
  176. "& .bss",
  177. "IRAM",
  178. "Flash code",
  179. "& rodata",
  180. "Total")
  181. print("%24s %10s %6s %6s %10s %8s %7s" % headings)
  182. for k in sorted(sizes.keys()):
  183. v = sizes[k]
  184. if ":" in k: # print subheadings for key of format archive:file
  185. sh,k = k.split(":")
  186. if sh != sub_heading:
  187. print(sh)
  188. sub_heading = sh
  189. data = v.get(".dram0.data", 0)
  190. bss = v.get(".dram0.bss", 0)
  191. iram = sum(t for (s,t) in v.items() if s.startswith(".iram0"))
  192. flash_text = v.get(".flash.text", 0)
  193. flash_rodata = v.get(".flash.rodata", 0)
  194. total = data + bss + iram + flash_text + flash_rodata
  195. print("%24s %10d %6d %6d %10d %8d %7d" % (k[:24],
  196. data,
  197. bss,
  198. iram,
  199. flash_text,
  200. flash_rodata,
  201. total))
  202. if __name__ == "__main__":
  203. main()