Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

  •  Cary Phillips
  •  Christina Tempelaar-Lietz
  •  John Mertic
  •  Joseph Goldstone
  •  Kimball Thurston
  •  Larry Gritz
  •  Nick Porcino
  •  Peter Hillman
  •  Rod Bogart

...

Community:

  •  Li Ji, ILM

Discussion:

  • Sparsely attended: public holiday in New Zealand (Kimball, Peter), Cary and Christina travelling.
  • RGB asks: is the order of attributes in the file significant? Consensus with LG: we don't think it was ever intended to be so. 
  • Li Ji: discussing benchmarking of the library, trying to organize such thing, and has academic collaborators. Among other things, need a good corpus of exr files that cover the right use cases, also must be open source content.
  • Nick points out that there is the openexr-images repo, too. But all agree that it doesn't necessarily cover good benchmarking cases. (E.g., no 4k images? the deep image examples are unrealistically simple?)
  • LG: Wish there were a freely usable production-level complexity test deep image. Nick: Generating images from ALAB or Intel Moore Lane scenes in DPEL would give good production complexity image for this and other modes (deep, albedo, cryptomatte ids, etc).
  • General talk: we've never had a perf benchmark for OpenEXR, no perf regression testing in any organized way, nor any way to know if over time we're improving perf (or making it worse) other than anecdotal/accidental noticing in studios.
  • Nick suggests Li Ji start a google doc to collect thoughts on the "spec" of what we need in benchmark test images and how to generate them. LJ: Already exists: https://github.com/lji-ilm/openexr-notes/blob/main/docs/benchmarkplanning.md