)]}'
{
  "id": "6c1742f97dd9917ef7e8bf94c38639594bd8d210",
  "repo": "external/github.com/python/cpython",
  "revision": "d858df20d0d82eeefbdbe9982cfd68207a0cd446",
  "path": "Parser/tokenizer.h"
}
