Files
trainlib/TODO.md
2026-03-22 20:24:10 -07:00

738 B

Long-term

  • Implement a dataloader in-house, with a clear, lightweight mechanism for collection-of-structures to structure-of-collections. For multi-proc handling (happens in torch's dataloader, as well as the BatchedDataset for two different purposes), we should rely on (a hopefully more stable) execlib.
  • Domains may be externalized (co3 or convlib)
  • Up next: CLI, fully JSON-ification of model selection + train.
  • Consider a "multi-train" alternative (or arg support in train()) for training many "rollouts" from the same base estimator (basically forks under different seeds). For architecture benchmarking above all, seeing average training behavior. Consider corresponding Plotter methods (error bars)