Skip to content
/ fomoh Public

PyTorch-based library that implements nested forward AD and interfaces with PyTorch models.

License

Notifications You must be signed in to change notification settings

SRI-CSL/fomoh

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Fomoh Logo

Fomoh is a PyTorch-based library that implements nested forward AD and interfaces with PyTorch models.

  • Perform forward-mode automatic differentiation over functions in PyTorch to get first and second order information.
  • Train neural network models using forward-mode automatic differentiation.
  • Includes first-order FGD and our proposed approach of FoMoH-KD.

For additional details, please refer to our paper: Second-Order Forward-Mode Automatic Differentiation for Optimization

To run the code, from this directory location you can install the fomoh Python package:

pip install .

Blog posts and Notebook Tutorials:

  • For basic usage and an introduction please refer to this tutorial that covers the content of this notebook: Fomoh_Rosenblock_Example.ipynb
  • For training a neural network please refer to Training Neural Networks using Fomoh

Examples:

  • To run the Rosenbrock example from the paper, in the ./scripts/Rosenbrock location, run: python Rosenbrock_ND_dim_comparison.py --epochs 100 --dim-obj 10 --save-file ./plots/rosenbrock_comparison_plane_dim_comparison_10D.pt --newton
  • To run a logistic regression example from the paper, e.g. FoMoH, in the ./scripts/logistic_regression location, first make the new folder best_results, then run: ./train_model.sh 0.1362 1024 FoMoH 0
  • To run a cnn example from the paper, e.g. FoMoH, in the ./scripts/cnn location, first make the new folder best_results_3000, then run: ./train_model.sh 0.544 2048 FoMoH 0

How to cite?

Please consider citing the following paper if you use Fomoh in your research:

@article{cobb2024second,
  title={Second-Order Forward-Mode Automatic Differentiation for Optimization},
  author={Cobb, Adam D and Baydin, Atılım Güneş and Pearlmutter, Barak A. and Jha, Susmit},
  journal={arXiv preprint arXiv:2408.10419},
  year={2024}
}

Acknowledgements

This material is based upon work supported by the United States Air Force and DARPA under Contract No. FA8750-23-C-0519. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the United States Air Force and DARPA.

About

PyTorch-based library that implements nested forward AD and interfaces with PyTorch models.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published