Show simple item record

dc.contributor.authorGupta, Deepak Kumar
dc.contributor.authorMago, Gowreesh
dc.contributor.authorChavan, Arnav
dc.contributor.authorPrasad, Dilip K.
dc.contributor.authorThomas, Rajat Mani
dc.date.accessioned2025-03-20T09:57:52Z
dc.date.available2025-03-20T09:57:52Z
dc.date.issued2024
dc.description.abstractTraditional deep learning models are trained and tested on relatively low-resolution images (< 300 px), and cannot be directly operated on large-scale images due to compute and memory constraints. We propose Patch Gradient Descent (PatchGD), an effective learning strategy that allows us to train the existing CNN and transformer architectures (hereby referred to as deep learning models) on large-scale images in an end-to-end manner. PatchGD is based on the hypothesis that instead of performing gradient-based updates on an entire image at once, it should be possible to achieve a good solution by performing model updates on only small parts of the image at a time, ensuring that the majority of it is covered over the course of iterations. PatchGD thus extensively enjoys better memory and compute efficiency when training models on large-scale images. PatchGD is thoroughly evaluated on PANDA, UltraMNIST, TCGA, and ImageNet datasets with ResNet50, MobileNetV2, ConvNeXtV2, and DeiT models under different memory constraints. Our evaluation clearly shows that PatchGD is much more stable and efficient than the standard gradient-descent method in handling large images, especially when the compute memory is limited. Code is available at https://github.com/nyunAI/PatchGD.en_US
dc.descriptionSource at <a href=https://www.jmlr.org/tmlr/index.html>https://www.jmlr.org/tmlr/index.html</a>.en_US
dc.identifier.citationGupta, Mago, Chavan, Prasad, Thomas. Pushing the Limits of Gradient Descent for Efficient Learning on Large Images. Transactions on Machine Learning Research (TMLR). 2024;2024en_US
dc.identifier.cristinIDFRIDAID 2367408
dc.identifier.issn2835-8856
dc.identifier.urihttps://hdl.handle.net/10037/36731
dc.language.isoengen_US
dc.publisherTMLRen_US
dc.relation.journalTransactions on Machine Learning Research (TMLR)
dc.relation.urihttps://openreview.net/pdf?id=6dS1jhdemD
dc.rights.accessRightsopenAccessen_US
dc.rights.holderCopyright 2024 The Author(s)en_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0en_US
dc.rightsAttribution 4.0 International (CC BY 4.0)en_US
dc.titlePushing the Limits of Gradient Descent for Efficient Learning on Large Imagesen_US
dc.type.versionpublishedVersionen_US
dc.typeJournal articleen_US
dc.typeTidsskriftartikkelen_US
dc.typePeer revieweden_US


File(s) in this item

Thumbnail

This item appears in the following collection(s)

Show simple item record

Attribution 4.0 International (CC BY 4.0)
Except where otherwise noted, this item's license is described as Attribution 4.0 International (CC BY 4.0)