Segment Anything in Food Images

Abstract

This paper introduces a new approach for food image segmentation utilizing the Segment Anything Model (SAM), with the additional refinement achieved through fine-tuning with Low-Rank Adaptation layers (LoRA). The segmentation task involves generating a binary mask for food in RGB images, with pixels categorized as background or food. We conduct various experiments to assess and compare the performance of our proposed method with previous approaches. Our findings indicate that our method consistently outperforms other techniques, achieving an accuracy of 94.14%. The improved accuracy of our approach highlights its potential for various applications in food image analysis, contributing to the advancement of computer vision techniques in the realm of food recognition and segmentation.

Cite

Text

Alahmari et al. "Segment Anything in Food Images." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024. doi:10.1109/CVPRW63382.2024.00375

Markdown

[Alahmari et al. "Segment Anything in Food Images." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2024.](https://mlanthology.org/cvprw/2024/alahmari2024cvprw-segment/) doi:10.1109/CVPRW63382.2024.00375

BibTeX

@inproceedings{alahmari2024cvprw-segment,
  title     = {{Segment Anything in Food Images}},
  author    = {Alahmari, Saeed S. and Gardner, Michael and Salem, Tawfiq},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2024},
  pages     = {3715-3720},
  doi       = {10.1109/CVPRW63382.2024.00375},
  url       = {https://mlanthology.org/cvprw/2024/alahmari2024cvprw-segment/}
}