Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
133 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Representation Synthesis by Probabilistic Many-Valued Logic Operation in Self-Supervised Learning (2309.04148v3)

Published 8 Sep 2023 in cs.CV

Abstract: In this paper, we propose a new self-supervised learning (SSL) method for representations that enable logic operations. Representation learning has been applied to various tasks, such as image generation and retrieval. The logical controllability of representations is important for these tasks. Although some methods have been shown to enable the intuitive control of representations using natural languages as the inputs, representation control via logic operations between representations has not been demonstrated. Some SSL methods using representation synthesis (e.g., elementwise mean and maximum operations) have been proposed, but the operations performed in these methods do not incorporate logic operations. In this work, we propose a logic-operable self-supervised representation learning method by replacing the existing representation synthesis with the OR operation on the probabilistic extension of many-valued logic. The representations comprise a set of feature-possession degrees, which are truth values indicating the presence or absence of each feature in the image, and realize the logic operations (e.g., OR and AND). Our method can generate a representation that has the features of both representations or only those features common to both representations. In addition, the expression of the ambiguous presence of a feature is realized by indicating the feature-possession degree by the probability distribution of truth values of the many-valued logic. We showed that our method performs competitively in single and multi-label classification tasks compared with prior SSL methods using synthetic representations. Moreover, experiments on image retrieval using MNIST and PascalVOC showed that the representations of our method can be operated by OR and AND operations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. “Exploring simple siamese representation learning,” in CVPR, 2021.
  2. “Bootstrap your own latent-a new approach to self-supervised learning,” NeurIPS, 2020.
  3. “Momentum contrast for unsupervised visual representation learning,” in CVPR, 2020.
  4. “An empirical study of training self-supervised vision transformers,” in ICCV, 2021.
  5. “A simple framework for contrastive learning of visual representations,” in International conference on machine learning. PMLR, 2020, pp. 1597–1607.
  6. “Barlow twins: Self-supervised learning via redundancy reduction,” in ICML, 2021.
  7. “Understanding self-supervised learning dynamics without contrastive pairs,” in ICML, 2021.
  8. “How useful is self-supervised pretraining for visual tasks?,” in CVPR, 2020.
  9. “Unsupervised representation learning by predicting image rotations,” in ICLR, 2018.
  10. “Unsupervised learning of visual representations by solving jigsaw puzzles,” in ECCV, 2016.
  11. “Representation uncertainty in self-supervised learning as variational inference,” in ICCV, 2023.
  12. “Emerging properties in self-supervised vision transformers,” in ICCV, 2021.
  13. “Learning transferable visual models from natural language supervision,” in ICML, 2021.
  14. “Dinov2: Learning robust visual features without supervision,” arXiv:2304.07193, 2023.
  15. “Language models are few-shot learners,” NeurIPS, 2020.
  16. “Zero-shot text-to-image generation,” in ICML, 2021.
  17. “Un-mix: Rethinking image mixtures for unsupervised visual representation learning,” in AAAI, 2022.
  18. “A simple data mixing prior for improving self-supervised learning,” in CVPR, 2022.
  19. “i-mix: A domain-agnostic strategy for contrastive representation learning,” in ICLR, 2021.
  20. “Mixco: Mix-up contrastive learning for visual representation,” NeurIPS Workshop, 2020.
  21. “Mix-up self-supervised learning for contrast-agnostic applications,” in ICME, 2022.
  22. “mixup: Beyond empirical risk minimization,” in ICLR, 2018.
  23. “Cutmix: Regularization strategy to train strong classifiers with localizable features,” in ICCV, 2019.
  24. Kurt Godel, “Zum intuitionistischen aussagenkalkul,” Anzeiger Akademie der Wissenschaften Wien, mathematisch-naturwissenschaftliche Klasse, 1932.
  25. “beta-vae: Learning basic visual concepts with a constrained variational framework,” in ICLR, 2016.
  26. “Infogan: Interpretable representation learning by information maximizing generative adversarial nets,” NeurIPS, 2016.
  27. “Cross-encoder for unsupervised gaze representation learning,” in ICCV, 2021.
  28. “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
  29. “Guided variational autoencoder for disentanglement learning,” in CVPR, 2020.
  30. “Mixsiam: A mixture-based approach to self-supervised representation learning,” arXiv preprint arXiv:2111.02679, 2021.
  31. “Deep residual learning for image recognition,” in CVPR, 2016.
  32. “Imagenet: A large-scale hierarchical image database,” in CVPR, 2009.
  33. “Learning multiple layers of features from tiny images,” 2009.
  34. “Large batch training of convolutional networks,” arXiv preprint arXiv:1708.03888, 2017.
  35. “Microsoft coco: Common objects in context,” in ECCV, 2014.
  36. “The pascal visual object classes (voc) challenge,” IJCV, 2010.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com