Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
96 tokens/sec
Gemini 2.5 Pro Premium
51 tokens/sec
GPT-5 Medium
36 tokens/sec
GPT-5 High Premium
34 tokens/sec
GPT-4o
96 tokens/sec
DeepSeek R1 via Azure Premium
91 tokens/sec
GPT OSS 120B via Groq Premium
466 tokens/sec
Kimi K2 via Groq Premium
148 tokens/sec
2000 character limit reached

Deep Polynomial Neural Networks (2006.13026v2)

Published 20 Jun 2020 in cs.LG, cs.CV, and stat.ML

Abstract: Deep Convolutional Neural Networks (DCNNs) are currently the method of choice both for generative, as well as for discriminative learning in computer vision and machine learning. The success of DCNNs can be attributed to the careful selection of their building blocks (e.g., residual blocks, rectifiers, sophisticated normalization schemes, to mention but a few). In this paper, we propose $\Pi$-Nets, a new class of function approximators based on polynomial expansions. $\Pi$-Nets are polynomial neural networks, i.e., the output is a high-order polynomial of the input. The unknown parameters, which are naturally represented by high-order tensors, are estimated through a collective tensor factorization with factors sharing. We introduce three tensor decompositions that significantly reduce the number of parameters and show how they can be efficiently implemented by hierarchical neural networks. We empirically demonstrate that $\Pi$-Nets are very expressive and they even produce good results without the use of non-linear activation functions in a large battery of tasks and signals, i.e., images, graphs, and audio. When used in conjunction with activation functions, $\Pi$-Nets produce state-of-the-art results in three challenging tasks, i.e. image generation, face verification and 3D mesh representation learning. The source code is available at \url{https://github.com/grigorisg9gr/polynomial_nets}.

Citations (82)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper's primary contribution is unclear due to the use of placeholder text in its provided template.
  • It appears to aim at integrating polynomial functions into deep neural network architectures for enhanced nonlinear modeling.
  • Key methodologies and experimental evaluations are not detailed, leaving actual performance and implementation insights undetermined.

Thank you for providing the document. It appears that the content you've provided is the LaTeX source code for a template file (IEEEtran.cls bare_adv.tex) used for preparing papers for IEEE Computer Society journals, not the actual content of the paper "Deep Polynomial Neural Networks" (2006.13026).

The file contains placeholder text such as "The abstract goes here", "Introduction text here", and "The conclusion goes here". Therefore, I cannot provide a summary of the technical contributions, implementation details, or practical applications of the "Deep Polynomial Neural Networks" paper based on this template file.

If you can provide the actual content of the paper "Deep Polynomial Neural Networks" (2006.13026), I would be happy to analyze it and discuss its practical implementation and application.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.