## Abstract

The total influence of a function is a central notion in analysis of Boolean functions, and characterizing functions that have small total influence is one of the most fundamental questions associated with it. The KKL theorem and the Friedgut junta theorem give a strong characterization of such functions whenever the bound on the total influence is o(log n). However, both results become useless when the total influence of the function is omega(log n). The only case in which this logarithmic barrier has been broken for an interesting class of functions was proved by Bourgain and Kalai, who focused on functions that are symmetric under large enough subgroups of S {n}. In this paper, we build and improve on the techniques of the Bourgain-Kalai paper and establish new concentration results on the Fourier spectrum of Boolean functions with small total influence. Our results include: 1)A quantitative improvement of the Bourgain-Kalai result regarding the total influence of functions that are transitively symmetric. 2)A slightly weaker version of the Fourier-Entropy Conjecture of Friedgut and Kalai. Our result establishes new bounds on the Fourier entropy of a Boolean function f, as well as stronger bounds on the Fourier entropy of low-degree parts of f. In particular, it implies that the Fourier spectrum of a constant variance, Boolean function f is concentrated on 2{O(I[f] log I[f])} characters, improving an earlier result of Friedgut. Removing the log I[f] factor would essentially resolve the Fourier-Entropy Conjecture, as well as settle a conjecture of Mansour regarding the Fourier spectrum of polynomial size DNF formulas. Our concentration result for the Fourier spectrum of functions with small total influence also has new implications in learning theory. More specifically, we conclude that the class of functions whose total influence is at most K is agnostically learnable in time 2{O(K log K)} using membership queries. Thus, the class of functions with total influence O(log n log log n) is agnostically learnable in text{poly}(n) time.

Original language | English |
---|---|

Title of host publication | Proceedings - 2020 IEEE 61st Annual Symposium on Foundations of Computer Science, FOCS 2020 |

Publisher | IEEE Computer Society |

Pages | 247-258 |

Number of pages | 12 |

ISBN (Electronic) | 9781728196213 |

DOIs | |

State | Published - Nov 2020 |

Externally published | Yes |

Event | 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020 - Virtual, Durham, United States Duration: 16 Nov 2020 → 19 Nov 2020 |

### Publication series

Name | Proceedings - Annual IEEE Symposium on Foundations of Computer Science, FOCS |
---|---|

Volume | 2020-November |

ISSN (Print) | 0272-5428 |

### Conference

Conference | 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020 |
---|---|

Country/Territory | United States |

City | Virtual, Durham |

Period | 16/11/20 → 19/11/20 |

### Bibliographical note

Funding Information:This work was done while author Minzer was a member in the Institute for Advanced Study, Princeton, supported partially by NSF grant DMS-1638352 and Rothschild Fellowship. Author Safra was supported by the European Research Council (ERC) under the European Unions Horizon 2020 research and innovation programme (Grant agreement No. 835152).

Publisher Copyright:

© 2020 IEEE.

## Keywords

- Fourier analysis
- Fourier-Entropy Conjecture
- Learning sparse functions