Hardness Magnification for all Sparse NP Languages
Lijie Chen
MIT
Ce Jin
Tsinghua U.
Ryan Williams
MIT
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
Minimum Circuit Size Problem
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
Minimum Circuit Size Problem
MCSP 𝑠 𝑚 ∈ 𝐍𝐏; solvable in 𝑛 ⋅ 2 ෨𝑂 𝑠 𝑚 time.
We believe MCSP[𝑚10] ∉ 𝐏/𝐩𝐨𝐥𝐲!
(otherwise, no strong PRGs exist [Razborov-Rudich])
(MCSP[𝑚10] ∉ 𝐏/𝐩𝐨𝐥𝐲 ⇒ “𝐏/𝐩𝐨𝐥𝐲-natural property useful
against 𝐏/𝐩𝐨𝐥𝐲” )
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
Minimum Circuit Size Problem
MCSP 𝑠 𝑚 ∈ 𝐍𝐏; solvable in 𝑛 ⋅ 2 ෨𝑂 𝑠 𝑚 time.
We believe MCSP[𝑚10] ∉ 𝐏/𝐩𝐨𝐥𝐲!
(otherwise, no strong PRGs exist [Razborov-Rudich])
(MCSP[𝑚10] ∉ 𝐏/𝐩𝐨𝐥𝐲 ⇒ “𝐏/𝐩𝐨𝐥𝐲-natural property useful
against 𝐏/𝐩𝐨𝐥𝐲” )
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
Minimum Circuit Size Problem
MCSP 𝑠 𝑚 ∈ 𝐍𝐏; solvable in 𝑛 ⋅ 2 ෨𝑂 𝑠 𝑚 time.
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛 depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
(McKay-Murray-Williams’19)
We believe MCSP[𝑚10] ∉ 𝐏/𝐩𝐨𝐥𝐲!
(otherwise, no strong PRGs exist [Razborov-Rudich])
(MCSP[𝑚10] ∉ 𝐏/𝐩𝐨𝐥𝐲 ⇒ “𝐏/𝐩𝐨𝐥𝐲-natural property useful
against 𝐏/𝐩𝐨𝐥𝐲” )
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
Minimum Circuit Size Problem
MCSP 𝑠 𝑚 ∈ 𝐍𝐏; solvable in 𝑛 ⋅ 2 ෨𝑂 𝑠 𝑚 time.
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛 depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
(McKay-Murray-Williams’19)
“Hardness Magnification”
(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛 depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
(McKay-Murray-Williams’19)
Similar magnification results for MKtP (Minimum time-bounded
Kolmogorov Complexity, Kt 𝑥 )
(“Gap-MKtP[𝑎, 𝑏]”: distinguish between Kt 𝑥 ≤ 𝑎 and Kt 𝑥 ≥ 𝑏)
Hardness Magnification for MCSP
(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛 depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
(McKay-Murray-Williams’19)
Similar magnification results for MKtP (Minimum time-bounded
Kolmogorov Complexity, Kt 𝑥 )
Kt(x) = “measure of how much info needed to generate 𝑥 quickly”MKtP ≈ MCSP with “EXP-oracle gates”
(“Gap-MKtP[𝑎, 𝑏]”: distinguish between Kt 𝑥 ≤ 𝑎 and Kt 𝑥 ≥ 𝑏)
Hardness Magnification for MCSP
Similar magnification results for MKtP (Minimum time-bounded
Kolmogorov Complexity, Kt 𝑥 )
Kt(x) = “measure of how much info needed to generate 𝑥 quickly”MKtP ≈ MCSP with “EXP-oracle gates”
(“Gap-MKtP[𝑎, 𝑏]”: distinguish between Kt 𝑥 ≤ 𝑎 and Kt 𝑥 ≥ 𝑏)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚) doesn’t have 𝑛3polylog 𝑛-size
(De Morgan) Formulas, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
(Oliveira-Pich-Santhanam’19)
Hardness Magnification for MCSP(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛 depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
(McKay-Murray-Williams’19)
Similar magnification results for MKtP (Minimum time-bounded
Kolmogorov Complexity, Kt 𝑥 )
Kt(x) = “measure of how much info needed to generate 𝑥 quickly”MKtP ≈ MCSP with “EXP-oracle gates”
(“Gap-MKtP[𝑎, 𝑏]”: distinguish between Kt 𝑥 ≤ 𝑎 and Kt 𝑥 ≥ 𝑏)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚) doesn’t have 𝑛3polylog 𝑛-size
(De Morgan) Formulas, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
(Oliveira-Pich-Santhanam’19)
(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛 depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
(McKay-Murray-Williams’19)
Hardness Magnification for MCSP
(Other Hardness Magnification Results)𝑛1−𝜀-approximate Clique [Sri’03]
Average-case MCSP [OS’18]
𝑘-Vertex-Cover [OS’18]
low-depth circuit LBs for 𝐍𝐂𝟏 [AK’10,CT’19]
sublinear-depth circuit LBs for 𝐏 [LW’13]
…
Weak LB
Strong LB
Magnification
Suggests new
approaches to
proving strong
lower bounds?
How to view Hardness Magnification?
It is argued that HM can bypass the Natural Proof Barrier [Razborov-Rudich]
Weak LB
Strong LB
Magnification
Suggests new
approaches to
proving strong
lower bounds?
How to view Hardness Magnification?
It is argued that HM can bypass the Natural Proof Barrier [Razborov-Rudich]
• A heuristic argument [AK’10, OS’18]: HM seems to yield strong LBs only for certain functions, not for most of them (violating the “largeness” condition of Natural Proofs)
Weak LB
Strong LB
Magnification
Suggests new
approaches to
proving strong
lower bounds?
How to view Hardness Magnification?
It is argued that HM can bypass the Natural Proof Barrier [Razborov-Rudich]
• A heuristic argument [AK’10, OS’18]: HM seems to yield strong LBs only for certain functions, not for most of them (violating the “largeness” condition of Natural Proofs)
• A real theorem [CHOPRS to appear in ITCS’20] In some cases, the required weak LB actually implies the non-existence of natural proofs
Weak LB
Strong LB
Magnification
Suggests new
approaches to
proving strong
lower bounds?
How to view Hardness Magnification?
Extending Known Lower Bounds?
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
[OPS’19]
Extending Known Lower Bounds?
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
We know how to prove 𝑛1.99-size formula
lower bound for Gap-MKtP ! [OPS’19]
[OPS’19]
Extending Known Lower Bounds?
Can we improve it by a factor of 𝑛1+𝜀?
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
We know how to prove 𝑛1.99-size formula
lower bound for Gap-MKtP ! [OPS’19]
[OPS’19]
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Extending Known Lower Bounds?
Formula-⊕: De Morgan Formulas
where each leaf node computes
XOR of a subset of input bits
[OPS’19]
Known LB against Formula-⊕(Tal’16) :
𝐅𝟐-Inner-Product ∉ 𝐅𝐨𝐫𝐦𝐮𝐥𝐚-⊕ 𝒏𝟐/ 𝐩𝐨𝐥𝐲𝐥𝐨𝐠 𝒏
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Extending Known Lower Bounds?
Stronger LB than required
Known LB against Formula-⊕(Tal’16) :
𝐅𝟐-Inner-Product ∉ 𝐅𝐨𝐫𝐦𝐮𝐥𝐚-⊕ 𝒏𝟐/ 𝐩𝐨𝐥𝐲𝐥𝐨𝐠 𝒏
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Stronger LB than required
Extending Known Lower Bounds?
Much easier than Gap-MKtP??!
Known LB against Formula-⊕(Tal’16) :
𝐅𝟐-Inner-Product ∉ 𝐅𝐨𝐫𝐦𝐮𝐥𝐚-⊕ 𝒏𝟐/ 𝐩𝐨𝐥𝐲𝐥𝐨𝐠 𝒏
Much easier than Gap-MKtP??!
Can we adapt the proof techniques to Gap-MKtP?
(Input length 𝑛 = 2𝑚)
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formulas, then
𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Stronger LB than required
Extending Known Lower Bounds?
Weak LB
Strong LB
Magnification
Indicates proving
“weak” lower bounds
are even harder than
previously thought??
Suggests new
approaches to
proving strong
lower bounds?
How to view Hardness Magnification?
Weak LB
Strong LB
Magnification
Indicates proving
“weak” lower bounds
are even harder than
previously thought??
Suggests new
approaches to
proving strong
lower bounds?
•Hardness magnification:
Proving almost-linear size lower bounds is already as hard as proving super-polynomial lower bounds…
How to view Hardness Magnification?
What is special about MCSP and MKtP?
Is it because they are “compression” problems?
(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formula, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
Observation: MCSP[𝑚10] and MKtP[𝑚10]
are sparse languages!
MCSP 𝑠(𝑚) is 2 ෨𝑂 𝑠(𝑚) -sparse;
there are at most 2 ෨𝑂 𝑠(𝑚) many circuits!
(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formula, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Problem: MCSP[𝑠 𝑚 ]·Given: 𝑓: {0,1}𝑚 → {0,1} as a truth table of length 𝑛 = 2𝑚
·Decide: Does 𝑓 have a circuit of size at most 𝑠(𝑚)?
What is special about MCSP and MKtP?
Is it because they are “compression” problems?
Our result: Hardness magnification holds for
all sparse 𝐍𝐏 languages!
(Input length 𝑛 = 2𝑚)
If MCSP[𝑚10] doesn’t have circuits of 𝑛 ⋅ polylog 𝑛 size and polylog 𝑛depth, then 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲.
If Gap-MKtP 𝑚10, 𝑚10 + 𝑂(𝑚)· doesn’t have 𝑛3polylog 𝑛-size (De Morgan) Formula, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
· doesn’t have 𝑛 ⋅ polylog 𝑛-size Formula-⊕, then 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏.
Observation: MCSP[𝑚10] and MKtP[𝑚10]
are sparse languages!
What is special about MCSP and MKtP?
Is it because they are “compression” problems?
MCSP 𝑠(𝑚) is 2 ෨𝑂 𝑠(𝑚) -sparse;
there are at most 2 ෨𝑂 𝑠(𝑚) many circuits!
HM for all sparse NP languages
Let 𝐿 be any 2𝑛𝑜 1
-sparse 𝐍𝐏 language.
· If 𝐿 doesn’t have 𝒏𝟏.𝟎𝟏-size circuits, then for all 𝒌, 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 .
· If 𝐿 doesn’t have 𝒏𝟑.𝟎𝟏-size formulas, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size formulas.
· If 𝐿 doesn’t have 𝒏𝟐.𝟎𝟏-size branching programs, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size branching programs.
Theorem 1:
HM for all sparse NP languages
Similar results for other models!
Let 𝐿 be any 2𝑛𝑜 1
-sparse 𝐍𝐏 language.
· If 𝐿 doesn’t have 𝒏𝟏.𝟎𝟏-size circuits, then for all 𝒌, 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 .
· If 𝐿 doesn’t have 𝒏𝟑.𝟎𝟏-size formulas, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size formulas.
· If 𝐿 doesn’t have 𝒏𝟐.𝟎𝟏-size branching programs, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size branching programs.
Theorem 1:
HM for all sparse NP languages
Similar results for other models!
Let 𝐿 be any 2𝑛𝑜 1
-sparse 𝐍𝐏 language.
· If 𝐿 doesn’t have 𝒏𝟏.𝟎𝟏-size circuits, then for all 𝒌, 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 .
· If 𝐿 doesn’t have 𝒏𝟑.𝟎𝟏-size formulas, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size formulas.
· If 𝐿 doesn’t have 𝒏𝟐.𝟎𝟏-size branching programs, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size branching programs.
Theorem 1:
Compared with [MMW’19]: Our techniques yield weaker
consequences (e.g. they get 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲), but apply to more
restricted models.
HM for all sparse NP languages
Similar results for other models!
Let 𝐿 be any 2𝑛𝑜 1
-sparse 𝐍𝐏 language.
· If 𝐿 doesn’t have 𝒏𝟏.𝟎𝟏-size circuits, then for all 𝒌, 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 .
· If 𝐿 doesn’t have 𝒏𝟑.𝟎𝟏-size formulas, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size formulas.
· If 𝐿 doesn’t have 𝒏𝟐.𝟎𝟏-size branching programs, then
for all 𝒌, 𝐍𝐏 doesn’t have 𝒏𝒌-size branching programs.
Theorem 1:
Compared with [MMW’19]: Our techniques yield weaker
consequences (e.g. they get 𝐍𝐏 ⊄ 𝐏/𝐩𝐨𝐥𝐲), but apply to more
restricted models.
(Best known formula LB: 𝒏𝟑/𝐩𝐨𝐥𝐲𝐥𝐨𝐠 𝒏) [Håstad 90s, Tal]
(Best known branching program LB: 𝒏𝟐/𝐩𝐨𝐥𝐲𝐥𝐨𝐠 𝒏) [Nečiporuk 60s]
Theorem 2:
(Input length 𝑛 = 2𝑚)
Similar results for other models!
Hardness Magnification for MCSP
If MCSP 𝑚10 doesn’t have 𝑛3polylog 𝑛-size (De Morgan)
Formulas, then 𝐏𝐒𝐏𝐀𝐂𝐄 ⊄ (nonuniform) 𝐍𝐂𝟏.
Theorem 2:
(Input length 𝑛 = 2𝑚)
Similar results for other models!
Hardness Magnification for MCSP
If MCSP 𝑚10 doesn’t have 𝑛3polylog 𝑛-size (De Morgan)
Formulas, then 𝐏𝐒𝐏𝐀𝐂𝐄 ⊄ (nonuniform) 𝐍𝐂𝟏.
Best MCSP lower bound (Cheraghchi-Kabanets-Lu-Myrisiotis’19) :
MCSP[2𝑚/10𝑚] requires 𝑛3−𝑜 1 -size formulas.
(doesn’t work for 𝑚10…)
Theorem 2:
Similar results for MKtP 𝑚10 and 𝐄𝐗𝐏 ⊄ 𝐍𝐂𝟏 (improving upon
[OPS’19] which required lower bounds for Gap-MKtP)
(Input length 𝑛 = 2𝑚)
Similar results for other models!
Hardness Magnification for MCSP
If MCSP 𝑚10 doesn’t have 𝑛3polylog 𝑛-size (De Morgan)
Formulas, then 𝐏𝐒𝐏𝐀𝐂𝐄 ⊄ (nonuniform) 𝐍𝐂𝟏.
Best MCSP lower bound (Cheraghchi-Kabanets-Lu-Myrisiotis’19) :
MCSP[2𝑚/10𝑚] requires 𝑛3−𝑜 1 -size formulas.
(doesn’t work for 𝑚10…)
Algorithms with small non-uniformity
Theorem 3:
Let 𝐿 be a 2𝑛𝑜 1
-sparse 𝐍𝐏 language not computable by an
𝒏𝟏.𝟎𝟏-time 𝒏𝟎.𝟎𝟏-space deterministic algorithm with 𝒏𝟎.𝟎𝟏 bits
of advice, then 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 for all 𝒌.
Algorithms with small non-uniformity
Theorem 3:
There is a 𝟐𝒏𝟎.𝟎𝟏
⋅ 𝒏 -sparse language 𝐿 ∈ 𝐃𝐓𝐈𝐌𝐄[ ෨𝑂(𝑛1.01)],
not computable by an 𝒏𝟏.𝟎𝟏-time deterministic algorithm with
𝒏𝟎.𝟎𝟏 bits of advice.
The hypothesis is “close” to what we can prove!
Let 𝐿 be a 2𝑛𝑜 1
-sparse 𝐍𝐏 language not computable by an
𝒏𝟏.𝟎𝟏-time 𝒏𝟎.𝟎𝟏-space deterministic algorithm with 𝒏𝟎.𝟎𝟏 bits
of advice, then 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 for all 𝒌.
(Adaptation of time hierarchy theorem)
Algorithms with small non-uniformity
Theorem 3:
There is a 𝟐𝒏𝟎.𝟎𝟏
⋅ 𝒏 -sparse language 𝐿 ∈ 𝐃𝐓𝐈𝐌𝐄[ ෨𝑂(𝑛1.01)],
not computable by an 𝒏𝟏.𝟎𝟏-time deterministic algorithm with
𝒏𝟎.𝟎𝟏 bits of advice.
The hypothesis is “close” to what we can prove!
Let 𝐿 be a 2𝑛𝑜 1
-sparse 𝐍𝐏 language not computable by an
𝒏𝟏.𝟎𝟏-time 𝒏𝟎.𝟎𝟏-space deterministic algorithm with 𝒏𝟎.𝟎𝟏 bits
of advice, then 𝐍𝐏 ⊄ 𝐒𝐈𝐙𝐄 𝒏𝒌 for all 𝒌.
(Adaptation of time hierarchy theorem)
Can we make it sparser?
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
Let 𝐿 be any 2𝑛𝑜 1
-sparse 𝐍𝐏 language.
· If 𝐿 doesn’t have 𝒏𝟑.𝟎𝟏-size formulas, then for every 𝒌, 𝐍𝐏doesn’t have 𝒏𝒌-size formulas.
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
Intuition
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
N N N N N Y N N N N Y N N Y N N
(Sparse) 𝐿 ∩ {0,1}𝑛
0000 0001 1110 1111…………
Intuition
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
N N N N N Y N N N N Y N N Y N N
Y Y N Y
(Sparse) 𝐿 ∩ {0,1}𝑛
(Dense) Auxiliary 𝐍𝐏 language 𝐾 ∩ {0,1}𝑛0.001/𝑘
(“kernel problem”)
0000 0001 1110 1111…………
00 01 10 11
Intuition
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
N N N N N Y N N N N Y N N Y N N
Y Y N Y
(Sparse) 𝐿 ∩ {0,1}𝑛
(Dense) Auxiliary 𝐍𝐏 language 𝐾 ∩ {0,1}𝑛0.001/𝑘
(“kernel problem”)
0000 0001 1110 1111…………
00 01 10 11
We will construct cubic-size
formulas for 𝐿, with oracle
access to 𝐾.
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
Set 𝑡 ≔ 𝒏𝟎.𝟎𝟎𝟏/𝒌 > log Sparsity of 𝐿 .
Standard hashing tricks imply:
There is a hash function 𝐻𝑠: {0,1}𝑛 → {0,1}𝑂(𝑡) that is
· Perfect: maps YES-instances of 𝐿 into distinct images
· described by an 𝑂 𝑡 -bit seed 𝑠
· linear over 𝐅2 (there is a “correct”
seed 𝑠 that makes the
hash function 𝐻𝑠 perfect)
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
Set 𝑡 ≔ 𝒏𝟎.𝟎𝟎𝟏/𝒌 > log Sparsity of 𝐿 .
Standard hashing tricks imply:
There is a hash function 𝐻𝑠: {0,1}𝑛 → {0,1}𝑂(𝑡) that is
· Perfect: maps YES-instances of 𝐿 into distinct images
· described by an 𝑂 𝑡 -bit seed 𝑠
· linear over 𝐅2
(Construction: pick some coordinates
from the Error Correcting Code)
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
(𝑡 ≔ 𝑛0.001/𝑘 > log Sparsity of 𝐿 )
(Perfect hash 𝐻𝑠: {0,1}𝑛 → {0,1}𝑂(𝑡) with seed s = 𝑂 𝑡 )
Input: Hash seed 𝑠, hash value ℎ, index 𝑖 ∈ [𝑛]Output: The 𝑖-th bit of some 𝑥 ∈ 𝐿 such that 𝐻𝑠 𝑥 = ℎ.
Define an 𝑂(𝑡)-input auxiliary NP problem 𝐾 (“kernel problem”):
For the “correct” 𝑠, this 𝑥is unique
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
𝐍𝐏 has 𝑛𝑘-size formulas ⇒ 𝐾 has formulas of size 𝑛0.001!
Input: Hash seed 𝑠, hash value ℎ, index 𝑖 ∈ [𝑛]Output: The 𝑖-th bit of some 𝑥 ∈ 𝐿 such that 𝐻𝑠 𝑥 = ℎ.
Define an 𝑂(𝑡)-input auxiliary NP problem 𝐾 (“kernel problem”):
(𝑡 ≔ 𝑛0.001/𝑘 > log Sparsity of 𝐿 )
(Perfect hash 𝐻𝑠: {0,1}𝑛 → {0,1}𝑂(𝑡) with seed s = 𝑂 𝑡 )
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
𝐍𝐏 has 𝑛𝑘-size formulas ⇒ 𝐾 has formulas of size 𝑛0.001!
On input (𝑠, ℎ, 𝑖), guess (𝑥, 𝑦), where 𝑦 witnesses 𝑥 ∈ 𝐿.Accept 𝑥𝑖 = 1 and 𝐻𝑠(𝑥) = ℎ.
Input: Hash seed 𝑠, hash value ℎ, index 𝑖 ∈ [𝑛]Output: The 𝑖-th bit of some 𝑥 ∈ 𝐿 such that 𝐻𝑠 𝑥 = ℎ.
Define an 𝑂(𝑡)-input auxiliary NP problem 𝐾 (“kernel problem”):
(𝑡 ≔ 𝑛0.001/𝑘 > log Sparsity of 𝐿 )
(Perfect hash 𝐻𝑠: {0,1}𝑛 → {0,1}𝑂(𝑡) with seed s = 𝑂 𝑡 )
Proof of Theorem 1.2
Assume: 𝐍𝐏 has 𝑛𝑘-size formulas for some 𝑘.
Goal: Design 𝑛3.01-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
Claim: for the “correct” 𝑠, the following decides L:
On input 𝑥 ∈ {0,1}𝑛, accept iff:
∀𝑖 ∈ 𝑛 , 𝐾 𝑠,𝐻𝑠 𝑥 , 𝑖 = 𝑥𝑖
Input: Hash seed 𝑠, hash value ℎ, index 𝑖 ∈ [𝑛]Output: The 𝑖-th bit of some 𝑥 ∈ 𝐿 such that 𝐻𝑠 𝑥 = ℎ.
Define an 𝑂(𝑡)-input auxiliary NP problem 𝐾 (“kernel problem”):
(𝑡 ≔ 𝑛0.001/𝑘 > log Sparsity of 𝐿 )
(Perfect hash 𝐻𝑠: {0,1}𝑛 → {0,1}𝑂(𝑡) with seed s = 𝑂 𝑡 )
On input 𝑥 ∈ {0,1}𝑛, accept iff:
∀𝑖 ∈ 𝑛 , 𝐾 𝑠,𝐻𝑠 𝑥 , 𝑖 = 𝑥𝑖
AND
𝐾
𝑠 1
=?
𝑥1
𝐻𝑠(𝑥)
⋯=?
𝐾
𝑠 𝑛𝐻𝑠(𝑥)
𝑥𝑛
𝐾: 𝑛0.001-size
Goal: Design 𝒏𝟑.𝟎𝟏-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
On input 𝑥 ∈ {0,1}𝑛, accept iff:
∀𝑖 ∈ 𝑛 , 𝐾 𝑠,𝐻𝑠 𝑥 , 𝑖 = 𝑥𝑖
AND
𝐾
𝑠 1
=?
𝑥1
𝐻𝑠(𝑥)
⋯=?
𝐾
𝑠 𝑛𝐻𝑠(𝑥)
𝑥𝑛
𝐾: 𝑛0.001-size
Hash seed 𝑠hardwired into formulas
Goal: Design 𝒏𝟑.𝟎𝟏-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
On input 𝑥 ∈ {0,1}𝑛, accept iff:
∀𝑖 ∈ 𝑛 , 𝐾 𝑠,𝐻𝑠 𝑥 , 𝑖 = 𝑥𝑖
AND
𝐾
𝑠 1
=?
𝑥1
𝐻𝑠(𝑥)⊕ ⋯⊕
⋯=?
𝐾
𝑠 𝑛𝐻𝑠(𝑥)⊕ ⋯⊕
𝑥𝑛
𝐾: 𝑛0.001-size
Hash seed 𝑠hardwired into formulas
Each bit of 𝐻𝑠 𝑥 is
an XOR function
(implemented by De
Morgan formulas of
size 𝑂 𝑛2 )
Goal: Design 𝒏𝟑.𝟎𝟏-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
On input 𝑥 ∈ {0,1}𝑛, accept iff:
∀𝑖 ∈ 𝑛 , 𝐾 𝑠,𝐻𝑠 𝑥 , 𝑖 = 𝑥𝑖
AND
𝐾
𝑠 1
=?
𝑥1
𝐻𝑠(𝑥)⊕ ⋯⊕
⋯=?
𝐾
𝑠 𝑛𝐻𝑠(𝑥)⊕ ⋯⊕
𝑥𝑛
𝐾: 𝑛0.001-size
Hash seed 𝑠hardwired into formulas
Each bit of 𝐻𝑠 𝑥 is
an XOR function
(implemented by De
Morgan formulas of
size 𝑂 𝑛2 )
Total size
𝒏 ⋅ 𝒏𝟎.𝟎𝟎𝟏 ⋅ 𝑶 𝒏𝟐
Goal: Design 𝒏𝟑.𝟎𝟏-size formulas for 2𝑛𝑜 1
-sparse 𝐍𝐏 language 𝐿.
• Are there any other natural sparse NP languages for which one can prove some concrete lower bounds?
Open Problems
• Are there any other natural sparse NP languages for which one can prove some concrete lower bounds?
• Is it possible to show hardness magnification results for “denser” variants of MCSP or MKtP, such as MCSP[𝟐𝒎/𝒎𝟑]?
Open Problems
• Are there any other natural sparse NP languages for which one can prove some concrete lower bounds?
• Is it possible to show hardness magnification results for “denser” variants of MCSP or MKtP, such as MCSP[𝟐𝒎/𝒎𝟑]?
Open Problems
Thank you!