CORE
πΊπ¦Β
Β make metadata, not war
Services
Research
Services overview
Explore all CORE services
Access to raw data
API
Dataset
FastSync
Content discovery
Recommender
Discovery
OAI identifiers
OAI Resolver
Managing content
Dashboard
Bespoke contracts
Consultancy services
Support us
Support us
Membership
Sponsorship
Community governance
Advisory Board
Board of supporters
Research network
About
About us
Our mission
Team
Blog
FAQs
Contact us
Sparsifying generalized linear models
Authors
Arun Jambulapati
James R. Lee
Yang P. Liu
Aaron Sidford
Publication date
29 November 2023
Publisher
View
on
arXiv
Abstract
We consider the sparsification of sums
F
:
R
n
β
R
F : \mathbb{R}^n \to \mathbb{R}
F
:
R
n
β
R
where
F
(
x
)
=
f
1
(
β¨
a
1
,
x
β©
)
+
β―
+
f
m
(
β¨
a
m
,
x
β©
)
F(x) = f_1(\langle a_1,x\rangle) + \cdots + f_m(\langle a_m,x\rangle)
F
(
x
)
=
f
1
β
(β¨
a
1
β
,
x
β©)
+
β―
+
f
m
β
(β¨
a
m
β
,
x
β©)
for vectors
a
1
,
β¦
,
a
m
β
R
n
a_1,\ldots,a_m \in \mathbb{R}^n
a
1
β
,
β¦
,
a
m
β
β
R
n
and functions
f
1
,
β¦
,
f
m
:
R
β
R
+
f_1,\ldots,f_m : \mathbb{R} \to \mathbb{R}_+
f
1
β
,
β¦
,
f
m
β
:
R
β
R
+
β
. We show that
(
1
+
Ξ΅
)
(1+\varepsilon)
(
1
+
Ξ΅
)
-approximate sparsifiers of
F
F
F
with support size
n
Ξ΅
2
(
log
β‘
n
Ξ΅
)
O
(
1
)
\frac{n}{\varepsilon^2} (\log \frac{n}{\varepsilon})^{O(1)}
Ξ΅
2
n
β
(
lo
g
Ξ΅
n
β
)
O
(
1
)
exist whenever the functions
f
1
,
β¦
,
f
m
f_1,\ldots,f_m
f
1
β
,
β¦
,
f
m
β
are symmetric, monotone, and satisfy natural growth bounds. Additionally, we give efficient algorithms to compute such a sparsifier assuming each
f
i
f_i
f
i
β
can be evaluated efficiently. Our results generalize the classic case of
β
p
\ell_p
β
p
β
sparsification, where
f
i
(
z
)
=
β£
z
β£
p
f_i(z) = |z|^p
f
i
β
(
z
)
=
β£
z
β£
p
, for
p
β
(
0
,
2
]
p \in (0, 2]
p
β
(
0
,
2
]
, and give the first near-linear size sparsifiers in the well-studied setting of the Huber loss function and its generalizations, e.g.,
f
i
(
z
)
=
min
β‘
{
β£
z
β£
p
,
β£
z
β£
2
}
f_i(z) = \min\{|z|^p, |z|^2\}
f
i
β
(
z
)
=
min
{
β£
z
β£
p
,
β£
z
β£
2
}
for
0
<
p
β€
2
0 < p \leq 2
0
<
p
β€
2
. Our sparsification algorithm can be applied to give near-optimal reductions for optimizing a variety of generalized linear models including
β
p
\ell_p
β
p
β
regression for
p
β
(
1
,
2
]
p \in (1, 2]
p
β
(
1
,
2
]
to high accuracy, via solving
(
log
β‘
n
)
O
(
1
)
(\log n)^{O(1)}
(
lo
g
n
)
O
(
1
)
sparse regression instances with
m
β€
n
(
log
β‘
n
)
O
(
1
)
m \le n(\log n)^{O(1)}
m
β€
n
(
lo
g
n
)
O
(
1
)
, plus runtime proportional to the number of nonzero entries in the vectors
a
1
,
β¦
,
a
m
a_1, \dots, a_m
a
1
β
,
β¦
,
a
m
β
Similar works
Full text
Available Versions
arXiv.org e-Print Archive
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:arXiv.org:2311.18145
Last time updated on 10/05/2024