\documentclass[11pt]{article} \usepackage{chao} \usepackage{natbib} \title{Sparsest Cut} \author{} \date{} \DeclareMathOperator*{\opt}{OPT} \DeclareMathOperator*{\len}{len} \newcommand{\scut}{\textsc{Sparsest Cut}} \newcommand{\nonuscut}{\textsc{Non-Uniform Sparsest Cut}} \newcommand{\expansion}{\textsc{Expansion}} \begin{document} \maketitle % \paragraph{Research problem} What is the best possible approximation rate of linear programming based approximation algorithms for \scut? What about algorithms for planar graphs? \section{Introduction} % Requirement: The introduction may have the following parts: % Establishing the importance (Background), % literature review (previous research contributions) % Gap/the specific problem % The present study (the research topic, the research purpose ) % The information should be presented in coherent paragraphs. \scut{} is a fundamental problem in graph algorithms with connections to various cut related problems. \begin{problem}[\nonuscut] The input is a graph $G=(V,E)$ with edge capacities $c:E\to \R_+$ and a set of vertex pairs $\{s_1,t_1\},\dots,\{s_k,t_k\}$ along with demand values $D_1,\dots,D_k\in \R_+$. The goal is to find a cut $\delta(S)$ of $G$ such that $\frac{c(\delta(S))}{\sum_{i:|S\cap \set{s_i,t_i}|=1}D_i}$ is minimized. \end{problem} In other words, \nonuscut{} finds the cut that minimizes its capacity divided by the sum of demands of the vertex pairs it separates. There are two important varients of \nonuscut{}. Note that we always consider unordered pair $\{s_i,t_i\}$, i.e., we do not distinguish $\{s_i,t_i\}$ and $\{t_i,s_i\}$. \scut{} is the uniform version of \nonuscut{}. The demand is 1 for every possible vertex pair $\{s_i,t_i\}$. In this case, we can remove from the input the pairs and demands. The goal becomes to minimize $\frac{c(\delta(S))}{|S||V\setminus S|}$. \expansion{} further simplifies the objective of \scut{} to $\min_{|S|\leq n/2}\frac{c(\delta(S))}{|S|}$. \subsection{importance and connections} These problems are interesting since they are related to central concepts in graph theory and help to design algorithms for hard problems on graph. One connections is expander graphs. The importance of expander graphs is thoroughly surveyed in \citep{hoory_expander_2006}. The optimum of \expansion{} is also known as Cheeger constant or conductance of a graph. \scut{} provides a 2-approximation of Cheeger constant, which is especially important in the context of expander graphs as it is a way to measure the edge expansion of a graph. \nonuscut{} is related to other cut problems such as Multicut and Balanced Separator. From a more mathematical perspective, the techniques developed for approximating \scut{} are deeply related to metric embedding, which is another fundamental problem in geometry. Besides theoretical interests, \scut{} is useful in practical scenarios such as in image segmentation and in some machine leaning algorithms. \subsection{related works} \nonuscut{} is APX-hard \citep{juliaJACMapxhard} and, assuming the Unique Game Conjecture, has no polynomial time constant factor aproximation algorithm\citep{chawla_hardness_2005}. \scut{} admits no PTAS \citep{uniformhardnessFocs07}, assuming a widely believed conjecture. The currently best approximation algorithm has ratio $O(\sqrt{\log n})$ and running time $\tilde{O}(n^2)$ \citep{arora_osqrtlogn_2010}. Prior to this currently optimal result, there is a long line of research optimizing both the approximation ratio and the complexity, see \citep{arora_expander_2004,leighton_multicommodity_1999}. There are also works concerning approximating \scut{} on special graph classes such as planar graphs \citep{lee_genus_2010}, graphs with low treewidth \citep{chlamtac_approximating_2010,gupta2013sparsestcutboundedtreewidth, Chalermsook_2024}. For an overview of the LP methods for \scut{}, see \citep{sparsest_cut_notes}. \subsection{open problems} One major open problem for \scut{} is the best approximation ratio for planar graphs. It is conjectured that the ratio for planar graphs is $O(1)$ but currently the best lowerbound is $O(\sqrt{\log n})$. For graphs treewidth $k$, an open problem is that whether there is a 2 approximation algorithm that runs in $2^{O(k)}\poly(n)$. \section{Literature Review} % Requirement: summarize previous research contributions and identify the gap or the specific problem The seminal work of \cite{leighton_multicommodity_1999} starts this line of research. They studied multicommodity flow problem and proved a $O(\log n)$ flow-cut gap. They also developed $O(\log n)$ approximation algorithm for multicommodity flow problems, which can imply $O(\log n)$ approximation for \scut{} and $O(\log^2 n)$ approximation for \nonuscut{}. The technique is called region growing. They also discovered a lowerbound of $\Omega(\log n)$ via expanders. Note that any algorithm achieving the $O(\log n)$ flow cut gap implies an $O(\log^2 n)$ approximation for \nonuscut{}, but better ratio is still possible through other methods. This paper showed that $O(\log^2 n)$ is the best approximation we can achieve using flow-cut gap. For \nonuscut{} \citep{leighton_multicommodity_1999} only guarantees a $O(\log^2 n)$ approximation. This is further improved by \citep{Linial_London_Rabinovich_1995} and \citep{lognGapAumann98}. \cite{lognGapAumann98} applied metric embedding to \nonuscut{} and obtained a $O(\log n)$ approximation. The connections between metric embedding and \nonuscut{} is influential. \nonuscut{} can be formulated as an integer program. \citeauthor{lognGapAumann98} considered the metric relaxation of the IP. They observed that \nonuscut{} is polynomial time solvable for trees and more generally for all $\ell_1$ metrics. The $O(\log n)$ approximation follows from the $O(\log n)$ distortion in the metric embedding theorem. \citep{arora_expander_2004} and \citep{arora_osqrtlogn_2010} further improved the approximation ratio for \scut{} to $O(\sqrt{\log n})$ via semidefinite relaxation. This is currently the best approximation ratio for \scut{}. There is also plenty of research concerning \scut{} on some graph classes, for example \citep{bonsma_complexity_2012}. One of the most popular class is graphs with constant treewidth. \citep{Chalermsook_2024} gave a $O(k^2)$ approximation algorithm with complexity $2^{O(k)}\poly(n)$. \citep{Cohen-Addad_Mömke_Verdugo_2024} obtained a 2-approximation algorithm for sparsest cut in treewidth $k$ graph with running time $2^{2^{O(k)}}\poly(n)$. \scut{} is easy on trees and the flow-cut gap is 1 for trees. One explaination mentioned in \citep{sparsest_cut_notes} is that shortest path distance in trees is an $\ell_1$ metric. There are works concerning planar graphs and more generally graphs with constant genus. \citep{leighton_multicommodity_1999} provided a $\Omega(\log n)$ lowerbound for flow-cut gap for \scut{}. However, it is conjectured that the gap is $O(1)$, while currently the best upperbound is still $O(\sqrt{\log n})$ \citep{rao_small_1999}. For graphs with constant genus, \citep{lee_genus_2010} gives a $O(\sqrt{\log g})$ approximation for \scut{}, where $g$ is the genus of the input graph. For flow-cut gap in planar graphs the techniques are mainly related to metric embedding theory. \section{LP} \begin{minipage}{0.47\linewidth} \begin{equation}\label{IP} \begin{aligned} \min& & \frac{\sum_e c_e x_e}{\sum_{i} D_i y_i}& & &\\ s.t.& & \sum_{e\in p} x_e&\geq y_i & &\forall p\in \mathcal{P}_{s_i,t_i}, \forall i\\ & & x_e,y_i&\in \{0,1\} \end{aligned} \end{equation} \end{minipage} \begin{minipage}{0.47\linewidth} \begin{equation}\label{LP} \begin{aligned} \min& & \sum_e c_e x_e& & &\\ s.t.& & \sum_i D_iy_i&=1 & &\\ & & \sum_{e\in p} x_e&\geq y_i & &\forall p\in \mathcal{P}_{s_i,t_i}, \forall i\\ & & x_e,y_i&>0 \end{aligned} \end{equation} \end{minipage} \bigskip \begin{minipage}{0.47\linewidth} \begin{equation}\label{dual} \begin{aligned} \max& & \lambda& & &\\ s.t.& & \sum_{p\in\mathcal{P}_{s_i,t_i}} y_p&\geq \lambda D_i & &\forall i\\ & & \sum_i \sum_{p\in \mathcal{P}_{s_i,t_i}, p\ni e}y_p&\geq c_e & &\forall e\\ & & y_p&\geq 0 \end{aligned} \end{equation} \end{minipage} \begin{minipage}{0.47\linewidth} \begin{equation}\label{metric} \begin{aligned} \min& & \sum_{uv\in E} c_{uv}d(u,v)& & &\\ s.t.& & \sum_i D_i d(s_i,t_i)&=1 & &\\ & & \text{$d$ is a metric on $V$} \end{aligned} \end{equation} \end{minipage} \newcommand{\lp}{LP\ref{LP}} \newcommand{\ip}{IP\ref{IP}} \newcommand{\dual}{LP\ref{dual}} \newcommand{\metric}{LP\ref{metric}} \begin{enumerate} \item \ip{} $\geq$ \lp{}. Given any feasible solution to \ip{}, we can scale all $x_e$ and $y_i$ simultaneously with factor $1/\sum_i D_i y_i$. The scaled solution is feasible for \lp{} and gets the same objective value. \item \lp{} $=$ \dual{}. by duality. \item \metric{} $=$ \lp{}. It is easy to see \metric{} $\geq$ \lp{} since any feasible metric to \metric{} induces a feasible solution to \lp{}. In fact, the optimal solution to \lp{} also induces a feasible metric. Consider a solution $x_e,y_i$ to \lp{}. Let $d_x$ be the shortest path metric on $V$ using edge length $x_e$. It suffices to show that $y_i=d_x(s_i,t_i)$. This can be seen from a reformulation of \lp{}. The constraint $\sum_i D_i y_i=1$ can be removed and the objective becomes $\sum_e c_e x_e / \sum_i D_i y_i$. This reformulation does not change the optimal solution. Now suppose in the optimal solution to \lp{} there is some $y_i$ which is strictly smaller than $d_x(s_i,t_i)$. Then the denominator $\sum_i D_i y_i$ in the objective of our reformulation can be larger, contradicting to the optimality of solution $x_e,y_i$. \end{enumerate} \bibliographystyle{plainnat} \bibliography{ref} \end{document}