no code implementations • 11 Feb 2024 • Bhisham Dev Verma, Rameshwar Pratap
In this work, we aim to propose faster and space efficient locality sensitive hash functions for Euclidean distance and cosine similarity for tensor data.
no code implementations • 6 Sep 2023 • Amit Deshpande, Rameshwar Pratap
However, in the presence of adversarial noise or outliers, $D^{2}$ sampling is more likely to pick centers from distant outliers instead of inlier clusters, and therefore its approximation guarantees \textit{w. r. t.}
no code implementations • 22 Aug 2023 • Rameshwar Pratap, Raghav Kulkarni
We note that a naive solution to this problem is to repeatedly recompute $\mathrm{minHash}$ with respect to the updated dimension.
no code implementations • 26 Apr 2022 • Amit Deshpande, Rameshwar Pratap
In this paper, we give a one-pass subset selection with an additive approximation guarantee for $\ell_{p}$ subspace approximation, for any $p \in [1, \infty)$.
no code implementations • 4 Mar 2022 • Rameshwar Pratap, Bhisham Dev Verma, Raghav Kulkarni
Tug-of-war} (or AMS) sketch gives a randomized sublinear space (and linear time) algorithm for computing the frequency moments, and the inner product between two frequency vectors corresponding to the data streams.
no code implementations • 1 Dec 2021 • Debajyoti Bera, Rameshwar Pratap, Bhisham Dev Verma
We show that FSketch is significantly faster, and the accuracy obtained by using its sketches are among the top for the standard unsupervised tasks of RMSE, clustering and similarity search.
no code implementations • 13 Nov 2021 • Bhisham Dev Verma, Rameshwar Pratap, Debajyoti Bera
In this work, we present a dimensionality reduction algorithm, aka.
no code implementations • 9 Sep 2021 • Debajyoti Bera, Rameshwar Pratap, Bhisham Dev Verma, Biswadeep Sen, Tanmoy Chakraborty
QUINT is the first of its kind that provides tremendous gain in terms of speed and space usage without compromising much on the accuracy of the downstream tasks.
no code implementations • 20 Mar 2021 • Amit Deshpande, Rameshwar Pratap
Our ideas also extend to give a reduction in the number of passes required by adaptive sampling algorithms for $\ell_{p}$ subspace approximation and subset selection, for $p \geq 2$.
no code implementations • 7 Dec 2020 • Avantika Singh, Chirag Vashist, Pratyush Gaurav, Aditya Nigam, Rameshwar Pratap
Here, in this paper, we propose an iris indexing scheme using real-valued deep iris features binarized to iris bar codes (IBC) compatible with the indexing structure.
no code implementations • 30 Jun 2020 • Amit Deshpande, Rameshwar Pratap
Any multiplicative approximation algorithm for the subspace approximation problem with outliers must solve the robust subspace recovery problem, a special case in which the $(1-\alpha)n$ inliers in the optimal solution are promised to lie exactly on a $k$-dimensional linear subspace.
no code implementations • 10 Oct 2019 • Rameshwar Pratap, Debajyoti Bera, Karthik Revanuru
We compare the performance of our algorithm with the state-of-the-art algorithms on the task of mean-square-error and ranking.
no code implementations • 16 Aug 2017 • Rameshwar Pratap, Ishan Sohony, Raghav Kulkarni
Ideally, the compressed representation of the data should be such, that the similarity between each pair of data points is preserved, while keeping the time and the randomness required for compression as low as possible.