Preyas Popat

Preyas Popat
Are you Preyas Popat?

Claim your profile, edit publications, add additional information:

Contact Details

Name
Preyas Popat
Affiliation
Location

Pubs By Year

Pub Categories

 
Computer Science - Learning (2)
 
Computer Science - Data Structures and Algorithms (2)
 
Computer Science - Computational Complexity (1)
 
Statistics - Machine Learning (1)

Publications Authored By Preyas Popat

Fractals are self-similar recursive structures that have been used in modeling several real world processes. In this work we study how "fractal-like" processes arise in a prediction game where an adversary is generating a sequence of bits and an algorithm is trying to predict them. We will see that under a certain formalization of the predictive payoff for the algorithm it is most optimal for the adversary to produce a fractal-like sequence to minimize the algorithm's ability to predict. Read More

Consider the classical problem of predicting the next bit in a sequence of bits. A standard performance measure is {\em regret} (loss in payoff) with respect to a set of experts. For example if we measure performance with respect to two constant experts one that always predicts 0's and another that always predicts 1's it is well known that one can get regret $O(\sqrt T)$ with respect to the best expert by using, say, the weighted majority algorithm. Read More

We prove that for an arbitrarily small constant $\eps>0,$ assuming NP$\not \subseteq$DTIME$(2^{{\log^{O(1/\eps)} n}})$, the preprocessing versions of the closest vector problem and the nearest codeword problem are hard to approximate within a factor better than $2^{\log ^{1-\eps}n}.$ This improves upon the previous hardness factor of $(\log n)^\delta$ for some $\delta > 0$ due to \cite{AKKV05}. Read More