conf.bib

@comment{{This file has been generated by bib2bib 1.92}}
@comment{{Command line: C:\alldata\data\webbib\bib2bib-1.92.exe -oc conf -ob conf.bib -c '$type = "INPROCEEDINGS"' cormode.bib}}
@inproceedings{CormodeJowhariMonemizadehMuthukrishnan17,
  author = {Graham Cormode and Hossein Jowhari and Morteza Monemizadeh and S. Muthukrishnan},
  title = {Streaming Algorithms for Matching Size Estimation in Sparse Graphs},
  booktitle = {European Symposium on Algorithms},
  year = 2017,
  url = {../papers/matching-esa.pdf},
  slides = {../slides/sparse-esa.pdf},
  pubsite = {http://drops.dagstuhl.de/opus/frontdoor.php?source_opus=7849},
  abstract = {Estimating the size of the maximum matching is a canonical problem in
graph analysis, and one that has attracted extensive study over a
range of different computational models. 
We present improved streaming algorithms for approximating the size of maximum matching with sparse (bounded arboricity) graphs. 

\begin{itemize}
\item {\em (Insert-Only Streams)}
We present a one-pass algorithm that takes  $O(\arbor\log n)$ space and approximates the size of the  maximum matching in 
graphs with arboricity $\arbor$ within a factor of $O(\arbor)$. This improves
significantly upon the state-of-the-art $\tilde{O}(\arbor n^{2/3})$-space
streaming algorithms, and is the first poly-logarithmic space
algorithm for this problem. 

\item {\em (Dynamic Streams)}
Given a dynamic graph stream (i.e., inserts and deletes) of edges of an underlying $\arbor$-bounded arboricity graph,  
we present an one-pass algorithm that uses  space $\tilde{O}(\arbor^{10/3}n^{2/3})$ 
and returns an $O(\arbor)$-estimator for the size of the maximum matching on the condition that the number
edge deletions in the stream is bounded by $O(\arbor n)$. 
For this class of inputs, our algorithm improves the state-of-the-art $\tilde{O}(\arbor n^{4/5})$-space algorithms, where the $\tilde{O}(.)$ notation hides logarithmic in $n$ dependencies.

\end{itemize}

In contrast to prior work, our results take more advantage of the streaming access to the input and characterize the matching size 
based on the ordering of the edges in the stream in addition to the degree
distributions and structural properties of the sparse graphs. 
}
}
@inproceedings{JorgensenYuCormode16,
  author = {Zach Jorgensen and
               Ting Yu and
               Graham Cormode},
  title = {Publishing Attributed Social Graphs with Formal Privacy Guarantees},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  pages = {107--122},
  year = {2016},
  url = {../papers/attributedgraph.pdf},
  slides = {../slides/graphanon-sigmod.pdf},
  poster = {../slides/graphanon16-poster.pdf},
  link = {http://dl.acm.org/citation.cfm?doid=2882903.2915215},
  abstract = {Many data analysis tasks rely on the abstraction of a graph to represent
relations between entities, with attributes on the nodes and edges.  
Since the relationships encoded are often sensitive, we seek effective
ways to release representative graphs which nevertheless protect the
privacy of the data subjects.
Prior work on this topic has focused primarily on the graph
structure in isolation, and has not provided ways to handle richer
graphs with correlated attributes.

We introduce an approach to release such graphs under the strong guarantee
of differential privacy.
We adapt existing graph models, and introduce a new one, and show how
to augment them with meaningful privacy.
This provides a complete workflow, where the input is a sensitive
graph, and the output is a realistic synthetic graph.
Our experimental study demonstrates that our process produces useful,
accurate attributed graphs. }
}
@inproceedings{ChitnisCormodeEsfandariHajiaghayiMcGregorMonemizadehVorotnikova16,
  title = {Kernelization via Sampling with Applications to Dynamic Graph Streams},
  author = {Rajesh Chitnis and Graham Cormode and Hossein Esfandiari and MohammadTaghi Hajiaghayi and Andrew McGregor and Morteza Monemizadeh and Sofya Vorotnikova},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  year = 2016,
  url = {../papers/vcsampling.pdf},
  link = {http://arxiv.org/abs/1505.01731},
  abstract = {In this paper we present a simple but powerful subgraph sampling primitive that is applicable in a variety of computational models including dynamic graph streams (where the input graph is defined by a sequence of edge/hyperedge insertions and deletions) and distributed systems such as MapReduce. In the case of dynamic graph streams, we use this primitive to prove the following results:

* 
{\em Matching:} Our main result for matchings is that there exists an ${O}\sim(k^2)$ space algorithm that returns the edges of a maximum matching on the assumption the cardinality is at most $k$. The best previous algorithm used ${O}\sim(kn)$ space where $n$ is the number of vertices in the graph
and we prove our result is optimal up to logarithmic factors. Our algorithm has ${O}\sim(1)$ update time. 
We also show that there exists an ${O}\sim(n^2/\alpha^3)$ space
algorithm that returns an $\alpha$-approximation for matchings of
arbitrary size. In independent work, Assadi et al.~(arXiv 2015) proved
this is optimal and provided an alternative algorithm. We  generalize our  exact and approximate algorithms to weighted matching. 
While there has been a substantial amount of work on approximate matching in insert-only graph streams, these are the first non-trivial results  in the dynamic setting. 

* {\em Vertex Cover and Hitting Set:} There exists an ${O}\sim(k^d)$ space algorithm that solves the minimum hitting set problem where $d$ is the cardinality of the input sets and $k$ is an upper bound on the size of the minimum hitting set. We prove this is  optimal up to logarithmic factors. Our algorithm has ${O}\sim(1)$ update time. The case $d=2$ corresponds to minimum vertex cover.

Finally, we consider a larger family of parameterized problems (including $b$-matching, disjoint paths, vertex coloring among others) for which our subgraph sampling primitive yields fast, small-space dynamic graph stream algorithms. We then show lower bounds for natural problems outside this family. 
}
}
@inproceedings{AhnCormodeGuhaMcGregorWirth15,
  author = {Kook Jin Ahn and
               Graham Cormode and
               Sudipto Guha and
               Andrew McGregor and
               Anthony Wirth},
  title = {Correlation Clustering in Data Streams},
  booktitle = {International Conference on Machine Learning,
               {ICML}},
  pages = {2237--2246},
  year = {2015},
  pubsite = {http://jmlr.org/proceedings/papers/v37/ahn15.html},
  url = {../papers/corrclustering.pdf},
  abstract = {
  In this paper, we address the
  problem of \emph{correlation clustering} in the dynamic data stream
  model.
  The stream consists of updates to the edge weights of a graph
  on~$n$ nodes and the goal is to find a node-partition such that the
  end-points of negative-weight edges are typically in different
  clusters whereas the end-points of positive-weight edges are typically
  in the same cluster. We present polynomial-time, $O(n\cdot {polylog}
  n)$-space approximation algorithms for natural problems that arise.
  
  We first develop data structures based on linear sketches that allow
  the ``quality'' of a given node-partition to be measured. We then 
  combine these data structures with convex programming and
  sampling techniques to solve the relevant approximation problem.
  However the standard LP and SDP formulations are not obviously
  solvable in $O(n\cdot {polylog} n)$-space. Our work presents
  space-efficient algorithms for the convex programming required, as well as
  approaches to reduce the adaptivity of the sampling. Note that the improved space
  and running-time bounds achieved from streaming algorithms
  are also useful for offline settings such as MapReduce models.}
}
@inproceedings{HeCormodeMachanavajjhalaProcopiucSrivastava15,
  author = {Xi He and
               Graham Cormode and
               Ashwin Machanavajjhala and
               Cecilia M. Procopiuc and
               Divesh Srivastava},
  title = {{DPT:} Differentially Private Trajectory Synthesis Using Hierarchical
               Reference Systems},
  booktitle = {Proceedings of the {VLDB} Endowment},
  volume = {8},
  pages = {1154--1165},
  year = {2015},
  url = {../papers/dpt.pdf},
  pubsite = {http://www.vldb.org/pvldb/vol8/p1154-he.pdf},
  abstract = {GPS-enabled devices are now ubiquitous, from airplanes and cars to 
smartphones and wearable technology.
This has resulted in a wealth of data about the movements of 
individuals and populations, which can be analyzed for useful information 
to aid in city and traffic planning, disaster preparedness and so on. 
However, the places that people go can disclose extremely sensitive 
information about them, and thus their use needs to be filtered 
through privacy preserving mechanisms. 
This turns out to be a highly challenging task: raw trajectories are 
highly detailed, and typically no pair is alike. 
Previous attempts fail either to provide adequate privacy protection, 
or to remain sufficiently faithful to the original behavior.

This paper presents DPT, a system to synthesize mobility data based 
on raw GPS trajectories of individuals while ensuring strong privacy 
protection in the form of $\epsilon$-differential privacy.
DPT makes a number of novel modeling and algorithmic contributions
including (i) discretization of raw trajectories using hierarchical 
reference systems (at multiple resolutions) to capture individual 
movements at differing speeds, (ii) adaptive mechanisms to select a 
small set of reference systems and construct prefix tree counts
privately, and (iii) use of direction-weighted sampling for improved utility. 
While there have been prior attempts to solve the subproblems required
to generate synthetic trajectories, to the best of our knowledge, 
ours is the first system that provides an end-to-end solution. 
We show the efficacy of our synthetic trajectory generation system 
using an extensive empirical evaluation.}
}
@inproceedings{ChitnisCormodeEsfandiariHajiaghayiMonemizadeh15,
  author = {Rajesh Chitnis and
               Graham Cormode and
               Hossein Esfandiari and
               MohammadTaghi Hajiaghayi and
               Morteza Monemizadeh},
  title = {New Streaming Algorithms for Parameterized Maximal
               Matching {\&} Beyond},
  booktitle = {Symposium on Parallelism in Algorithms},
  pages = {56--58},
  year = {2015},
  url = {../papers/maxmatching.pdf},
  abstract = {
   Very recently at SODA'15, we studied maximal matching via the framework of
   \emph{parameterized streaming}, where we sought solutions
   under the promise that no maximal matching exceeds $k$ in size.
   In this paper, we revisit this problem and provide a much simpler algorithm for this problem.
   We are also able to apply the same technique to the \emph{Point Line Cover} problem.}
}
@inproceedings{ZhangCormodeProcopiucSrivastavaXiao15,
  author = {Jun Zhang and Graham Cormode and Magda Procopiuc and Divesh Srivastava and Xiaokui Xiao},
  title = {Private Release of Graph Statistics using Ladder Functions},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  year = 2015,
  url = {../papers/staircase.pdf},
  abstract = {
  Protecting the privacy of individuals in graph structured data while
  making accurate versions of the data available is one of the most
  challenging problems in data privacy.
  Most efforts to date to perform this data
  release end up mired in complexity, overwhelm the signal with noise, and
  are not effective for use in practice.
  In this paper, we introduce a new method which guarantees differential
  privacy.  It specifies a probability distribution over possible outputs
  that is carefully defined to maximize the utility for the given input,
  while still providing the required privacy level.  The distribution is
  designed to form a `ladder', so that each output achieves the highest
  `rung' (maximum probability) compared to less preferable outputs.  We
  show how our ladder framework can be applied to problems of counting the
  number of occurrences of subgraphs, a vital objective in graph analysis,
  and give algorithms whose cost is comparable to that of computing the
  count exactly.  Our experimental study confirms that our method
  outperforms existing methods for counting triangles and stars in
  terms of accuracy, and provides solutions for some problems for which
  no effective method was previously known.
  The results of our algorithms can be used to estimate the parameters of
  suitable graph models, allowing synthetic graphs to be sampled.
}
}
@inproceedings{ChakrabartiCormodeMcGregorThalerVenkatasubramanian15,
  author = {Amit Chakrabarti and Graham Cormode and Andrew McGregor and Justin Thaler and Suresh Venkatasubramanian},
  year = 2015,
  booktitle = {Computational Complexity Conference},
  title = {Verifiable Stream Computation and Arthur-Merlin Communication},
  url = {../papers/oipccc.pdf},
  abstract = {
In the setting of streaming interactive proofs (SIPs), a client (verifier)
needs to compute a given function on a massive stream of data, arriving
online, but is unable to store even a small fraction of the data. It
outsources the processing to a third party service (prover), but is unwilling
to blindly trust answers returned by this service. Thus, the service cannot
simply supply the desired answer; it must {\em convince} the verifier of its
correctness via a short interaction after the stream has been seen.

In this work we study ``barely interactive'' SIPs. Specifically, we show that
two or three rounds of interaction suffice to solve several query
problems---including Index, Median, Nearest Neighbor Search, Pattern Matching,
and Range Counting---with polylogarithmic space and communication costs.  Such
efficiency with $O(1)$ rounds of interaction was thought to be impossible
based on previous work.

On the other hand, we initiate a formal study of the limitations of
constant-round SIPs by introducing a new hierarchy of communication models
called Online Interactive Proofs (OIPs). The online nature of these models is
analogous to the streaming restriction placed upon the verifier in an SIP. We
give upper and lower bounds that (1) characterize, up to quadratic blowups,
every finite level of the OIP hierarchy in terms of other well-known
communication complexity classes, (2) separate the first four levels of the
hierarchy, and (3) reveal that the hierarchy collapses to the fourth level.
Our study of OIPs reveals marked contrasts and some parallels with the classic
Turing Machine theory of interactive proofs, establishes limits on the power
of existing techniques for developing constant-round SIPs, and provides a new
characterization of (non-online) Arthur--Merlin communication in terms of an
online model.
}
}
@inproceedings{CormodeJorgensenYu15,
  author = {Zach Jorgensen and Ting Yu and Graham Cormode},
  title = {Conservative or Liberal? Personalized Differential Privacy},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2015,
  url = {../papers/pdp.pdf},
  abstract = {
  Differential privacy is widely accepted as a powerful framework for 
  providing strong, formal privacy guarantees for aggregate data analysis. 
  A limitation of the model is that the same level of privacy protection 
  is afforded for all individuals. However, it is common that the data 
  subjects have quite different expectations regarding the acceptable 
  level of privacy for their data. Consequently, differential privacy may 
  lead to insufficient privacy protection for some users, while 
  over-protecting others. 
  
  We argue that by accepting that not all users require the same level of 
  privacy, a higher level of utility can often be attained by not 
  providing excess privacy to those who do not want it. We 
  propose a new privacy definition called \emph{personalized differential 
  privacy} (PDP), a generalization of differential privacy in which users 
  specify a personal privacy requirement for their data. We then introduce 
  several novel mechanisms for achieving PDP. Our primary mechanism is a 
  general one that automatically converts any existing 
  differentially private algorithm into one that satisfies PDP. We also 
  present a more direct approach for achieving PDP, 
  inspired by the well-known exponential mechanism. We demonstrate 
  our framework through extensive experiments on real and synthetic data. }
}
@inproceedings{ChitnisCormodeHajiaghayiMonemizadeh15,
  title = {Parameterized Streaming: Maximal Matching and Vertex Cover},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  year = 2015,
  author = {Rajesh Chitnis and Graham Cormode and MohammadTaghi Hajiaghayi and Morteza Monemizadeh},
  url = {../papers/paramstreaming.pdf},
  abstract = {
    As graphs continue to grow in size, we seek ways to effectively process such data at scale.
    The model of streaming graph processing, in which a compact summary is maintained as each
    edge insertion/deletion is observed, is an attractive one.
    However, few results are known for optimization problems over such dynamic graph streams.
    
    In this paper, we introduce a new approach to handling graph streams,
    by instead seeking solutions for the parameterized versions of these problems.
    Here, we are given a parameter $k$ and the objective is to decide
    whether there is a solution bounded by $k$.
    By combining kernelization techniques with randomized sketch structures,
    we obtain the first streaming algorithms for the parameterized
    versions of Maximal Matching and Vertex Cover.
    We consider various models for a graph stream on $n$ nodes:
    the insertion-only model where the edges can only be added,
    and the dynamic model where edges can be both inserted and deleted.
    More formally, we show the following results:
    
    
    $*$
    In the insertion only model, there is a one-pass deterministic algorithm
    for the parameterized Vertex Cover problem which computes a sketch
    using $O\sim(k^{2})$ space [
$O\sim(f(k))=O(f(k)\cdot\log^{O(1)} m)$, where $m$ is the number of edges.]
      such that at each
      timestamp in time $\sim{O}(2^{k})$ it can either extract a solution of
      size at most $k$ for the current instance, or report that no such solution exists.
    We also show a tight lower bound of $\Omega(k^2)$ for the space
    complexity of any (randomized) streaming algorithms for
    the parameterized Vertex Cover, even in the insertion-only model.
    
    $*$ In the dynamic model, and under the \emph{promise} that at each
    timestamp there is a maximal matching of size at most $k$, there is a one-pass
    $O\sim(k^2)$-space (sketch-based) dynamic algorithm that maintains a maximal matching
    with worst-case update time [The time to update the current maximal matching upon an insertion or deletion.]
    $O\sim(k^2)$. This algorithm partially solves Open Problem $64$ from the List of open problems in sublinear algorithms.
    An application of this dynamic matching algorithm is a one-pass
    $O\sim(k^{2})$-space streaming algorithm for the parameterized Vertex Cover problem that in time $O\sim(2^{k})$
    extracts a solution for the final instance with probability $1-\delta/{n^{O(1)}}$, where $\delta<1$.
    To the best of our knowledge, this is the first graph streaming algorithm that combines linear
    sketching with sequential operations that depend on the graph at the current time.
    
    $*$ In the dynamic model without any promise, there is a one-pass
    randomized algorithm for the parameterized Vertex Cover problem
    which computes a sketch using $O\sim(nk)$ space such that in time $O\sim(nk+2^{k})$ it can either extract a solution of size
    at most $k$ for the final instance, or report that no such solution exists.
    
}
}
@inproceedings{CormodeProcopiucSrivastavaXiaoZhang14,
  title = {PrivBayes: Private Data Release via Bayesian Networks},
  author = {Graham Cormode and Magda Procopiuc and Divesh Srivastava and Xiaokui Xiao and Jun Zhang},
  year = 2014,
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  url = {../papers/PrivBayes.pdf},
  slides = {../slides/PrivBayes.pdf},
  abstract = {
    Privacy-preserving data publishing is an important problem that has
    been the focus of extensive study.
    The state-of-the-art goal for this problem is differential privacy,
    which offers a strong degree of privacy protection without making
    restrictive assumptions about the adversary.
    Existing techniques using differential privacy, however, cannot
    effectively handle the publication of high-dimensional data.
    In particular, when the input dataset contains a large number of
    attributes, existing methods require injecting a prohibitive amount of
    noise compared to the signal in the data, which renders the
    published data next to useless.
    
    To address the deficiency of the existing methods, this paper presents
    PrivBayes, a differentially private method for releasing
    high-dimensional data.
    Given a dataset $D$, PrivBayes first constructs a Bayesian network
    $\mathcal{N}$, which (i) provides a succinct model of the correlations
    among the attributes in $D$ and (ii) allows us to approximate the
    distribution of data in $D$ using a set $\mathcal{P}$ of
    low-dimensional marginals of $D$.
    After that, PrivBayes injects noise
    into each marginal in $\mathcal{P}$ to ensure differential privacy,
    and then uses the noisy marginals and the Bayesian network to
    construct an approximation of the data distribution in $D$. Finally,
    PrivBayes samples tuples from the approximate distribution to
    construct a synthetic dataset, and then releases the synthetic
    data. Intuitively, PrivBayes circumvents the curse of dimensionality,
    as it injects noise into the low-dimensional marginals in
    $\mathcal{P}$ instead of the high-dimensional dataset $D$.
    Private construction of Bayesian networks turns out to be
    significantly challenging, and we introduce a novel approach that uses
    a surrogate function for mutual information to build the model more
    accurately. We experimentally evaluate PrivBayes on real data, and demonstrate that it significantly outperforms existing solutions in terms of accuracy.  
}
}
@inproceedings{MaMuthukrishnanThompsonCormode14,
  title = {Modeling Collaboration in Academia: A Game Theoretic Approach},
  author = {Qiang Ma and S. Muthukrishnan and Brian Thompson and Graham Cormode},
  booktitle = {WWW Workshop on Big Scholarly Data},
  year = 2014,
  url = {../papers/hgame.pdf},
  abstract = {
    In this work, we aim to understand the mechanisms driving academic collaboration. We begin by building a model for how researchers split their effort between multiple papers, and how collaboration affects the number of citations a paper receives, supported by observations from a large real-world publication and citation dataset, which we call the {\em h-Reinvestment model}. Using tools from the field of Game Theory, we study
    researchers' collaborative behavior over time under this model, with the premise that each researcher wants to maximize his or her academic success. We find analytically that there is a strong incentive to collaborate rather than work in isolation, and that studying collaborative behavior through a game-theoretic lens is a promising approach to help us better understand the nature and dynamics of academic collaboration.}
}
@inproceedings{CormodeMuthukrishnanYan14,
  author = {Graham Cormode and S. Muthukrishnan and Jinyun Yan},
  title = {People Like Us: Mining Scholarly Data for Comparable Researchers},
  booktitle = {WWW Workshop on Big Scholarly Data},
  year = 2014,
  url = {../papers/comparableppl.pdf},
  abstract = {
    We present the problem of finding comparable researchers for any given researcher. 
    This problem has many motivations.
    Firstly, know thyself.
    The answers of where we stand among research community and who we are
    most alike may not be easily found by existing evaluations of ones' research 
    mainly based on citation counts.
    Secondly, there are many situations where one needs to find comparable researchers e.g., for reviewing peers, constructing programming committees or
    compiling teams for grants. %, etc. 
    It is often done through an ad hoc and informal basis. 
    
    Utilizing the large scale scholarly data accessible on the web, we address the problem of 
    automatically finding comparable researchers.
    We propose a standard to quantify the quality of research output, via the quality of publishing venues. 
    We represent a researcher as a sequence of her publication records, and develop a framework of comparison of researchers by sequence matching. 
    Several variations of comparisons are considered including matching by
    quality of publication venue and research topics, and performing prefix matching.  
    We evaluate our methods on a large corpus and
    demonstrate the effectiveness  of our methods through examples.
    In the end, we identify several promising directions for further work. }
}
@inproceedings{ChakrabartiCormodeGoyalThaler14,
  author = {Amit Chakrabarti and Graham Cormode and Navin Goyal and Justin Thaler},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  year = 2014,
  title = {Annotations for Sparse Data Streams},
  url = {../papers/sparseannotations.pdf},
  link = {http://arxiv.org/abs/1304.3816},
  abstract = {
  Motivated by the surging popularity of commercial cloud computing services, a
  number of recent works have studied \emph{annotated data streams} and variants
  thereof.
  In this setting, a computationally weak \emph{verifier} (cloud user), lacking
  the resources to store and manipulate his massive input locally, accesses a
  powerful but untrusted \emph{prover} (cloud service).  The verifier must work
  within the restrictive data streaming paradigm.  The prover, who can
  \emph{annotate} the data stream as it is read, must not just supply the final
  answer but also convince the verifier of its correctness. Ideally, both the amount
  of annotation from the prover and the space used by the verifier should be
  sublinear in the relevant input size parameters.
  
  A rich theory of such algorithms---which we call \emph{schemes}---has started
  to emerge. Prior work has shown how to leverage the prover's power to
  efficiently solve problems that have no non-trivial standard data stream
  algorithms.  However, even though optimal schemes are now known for several
  basic problems, such optimality holds only for streams whose length is commensurate with the
  size of the \emph{data universe}.  In contrast, many real-world data sets are
  relatively \emph{sparse}, including graphs that contain only $o(n^2)$ edges, and IP
  traffic streams that contain much fewer than the total number of possible IP
  addresses, $2^{128}$ in IPv6.
  
  Here we design the first annotation schemes that allow both the annotation and
  the space usage to be sublinear in the total number of stream {\em updates}
  rather than the size of the data universe.  We solve significant problems,
  including variations of INDEX, \textsc{set-disjointness}, and
  \textsc{frequency-moments}, plus several natural problems on graphs. On the
  other hand, we give a new lower bound that, for the first time, rules out
  smooth tradeoffs between annotation and space usage for a specific problem.
  Our technique brings out new nuances in Merlin--Arthur communication
  complexity models, and provides a separation between online versions of the MA
  and AMA models. }
}
@inproceedings{CormodeMuthukrishnanYan13b,
  author = {Graham Cormode and S. Muthukrishnan and Jinyun Yan},
  title = {First Author Advantage: Citation Labeling in Research},
  booktitle = {Proceedings of the Computational Scientometrics: Theory and Applications Workshop at CIKM},
  year = 2013,
  url = {../papers/firstauthor.pdf},
  link = {http://arxiv.org/abs/1310.1968},
  abstract = {
  Citations among research papers, and the networks they form, are the primary
  object of study in scientometrics. The act of making a citation reflects
  the citerís knowledge of the related literature, and of the work being cited.
  We aim to gain insight into this process by studying citation keys: user-generated
  labels to identify a cited work. Our main observation is that the
  first listed author is disproportionately represented in such labels, implying
a strong mental bias towards the first author.}
}
@inproceedings{CormodeGongProcopiucShenSrivastavaYu13,
  author = {Graham Cormode and Xi Gong and Cecilia M. Procopiuc and Entong Shen and Divesh Srivastava and Ting Yu},
  booktitle = {{ACM} Conference on Information and Knowledge Management ({CIKM})},
  year = 2013,
  title = {{UMicS}: From Anonymized Data to Usable MicroData},
  url = {../papers/umics.pdf},
  abstract = {
  There is currently a tug-of-war going on surrounding data
  releases. On one side, there are many strong reasons pulling to
  release data to other parties: business factors, freedom of
  information rules, and scientific sharing agreements. On the
  other side, concerns about individual privacy pull back, and
  seek to limit releases. Privacy technologies such as
  differential privacy have been proposed to resolve this
  deadlock, and there has been much study of how to perform
  private data release of data in various forms. The focus of
  such works has been largely on the {\em data owner}: what
  process should they apply to ensure that the released data
  preserves privacy whilst still capturing the input data
  distribution accurately. Almost no attention has been paid to
  the needs of the {\em data user}, who wants to make use of the
  released data within their existing suite of tools and data.
  The difficulty of making use of data releases is a major
  stumbling block for the widespread adoption of data privacy
  technologies.
  
  In this paper, instead of proposing new privacy mechanisms for
  data publishing, we consider the whole data release process,
  from the data owner to the data user. We lay out a set of
  principles for privacy tool design that highlights the
  requirements for {\em interoperability}, {\em extensibility}
  and {\em scalability}. We put these into practice with {\sf UMicS},
  an end-to-end prototype system to control the release and use
  of private data. An overarching tenet is that it should be
  possible to integrate the released data into the data user's
  systems with the minimum of change and cost. We describe how to
  instantiate {\sf UMicS} in a variety of usage scenarios. We show how
  using data modeling techniques from machine learning can
  improve the utility, in particular when combined with
  background knowledge that the data user may possess. We
  implement {\sf UMicS}, and evaluate it over a selection of data sets
  and release cases. We see that {\sf UMicS} allows for very effective
use of released data, while upholding our privacy principles.}
}
@inproceedings{CormodeDeligianakkisGarofalakisPapadopoulos13,
  title = {Lightweight Authentication of Linear Algebraic Queries on Data Streams},
  author = {Stavros Papadopoulos and Graham Cormode and Antonis Deligiannakis and Minos Garofalakis},
  year = 2013,
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  url = {../papers/streamauthsigmod.pdf},
  poster = {../slides/streamauthposter.pdf},
  abstract = {
  We consider a \emph{stream outsourcing} setting, where a data owner
  delegates the management of a set of disjoint data streams to an
  untrusted server. The owner \emph{authenticates} his streams via
  signatures. The server processes continuous queries on the union of
  the streams for clients trusted by the owner. Along with the results,
  the server sends proofs of result correctness derived from the owner's
  signatures, which are easily verifiable by the clients. We design
  novel constructions for a collection of fundamental problems over streams
  represented as \emph{linear algebraic} queries. In particular, our basic schemes
  authenticate \emph{dynamic vector sums} and
  \emph{dot products}, 
  as well as \emph{dynamic matrix products}. 
  These techniques can be adapted for authenticating a wide range
  of important operations in streaming environments, including group by
  queries, joins, in-network aggregation, similarity matching, and event processing.
  All our schemes are very \emph{lightweight}, and offer strong \emph{cryptographic}
  guarantees derived from formal definitions and proofs. 
  We experimentally confirm the practicality of our schemes in the
  performance sensitive streaming setting.   }
}
@inproceedings{CormodeYi13,
  title = {Quantiles over Data Streams: An Experimental Study},
  author = {Lu Wang and Ge Luo and Ke Yi and Graham Cormode},
  year = 2013,
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  url = {../papers/nquantiles.pdf},
  abstract = {
    A fundamental problem in data management and analysis is to generate
    descriptions of the distribution of data.  It is most common to give such
    descriptions in terms of the cumulative distribution, which is
    characterized by the quantiles of the data.  The design and engineering
    of efficient methods to find these quantiles has attracted much study,
    especially in the case where the data is described incrementally, and we
    must compute the quantiles in an online, streaming fashion.  Yet while
    such algorithms have proved to be tremendously useful in practice, there
    has been limited formal comparison of the competing methods, and no
    comprehensive study of their performance.  In this paper, we remedy this
    deficit by providing a taxonomy of different methods, and describe
    efficient implementations.  In doing so, we propose and analyze
    variations that have not been explicitly studied before, yet which turn
    out to perform the best.  To illustrate this, we provide detailed
    experimental comparisons demonstrating the tradeoffs between space, time,
  and accuracy for quantile computation.}
}
@inproceedings{CormodeProcopiucShenSrivastavaYu13,
  author = {Graham Cormode and Cecilia M. Procopiuc and Entong Shen and Divesh Srivastava and Ting Yu},
  booktitle = {Privacy-Preserving Data Publication and Analysis (PrivDB)},
  title = {Empirical Privacy and Empirical Utility of Anonymized Data},
  att_authors = {gc2602,ds8961,cp2838},
  att_private = {false},
  year = {2013},
  url = {../papers/empiricalpriv.pdf},
  abstract = {Procedures to anonymize data sets are vital for companies, government
agencies and other bodies to meet their obligations to share data
without compromising the privacy of the individuals contributing to it.
Despite much work on this topic, the area has not yet reached stability.
Early models ($k$-anonymity and $l$-diversity) are now thought to
offer insufficient privacy.
Noise-based methods like differential privacy are seen as providing
stronger privacy, but less utility.
However, across all methods
sensitive information of some individuals can often be
inferred with relatively high accuracy.

In this paper, we reverse the idea of a `privacy attack,' by incorporating it
into a measure of privacy.
Hence, we advocate the notion of {\em empirical privacy}, based on the posterior beliefs of an adversary,
and their ability to draw inferences about sensitive values in the
data.
This is not a new model, but rather a unifying view:
it allows us to study several well-known privacy models
which are not directly comparable otherwise.
We also consider an empirical approach to measuring utility, based on
a workload of queries.
Consequently, we are able to place different privacy models including
differential privacy and early syntactic models on
the same scale, and compare their privacy/utility tradeoff.
We learn that, in practice, the difference between differential
privacy and various syntactic models
is less dramatic than previously thought, but there are still clear
domination relations between them.}
}
@inproceedings{CormodeMirylenkaPalpanasSrivastava13,
  author = {Graham Cormode and Katsiaryna Mirylenka and Themis Palpanas and Divesh Srivastava},
  title = {Finding Interesting Correlations with Conditional Heavy Hitters},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2013,
  att_authors = {gc2602,ds8961},
  att_private = {false},
  url = {../papers/condhh.pdf},
  poster = {../slides/condhhposter.pdf},
  slides = {../slides/condhhicde.pdf},
  abstract = {The notion of {\em heavy hitters}---items that make up a large
fraction of the population---has been successfully used in a variety
of applications across sensor and RFID monitoring, network data
analysis, event mining, and more.  Yet this notion often fails to
capture the semantics we desire when we observe data in the form of
correlated pairs. Here, we are interested in items that are {\em
  conditionally} frequent: when a particular item is frequent within
the context of its parent item. In this work, we introduce and
formalize the notion of Conditional Heavy Hitters to identify such
items, with applications in network monitoring, and Markov chain modeling.
We introduce several streaming algorithms that allow us to find
conditional heavy hitters efficiently, and provide analytical results.
Different algorithms are successful for different input
characteristics.  We perform an experimental evaluation to demonstrate
the efficacy of our methods, and to study which algorithms are most
suited for different types of data. }
}
@inproceedings{CormodeProcopiucSrivastavaYaroslavtsev13,
  author = {Graham Cormode and Cecilia M. Procopiuc and Divesh Srivastava and Grigory Yaroslavtsev},
  title = {Accurate and Efficient Private Release of Datacubes and Contingency Tables},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2013,
  att_authors = {gc2602,ds8961,cp2838},
  att_private = {false},
  url = {../papers/privcube.pdf},
  abstract = {A central problem in releasing aggregate information about
sensitive data is to do so accurately
while providing a privacy guarantee on the output.
Recent work focuses on the class of {\em linear queries}, which include basic
counting queries, data cubes, and contingency tables. The goal is to
maximize the utility of their output, while giving a rigorous privacy
guarantee.
Most results follow a common template: pick a ``strategy'' set of linear
queries to apply to the data, then use the noisy answers to these
queries to reconstruct the queries of interest.
This entails either picking a strategy set that is hoped to be good
for the queries, or performing a costly search over the space
of all possible strategies.

In this paper, we propose a new approach that balances accuracy and
efficiency:
  we show how to improve the accuracy of a given query set by
  answering some strategy queries more accurately than others.
This leads to an efficient optimal noise allocation for many popular
strategies,
including wavelets, hierarchies, Fourier coefficients and more.
For the important case of marginal queries we show that this strictly improves on previous
methods, both analytically and empirically.
Our results also extend to ensuring that the returned query answers
are consistent with an (unknown) data set at minimal extra cost in
terms of time and noise.}
}
@inproceedings{CormodeFirmani13,
  author = {Graham Cormode and Donatella Firmani},
  title = {On Unifying the Space of $l_0$-Sampling Algorithms},
  booktitle = {{SIAM} Meeting on Algorithm Engineering and Experiments},
  year = 2013,
  att_authors = {gc2602},
  att_private = {false},
  url = {../papers/l0samp.pdf},
  abstract = {
    The problem of building an $l_0$-sampler is to sample
    near-uniformly from the support set of a dynamic multiset.  
    This problem has a variety of applications within data analysis,
    computational geometry and graph algorithms. 
    In this paper, we abstract a set of steps for building an $l_0$-sampler, based on sampling, recovery and selection. 
    We analyze the implementation of an $l_0$-sampler within this
    framework, and show how prior constructions of 
    $l_0$-samplers can all be expressed in terms of these steps. 
    Our experimental contribution is to provide a first detailed study of the
    accuracy and computational cost of $l_0$-samplers.}
}
@inproceedings{GoyalDaumeCormode12,
  author = {Amit Goyal and
               Hal {Daum{\'e} III} and
               Graham Cormode},
  title = {Sketch Algorithms for Estimating Point Queries in {NLP}},
  booktitle = {EMNLP-CoNLL},
  att_authors = {gc2602},
  att_private = {false},
  url = {../papers/nlpsketch.pdf},
  year = {2012},
  pages = {1093-1103},
  abstract = {
Many NLP tasks rely on accurate statistics
from large corpora. Tracking complete
statistics is memory intensive, so recent
work has proposed using compact approximate
``sketches'' of frequency distributions.
We describe 10 sketch methods, including existing
and novel variants. We compare and
study the errors (over-estimation and underestimation)
made by the sketches. We evaluate
several sketches on three important NLP problems.
Our experiments show that {\em one} sketch
performs best for all the three tasks.}
}
@inproceedings{AgarwalCormodeHuangPhillipsWeiYi12,
  title = {Mergeable Summaries},
  author = {Pankaj Agarwal and Graham Cormode and Zengfeng Huang and Jeff Phillips and Zheiwei Wei and Ke Yi},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  att_authors = {gc2602},
  att_private = {false},
  year = 2012,
  url = {../papers/mergeable.pdf},
  abstract = {
We study the {\em mergeability} of data summaries.  Informally speaking,
mergeability requires that, given two summaries on two data sets, there is
a way to merge the two summaries into a single summary on the union of the two data sets, while preserving the error and size guarantees.  This property means that the summaries can be merged in a way like other algebraic
operators such as sum and max, 
which is especially useful for computing summaries on massive distributed
data.  Several data summaries are trivially mergeable by construction, most
notably all the {\em sketches} that are linear functions of the data sets.
But some other fundamental
ones like those for heavy hitters and quantiles, are not (known to be)
mergeable.  In this paper, we demonstrate that these summaries are indeed
mergeable or can be made mergeable after appropriate modifications.
Specifically, we show that for $\epsilon$-approximate heavy hitters, there is a
deterministic mergeable summary of size $O(1/\epsilon)$; for $\epsilon$-approximate
quantiles, there is a deterministic summary of size $O(\frac{1}{\epsilon}
\log(\epsilon n))$ that has a restricted form of mergeability, and a randomized
one of size $O(\frac{1}{\epsilon} \log^{3/2}\frac{1}{\epsilon})$ with full
mergeability.  We also extend our results to geometric summaries such as
$\epsilon$-approximations and $\epsilon$-kernels.

We also achieve two results of independent interest: 
(1) we provide the best known randomized streaming bound for $\epsilon$-approximate quantiles that depends only on $\epsilon$, of size $O(\frac{1}{\epsilon} \log^{3/2}\frac{1}{\epsilon})$, and 
(2) we demonstrate that the MG and the SpaceSaving summaries for heavy hitters are isomorphic.}
}
@inproceedings{CohenCormodeDuffield12,
  author = {Edith Cohen and Graham Cormode and Nick Duffield},
  year = 2012,
  att_authors = {gc2602,nd1321},
  att_private = {false},
  booktitle = {{ACM} Conference on Measurement and Modeling of Computer Systems ({SIGMETRICS})},
  url = {../papers/sampdel.pdf},
  title = {Don't Let The Negatives Bring You Down: Sampling from Streams of Signed Updates},
  abstract = {
Random sampling has been proven time and time again to be a powerful tool 
for working with large data. Queries over the full dataset are replaced by 
approximate queries over the smaller (and hence easier to store and 
manipulate) sample. The sample constitutes a flexible summary that 
supports a wide class of queries. But in many applications, datasets are 
modified with time, and it is desirable to update samples without 
requiring access to the full underlying datasets. In this paper, we 
introduce and analyze novel techniques for sampling over dynamic data, 
modeled as a stream of modifications to weights associated with each key.

While sampling schemes designed for stream applications can often readily 
accommodate positive updates to the dataset, much less is known for the 
case of negative updates, where weights are reduced or items deleted 
altogether. We primarily consider the turnstile model of streams, and 
extend classic schemes to incorporate negative updates. Perhaps 
surprisingly, the modifications to handle negative updates turn out to be 
natural and seamless extensions of the well-known positive 
update-only algorithms. We show that they produce unbiased estimators, and 
we relate their performance to the behavior of corresponding algorithms on 
insert-only streams with different parameters. A careful analysis is 
necessitated, in order to account for the fact that sampling choices for 
one key now depend on the choices made for other keys.

In practice, our solutions turn out to be efficient and accurate.
Compared to recent algorithms for $L_p$ sampling which can be applied to this 
problem, they are significantly more reliable, and dramatically faster.   }
}
@inproceedings{CormodeYi12,
  author = {Graham Cormode and Ke Yi},
  year = 2012,
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {Scientific and Statistical Database Management ({SSDBM})},
  slides = {../slides/cdsliwinssdbm.pdf},
  url = {../papers/cdsliwinssdbm.pdf},
  title = {Tracking Distributed Aggregates over Time-based Sliding Windows},
  abstract = {
The area of distributed monitoring requires tracking the value of a
function of distributed data as new observations are made.
An important case is when attention is restricted to only a recent
time period, such as the last hour of readings---the sliding window
case. 
In this paper, we introduce a novel paradigm for handling such
monitoring problems, which we dub the ``forward/backward'' approach. 
This view allows us to provide 
optimal or near-optimal solutions for several fundamental problems, such
as counting, tracking frequent items, and maintaining order
statistics. 
The resulting protocols  improve on
previous work or give the first solutions for some problems, 
and operate efficiently in terms of space and time needed.  
Specifically, we obtain optimal $O(\frac{k}{\epsilon} \log (\epsilon n/k))$
communication per window of $n$ updates for tracking counts and heavy
hitters with accuracy $\epsilon$ across $k$ sites; and near-optimal
communication of $O(\frac{k}{\epsilon} \log^2(1/\epsilon) \log (n/k))$
for quantiles.  
We also present solutions for problems such as tracking distinct
items, entropy, and convex hull and diameter of point sets.  
}
}
@inproceedings{CormodeMuthukrishnanYan12,
  author = {Graham Cormode and S. Muthukrishnan and Jinyun Yan},
  year = 2012,
  att_authors = {gc2602},
  att_private = {false},
  url = {../papers/fun12.pdf},
  link = {http://arxiv.org/abs/1202.2638},
  booktitle = {Proceedings of the International Conference on Fun with Algorithms ({FUN})},
  title = {Scienceography: the study of how science is written},
  abstract = {
Scientific literature has itself been the subject of much scientific
study, for a variety of reasons: understanding how results are
communicated, how ideas spread, and assessing the influence of areas
or individuals.  
However, most prior work has focused on extracting and analyzing
citation and stylistic patterns. 
In this work, we introduce the notion of `scienceography', which
focuses on the {\em writing} of science. 
We provide a first large scale study using data derived from
the arXiv e-print repository. 
Crucially, our data includes the ``source code'' of scientific papers---the \LaTeX source---which enables us to study features not present in the ``final product'',
such as the tools used and private comments between authors. 
Our study identifies broad patterns and trends in two example areas---computer science and mathematics---as well as highlighting key
differences in the way that science is written in these fields. 
Finally, we outline future directions to extend the new
topic of scienceography.   
}
}
@inproceedings{CormodeProcopiucSrivastavaTran12,
  author = {Graham Cormode and Magda Procopiuc and Divesh Srivastava and Thanh Tran},
  att_authors = {gc2602,ds8961,cp2838},
  att_private = {false},
  booktitle = {International Conference on Database Theory (ICDT)},
  url = {../papers/sparsedp.pdf},
  year = 2012,
  link = {http://arxiv.org/abs/1103.0825},
  title = {Differentially Private Publication of Sparse Data},
  abstract = {The problem of privately releasing data is to provide a version of a 
dataset without revealing sensitive information about the individuals 
who contribute to the data.  The model of differential privacy allows 
such private release while providing strong guarantees on the output. 
A basic mechanism achieves differential privacy by adding noise to the 
frequency counts in the contingency tables (or, a subset of the count 
data cube) derived from the dataset.  However, when the dataset is 
sparse in its underlying space, as is the case for most multi-attribute 
relations, then the effect of adding noise is to vastly increase the 
size of the published data: it implicitly creates a huge number of 
dummy data points to mask the true data, making it almost impossible
to work with. 

We present techniques to overcome this roadblock and allow efficient 
private release of sparse data, while maintaining the guarantees of 
differential privacy.  Our approach is to release a compact summary of 
the noisy data.  Generating the noisy data and then summarizing it 
would still be very costly, so we show how to shortcut this step, and 
instead directly generate the summary from the input data, without 
materializing the vast intermediate noisy data.  We instantiate this 
outline for a variety of sampling and filtering methods, and show how 
to use the resulting summary for approximate, private, query answering. 
Our experimental study shows that this is an effective, practical
solution: in some examples we generate output that is 1000 times
smaller than the naive method, in less than 1\% of the time while
achieving comparable and occasionally improved utility over the 
costly materialization approach.
}
}
@inproceedings{CormodeProcopiucShenSrivastavaYu12,
  author = {Graham Cormode and Magda Procopiuc and Entong Shen and Divesh Srivastava and Ting Yu},
  title = {Differentially Private Spatial Decompositions},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  att_authors = {gc2602,ds8961,cp2838},
  att_private = {false},
  year = 2012,
  link = {http://arxiv.org/abs/1103.5170},
  url = {../papers/spatialpriv.pdf},
  abstract = {Differential privacy has recently emerged as the de facto standard for private
data release.
This makes it possible to provide
strong theoretical guarantees on the privacy and utility of released
data.
While it is well-understood how to release data based on counts and simple
functions under this guarantee, it remains to provide general purpose
techniques that are useful for a wider variety of queries.
In this paper, we focus on spatial data, i.e., any multi-dimensional
data that can be indexed by a tree structure.
Directly applying existing differential privacy methods to this type
of data simply generates noise.

We propose instead the class of ``private spatial decompositions'':
these adapt standard spatial indexing methods such
as quadtrees and kd-trees to provide a private description of the
data distribution.
Equipping such structures with differential privacy requires several
steps to ensure that they provide meaningful privacy guarantees.
Various basic steps, such as choosing splitting points and describing
the distribution of points within a region, must be done privately,
and the guarantees of the different building blocks must be composed into
an overall guarantee.
Consequently, we expose the design space for private spatial
decompositions, and analyze some key examples.
A major contribution of our work is to
provide new techniques for parameter setting and
post-processing of the output to improve
the accuracy of query answers.
Our experimental study demonstrates that it is possible to build such
decompositions efficiently, and use them to answer a variety of
queries privately and with high accuracy.}
}
@inproceedings{LuBangaloreCormodeHadjieleftheriouSrivastava12,
  author = {Meiyu Lu and Srinivas Bangalore and Graham Cormode and 
Marios Hadjieleftheriou and Divesh Srivastava},
  title = {A Dataset Search Engine for the Research Document Corpus},
  att_authors = {gc2602,ds8961,sb7658,mh6516},
  att_private = {false},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  url = {../papers/dataset.pdf},
  year = 2012,
  abstract = {
A key step in validating a proposed idea or system
is to evaluate over a suitable dataset. However, to this date there
have been no useful tools for researchers to understand which
datasets have been used for what purpose, or in what prior
work. Instead, they have to manually browse through papers
to find the suitable datasets and their corresponding URLs,
which is laborious and inefficient. To better aid the dataset
discovery process, and provide a better understanding of how
and where datasets have been used, we propose a framework to
effectively identify datasets within the scientific corpus. The key
technical challenges are identification of datasets, and discovery
of the association between a dataset and the URLs where they
can be accessed. Based on this, we have built a user friendly
web-based search interface for users to conveniently explore the
dataset-paper relationships, and find relevant datasets and their
properties.}
}
@inproceedings{CormodeShenSrivastavaYu12,
  author = {Graham Cormode and Entong Shen and Divesh Srivastava and Ting Yu},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  att_authors = {gc2602},
  att_private = {false},
  url = {../papers/licm.pdf},
  year = 2012,
  title = {Aggregate Query Answering on Possibilistic Data with Cardinality Constraints},
  abstract = {Uncertainties in data arise for a number of reasons: when the data
set is
incomplete, contains conflicting information or has been deliberately
perturbed or coarsened to remove sensitive details. An important case
which arises in many real applications
is when the data describes a set of {\em possibilities}, but with
{\em cardinality constraints}. These
constraints represent correlations between tuples encoding, e.g. that
at most two possible records are correct, or that there is an
(unknown) one-to-one mapping between a set of tuples and attribute
values. Although there has been much effort to handle
uncertain data, current systems are not equipped to handle such
correlations, beyond simple mutual exclusion and co-existence
constraints. Vitally, they have little support for efficiently handling
{\em aggregate} queries on such data.

In this paper, we aim to address some of these deficiencies, by
introducing LICM (Linear Integer Constraint Model), which can
succinctly represent many types of tuple correlations, particularly a
class of {\em cardinality constraints}. We motivate and explain the
model with examples from data cleaning and masking sensitive data,
to show that it enables
modeling and querying such data, which was not previously possible. We
develop an efficient strategy to answer conjunctive and aggregate queries
on possibilistic data by describing how to implement relational
operators over data in the model. LICM compactly integrates the
encoding of correlations, query answering and lineage recording. In
combination with off-the-shelf linear integer programming solvers, our
approach provides exact bounds for aggregate queries. Our prototype
implementation demonstrates that query answering with LICM can be
effective and scalable.}
}
@inproceedings{CormodeMitzenmacherThaler12,
  author = {Graham Cormode and Michael Mitzenmacher and Justin Thaler},
  title = {Practical Verified Computation with Streaming Interactive Proofs},
  booktitle = {Innovations in Theoretical Computer Science (ITCS)},
  year = 2012,
  att_authors = {gc2602},
  att_private = {false},
  url = {../papers/sipitcs.pdf},
  link = {http://arxiv.org/abs/1105.2003},
  abstract = {
When delegating computation to a service provider, as in the cloud computing
paradigm, we
seek some reassurance that the output is correct and complete. 
Yet recomputing the output as a check is inefficient and expensive, and
it may not even be feasible to store all the data locally. 
We are therefore interested in
what can be validated by a streaming (sublinear space) user, who
cannot store the full input, or perform the full computation herself.
Our aim in this work is to advance a recent line of work on ``proof
systems'' in which the service provider {\em proves} the correctness of
its output to a user.
The goal is to minimize the time and space costs of both parties in
generating and checking the proof. 
Only very recently have there been attempts to implement such proof
systems, and thus far these have been quite limited in functionality.  

Here, our approach is two-fold. 
First, we describe a carefully crafted instantiation of one of the
most efficient 
general-purpose constructions for arbitrary computations 
(streaming or otherwise), due to Goldwasser,
Kalai, and Rothblum.
This requires several new insights and enhancements to 
move the methodology from a theoretical result to a practical
possibility. Our main contribution is in achieving a prover that 
runs in time $O(S(n) \log S(n))$, where $S(n)$ is the size of an arithmetic circuit computing the function of interest;
this compares favorably to the poly$(S(n))$ runtime for the prover promised in prior work. Our experimental results
demonstrate that a practical general-purpose protocol for verifiable computation may be significantly closer to reality than previously
realized.

Second, we describe a set of techniques that achieve
genuine scalability for protocols
fine-tuned for specific important problems in streaming and database
processing. Focusing in particular on \emph{non-interactive} protocols for problems ranging from matrix-vector multiplication
to bipartite perfect matching,
we build on prior work \cite{annotations, graphstream} to achieve a
prover that runs in nearly linear-time, 
while obtaining optimal tradeoffs between communication cost and the user's working memory. 
Existing techniques required (substantially)
superlinear time for the prover. Finally, we develop improved \emph{interactive} protocols for specific problems based on a linearization 
technique originally due to Shen. 
We argue that
even if general-purpose methods improve,
special purpose protocols will remain valuable in real-world settings for
key problems, and hence special attention to specific problems is
warranted.
}
}
@inproceedings{CormodeThalerYi12,
  title = {Verifying Computations with Streaming Interactive Proofs},
  author = {Graham Cormode and Justin Thaler and Ke Yi},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  month = sep,
  year = 2012,
  url = {../papers/ipstreamvldb.pdf},
  link = {http://arxiv.org/abs/1109.6882},
  abstract = {
  When computation is outsourced, the data owner would like to be
  reassured that the desired computation has been performed correctly by
  the service provider.
  In theory, methods based on proof systems can give the data
    owner the necessary assurance, but previous work does not give a
    sufficiently scalable and practical solution, requiring a lot of time,
    space or computational power for both parties.  
  In this paper, we develop
    new proof protocols for verifying computations which are streaming in
    nature: the verifier (data owner) needs only logarithmic space and a single pass over the input, and
    after observing the input follows a simple
    protocol with a prover (service provider) that takes logarithmic
    communication spread over a logarithmic number
    of rounds.  
  These ensure that the computation is performed correctly: that the
  service provider has not made any errors or missed out some data. 
  The guarantee is very strong: even if the service provider
  deliberately tries to cheat, there is only vanishingly small
  probability of doing so undetected, while a correct computation is
  always accepted. 
    
  We first observe that some existing theoretical results 
  can be modified to work with streaming verifiers, 
  which shows that there exist efficient protocols for
  problems in the complexity classes $\mathsf{NP}$ and $\mathsf{NC}$. 
  Our main results 
  seek to bridge the gap between theory and practice by developing
    usable protocols
    for a variety of problems of central importance in 
    streaming and database processing. 
  All of our protocols achieve statistical 
    soundness and most require only logarithmic communication between
    prover and verifier.
  We also experimentally demonstrate their practicality and scalability. All
    these problems require linear space in the traditional streaming model,
    showing that adding a prover can exponentially reduce the effort needed
    by the verifier.}
}
@inproceedings{AgarwalCormodeHuangPhillipsWeiYi11,
  title = {Mergeable Coresets},
  author = {Pankaj Agarwal and Graham Cormode and Zengfeng Huang and Jeff Phillips and Zheiwei Wei and Ke Yi},
  booktitle = {Third Workshop on Massive Data Algorithmics (MASSIVE)},
  att_authors = {gc2602},
  att_private = {false},
  year = 2011,
  url = {../papers/massive.pdf},
  abstract = {
  We study the {\em mergeability} of data summaries.  Informally speaking,
  mergeability requires that, given two summaries on two data sets, there is
  a way to merge the two summaries into a summary on the two data sets
  combined together, while preserving the error and size guarantees.  This
  property means that the summary can be treated like other algebraic
  objects such as sum and max, 
  % essentially allows the summary to be used as if they were some
  %simple decomposable aggregates like sum and max, 
  which is especially useful for computing summaries on massive distributed
  data.  Many data summaries are trivially mergeable by construction, most
  notably those based on linear transformations.  But some other fundamental
  ones like those for heavy hitters and quantiles, are not (known to be)
  mergeable.  In this paper, we demonstrate that these summaries are indeed
  mergeable or can be made mergeable after appropriate modifications.
  Specifically, we show that for $\epsilon$-approximate heavy hitters, there is a
  deterministic mergeable summary of size $O(1/\epsilon)$; for $\epsilon$-approximate
  quantiles, there is a deterministic summary of size $O(\frac{1}{\epsilon}
  \log(\epsilon n))$ that has a restricted form of mergeability, and a randomized
  one of size $O(\frac{1}{\epsilon} \log^{3/2}\frac{1}{\epsilon})$ with full
  mergeability.  We also extend our results to geometric summaries such as
  $\epsilon$-approximations and $\epsilon$-kernels.
}
}
@inproceedings{Cormode11,
  author = {Graham Cormode},
  booktitle = {{ACM} {SIGKDD} Conference},
  att_authors = {gc2602},
  att_private = {false},
  year = 2011,
  url = {../papers/attackdd.pdf},
  poster = {../slides/attackddposter.pdf},
  link = {http://arxiv.org/abs/1011.2511},
  title = {Personal Privacy vs Population Privacy: Learning to Attack Anonymization},
  abstract = {
Over the last decade
 great strides have been made in developing techniques to
compute functions privately. 
In particular, Differential Privacy gives strong promises about
conclusions that can be drawn about an individual. 
In contrast, various syntactic methods for providing privacy (criteria
such as $k$-anonymity and $l$-diversity) have been criticized for
still allowing private information of an individual to be inferred.
 
In this paper, we consider the ability of an attacker to use data
meeting privacy definitions to build an accurate classifier. 
We demonstrate that even under Differential Privacy, such classifiers can be
used to infer ``private'' attributes accurately in realistic data. 
We compare this to similar approaches for inference-based attacks on
other forms of anonymized data.
We show how the efficacy of all these attacks can be measured on the
same scale, based on the probability of successfully inferring a
private attribute. 
We observe that the
accuracy of inference of private attributes for differentially private
data and $l$-diverse data can be quite similar. }
}
@inproceedings{CohenCormodeDuffield11b,
  author = {Edith Cohen and Graham Cormode and Nick Duffield},
  att_authors = {gc2602,nd1321,ec1458},
  att_private = {false},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  year = 2011,
  slides = {../slides/structure-vldb.pdf},
  url = {../papers/rangeaware.pdf},
  title = {Structure-Aware Sampling: Flexible and Accurate Summarization},
  abstract = {
  In processing large quantities of data, a fundamental problem is to
  obtain a summary which supports approximate query answering.  Random
  sampling yields flexible summaries which naturally support subset-sum
  queries with unbiased estimators and well-understood confidence
  bounds.
  Classic sample-based summaries, however, are designed for arbitrary
  subset queries and are oblivious to the structure in the set of keys.
  The particular structure, such as hierarchy, order, or product space
  (multi-dimensional), makes {\em range queries} much more relevant for
  most analysis of the data.
  
  Dedicated summarization algorithms for range-sum queries have also been
  extensively studied.  They can outperform existing sampling schemes in
  terms of accuracy on range queries per summary size. Their accuracy,
  however, rapidly degrades when, as is often the case, the query spans
  multiple ranges.  
  They are also less flexible---being targeted for range sum
  queries alone---and are often quite costly to build and use.
  
  In this paper we propose and evaluate variance optimal sampling
  schemes that are {\em structure-aware}.  
  These summaries improve over 
  the accuracy of existing {\em structure-oblivious} sampling schemes on
  range queries while retaining the benefits of sample-based summaries:
  flexible summaries, with high accuracy on 
both range queries and arbitrary subset queries.
  }
}
@inproceedings{CormodeYi11,
  author = {Graham Cormode and Ke Yi},
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {{ACM} Principles of Distributed Computing ({PODC})},
  year = 2011,
  link = {http://portal.acm.org/citation.cfm?doid=1993806.1993839},
  url = {../papers/cdhhba.pdf},
  slides = {../slides/cdslidingwin-ba.pdf},
  title = {Tracking Distributed Aggregates over Time-based Sliding Windows (Brief Announcement)},
  abstract = {The area of distributed monitoring requires tracking the value of a
function of distributed data as new observations are made.
An important case is when attention is restricted to only a recent
time period, such as the last hour of readings---the sliding window
case. 
In this announcement, we outline a novel paradigm for handling such
monitoring problems, which we dub the ``forward/backward'' approach. 
This provides clean solutions for several fundamental problems, such
as counting, tracking frequent items, and maintaining order
statistics. 
We obtain efficient protocols for these problems that improve on
previous work, and are easy to implement.  
Specifically, we obtain optimal $O(\frac{k}{\epsilon} \log (\epsilon n/k))$
communication per window of $n$ updates for tracking counts and heavy
hitters with accuracy $\epsilon$ across $k$ sites; and near-optimal
communication of $O(\frac{k}{\epsilon} \log^2(1/\epsilon)$ $\log (n/k))$
for quantiles.  }
}
@inproceedings{CohenCormodeDuffield11,
  author = {Edith Cohen and Graham Cormode and Nick Duffield},
  booktitle = {{ACM} Conference on Measurement and Modeling of Computer Systems ({SIGMETRICS})},
  year = 2011,
  link = {http://portal.acm.org/citation.cfm?doid=1993744.1993763},
  url = {../papers/structure-sigmetrics.pdf},
  slides = {../slides/structure-sigmetrics.pdf},
  title = {Structure-Aware Sampling on Data Streams},
  abstract = {
  The massive data streams observed in network monitoring, data processing and
  scientific studies are typically too large to store.  
  For many applications over such data, we must obtain compact summaries
  of the stream. 
  These summaries should allow accurate answering of post hoc queries
  with estimates which approximate the true answers over the original stream.
  The data often has an underlying structure which makes certain subset
  queries, in particular {\em range queries}, more relevant than arbitrary
  subsets.  Applications such as access control, change
  detection, and heavy hitters typically involve subsets that are ranges
  or unions thereof.
  
  Random sampling is a natural summarization tool, 
  being easy to implement and flexible to use. 
  Known sampling methods are 
  good for arbitrary queries but fail to optimize for the common
  case of range queries.  
  Meanwhile, specialized summarization algorithms have been proposed 
  for range-sum queries and related problems. 
  These can outperform sampling giving fixed space resources, but lack
  its flexibility and simplicity. 
  Particularly, their accuracy degrades when queries span multiple
  ranges. 
  
  We define new stream sampling algorithms
  with a smooth and tunable trade-off between
  accuracy on range-sum queries and arbitrary subset-sum queries.  
  The technical key is to relax requirements on the variance over all
  subsets to enable better performance on the ranges of interest.
  This boosts the accuracy on range queries while retaining the prime
  benefits of sampling, in particular flexibility and accuracy, with tail
  bounds guarantees.
  Our experimental study indicates that structure-aware summaries
  drastically improve range-sum accuracy with respect to
  state-of-the-art stream sampling algorithms and outperform
  deterministic methods on range-sum queries and hierarchical heavy
  hitter queries.}
}
@inproceedings{CormodeKarloffWirth10,
  author = {Graham Cormode and Howard Karloff and Tony Wirth},
  att_authors = {gc2602,hk1971},
  att_private = {false},
  title = {Set Cover Algorithms For Very Large Datasets},
  booktitle = {{ACM} Conference on Information and Knowledge Management ({CIKM})},
  year = 2010,
  url = {../papers/ckw.pdf},
  slides = {../slides/setcover.pdf},
  abstract = {
  The problem of {\sc Set Cover}---to find the smallest subcollection of
  sets that covers some universe---is at the heart of many data and
  analysis tasks.  It arises in a wide range of settings, including
  operations research, machine learning, planning, data quality and data
  mining. Although finding an optimal solution is NP-hard, the greedy
  algorithm is widely used, and typically finds solutions that are close
  to optimal. 
  
  However, a direct implementation of the greedy approach, which
  picks the set with the largest number of uncovered items at each step,
  does not behave well when the input is very large and disk resident. The greedy
  algorithm must make many random accesses to disk, which are
  unpredictable and costly in comparison to linear scans. In order to
  scale {\sc Set Cover} to large datasets, we provide a new
  algorithm which finds a solution that is provably close to that of
  greedy, but which is much more efficient to implement using modern
  disk technology. Our experiments show a ten-fold improvement in
  speed on moderately-sized datasets, and an even greater improvement on
larger datasets. }
}
@inproceedings{CormodeMuthukrishnanYiZhang10,
  author = {Graham Cormode and S. Muthukrishnan and Ke Yi and Qin Zhang},
  att_authors = {gc2602},
  att_private = {false},
  title = {Optimal sampling from distributed streams},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  year = 2010,
  url = {../papers/cdsample.pdf},
  slides = {../slides/cdsample.pdf},
  pubsite = {http://portal.acm.org/citation.cfm?doid=1807085.1807099},
  abstract = {  A fundamental problem in data management is to draw a sample of a large
  data set, for approximate query answering, selectivity estimation, and
  query planning.  With large, streaming data sets, this problem becomes
  particularly difficult when the data is shared across multiple
  distributed sites.  The challenge is to ensure that a sample is drawn
  uniformly across the union of the data while minimizing the communication
  needed to run the protocol and track parameters of the evolving data.  At
  the same time, it is also necessary to make the protocol lightweight, by
  keeping the space and time costs low for each participant.
  In this paper, we present communication-efficient protocols for sampling
  (both with and without replacement) from $k$ distributed streams.  These
  apply to the case when we want a sample from the full streams, and to the
  sliding window cases of only the $W$ most recent items, or arrivals
  within the last $w$ time units.  We show that our protocols are optimal,
  not just in terms of the communication used, but also that they use
  minimal or near minimal (up to logarithmic factors) time to process each
  new item, and space to operate.}
}
@inproceedings{BhagatCormodeKrishnamurthySrivastava10A,
  author = {Smriti Bhagat and Graham Cormode and Balachander Krishnamurthy and Divesh Srivastava},
  att_authors = {gc260,ds8961,bk1836},
  att_private = {false},
  title = {Privacy in dynamic social networks},
  booktitle = {World Wide Web Conference ({WWW})},
  year = 2010,
  pubsite = {http://portal.acm.org/citation.cfm?doid=1772690.1772803},
  url = {../papers/danonwww.pdf},
  abstract = {
  Anonymization of social networks before they are published or shared
  has become an important research question. Recent work on anonymizing
  social networks has looked at privacy preserving techniques for
  publishing a single instance of the network. However, social networks
  evolve and a single instance is inadequate for analyzing the evolution
  of the social network or for performing any longitudinal data
  analysis. We study the problem of repeatedly publishing social network
  data as the network evolves, while preserving privacy of users. 
  Publishing multiple instances of the same network independently
  has privacy risks, since stitching the information
  together may allow an adversary to identify users in the networks.
  
  We propose methods to anonymize a dynamic network such that the privacy of
  users is preserved when new nodes and edges are added to the published
  network. 
  These methods make 
  use of link prediction algorithms to model the evolution of the
  social network. Using this predicted graph to perform group-based
  anonymization, the loss in privacy caused by new edges can be reduced.
  We evaluate the privacy loss on publishing multiple social network instances using our methods.}
}
@inproceedings{CormodeMitzenmacherThaler10,
  author = {Graham Cormode and Michael Mitzenmacher and Justin Thaler},
  att_authors = {gc2602},
  att_private = {false},
  title = {Streaming Graph Computations with a Helpful Advisor},
  booktitle = {European Symposium on Algorithms},
  year = 2010,
  url = {../papers/graphip.pdf},
  link = {http://arxiv.org/abs/1004.2899},
  abstract = {Motivated by the trend to outsource work to commercial cloud computing
services, we consider a variation of the streaming paradigm where a
streaming algorithm can be assisted by a powerful helper that can
provide annotations to the data stream.  We extend previous work on
such {\em annotation models} by considering a number of graph
streaming problems.  Without annotations, streaming algorithms for 
graph problems generally require significant memory;  we show that
for many standard problems, including all graph problems that can 
be expressed with totally unimodular integer programming formulations,
only constant memory is
needed for single-pass algorithms given linear-sized annotations.
We also obtain a protocol achieving \textit{optimal} tradeoffs between
annotation length and memory usage for matrix-vector multiplication;
this result contributes to a trend of recent research on numerical
linear algebra in streaming models.}
}
@inproceedings{ChakrabartiCormodeKondapallyMcGregor10,
  author = {Amit Chakrabarti and Graham Cormode and Ranga Kondapally and Andrew McGregor},
  att_authors = {gc2602},
  att_private = {false},
  title = {Information Cost Tradeoffs for Augmented Index and Streaming Language Recognition},
  booktitle = {{IEEE} Foundations of Computer Science ({FOCS})},
  year = 2010,
  url = {../papers/pqfocs.pdf},
  link = {http://eccc.hpi-web.de/report/2010/076},
  abstract = {
This paper makes three main contributions to the theory of communication
complexity and stream computation. First, we present new bounds on the
information complexity of {\sc Augmented Index}. In contrast to analogous results for {\sc Index}
by Jain, Radhakrishnan and Sen [\emph{J.~ACM}, 2009], we have to overcome the
significant technical challenge that protocols for {\sc Augmented Index} may violate the
``rectangle property'' due to the inherent input sharing. Second, we use these
bounds to resolve an open problem of Magniez, Mathieu and Nayak [\emph{STOC},
2010] that asked about the multi-pass complexity of recognizing Dyck
languages. This results in a natural separation between the standard
multi-pass model and the multi-pass model that permits reverse passes.  Third,
we present the first \emph{passive memory checkers} that verify the
interaction transcripts of priority queues, stacks, and double-ended queues.
We obtain tight upper and lower bounds for these problems, thereby addressing
an important sub-class of  the memory checking framework of Blum et al.
[\emph{Algorithmica}, 1994]. }
}
@inproceedings{CormodeLiLiSrivastava10,
  author = {Graham Cormode and Ninghui Li and Tiancheng Li and Divesh Srivastava},
  att_authors = {gc2602,ds8961},
  att_private = {false},
  title = {Minimizing Minimality and Maximizing Utility: 
Analyzing Method-based attacks on Anonymized Data},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  year = 2010,
  abstract = {The principle of {\em anonymization} for data sharing has become a very popular paradigm for the
preservation of privacy of the data subjects. Since the introduction of $k$-anonymity, dozens of
methods and enhanced privacy definitions have been proposed. However, over-eager attempts to
minimize the information lost by the anonymization potentially allow private information to be
inferred. Proof-of-concept of this ``minimality attack'' has been demonstrated for a variety of
algorithms and definitions \cite{WFW+09}.\\
In this paper, we provide a comprehensive analysis and study of this attack, and demonstrate that
with care its effect can be almost entirely countered. The attack allows an adversary to increase
his (probabilistic) belief in certain facts about individuals over the data. We show that (a) a
large class of algorithms are not affected by this attack, (b) for a class of algorithms that have
a ``symmetric'' property, the attacker's belief increases by at most a small constant, and (c) even
for an algorithm chosen to be highly susceptible to the attack, the attacker's belief when using
the attack increases by at most a small constant factor. We also provide a series of experiments
that show in all these cases that the confidence about the sensitive value of any individual
remains low in practice, while the published data is still useful for its intended purpose. From
this, we conclude that the impact of such method-based attacks can be minimized.}
}
@inproceedings{BhagatCormodeKrishnamurthySrivastava10B,
  author = {Smriti Bhagat and Graham Cormode and Balachander Krishnamurthy and Divesh Srivastava},
  att_authors = {gc2602,ds8961,bk1836},
  att_private = {false},
  title = {Prediction Promotes Privacy In Dynamic Social Networks},
  booktitle = {Workshop on Online Social Networks ({WOSN})},
  year = 2010,
  url = {../papers/danonwosn.pdf},
  abstract = {
Recent work on anonymizing online social networks (OSNs) 
has looked at privacy preserving techniques 
for publishing a single instance of the network. However, OSNs evolve and 
a single instance is inadequate for analyzing their evolution or performing 
longitudinal data analysis. We study the problem of repeatedly publishing 
OSN data as the network evolves while preserving privacy of users. 
Publishing multiple instances independently has privacy risks, since stitching 
the information together may allow an adversary to identify users. We provide 
methods to anonymize a dynamic network when new nodes and edges are added 
to the published network.
These methods 
use link prediction algorithms to model the evolution. Using this predicted 
graph to perform group-based anonymization, the loss in privacy caused by 
new edges can be eliminated almost entirely. 
We propose metrics for privacy loss, and evaluate them for publishing multiple OSN instances.}
}
@inproceedings{BhagatCormodeKrishnamurthySrivastava09,
  title = {Class-based graph anonymization for social network data},
  author = {Smriti Bhagat and Graham Cormode and Balachander Krishnamurthy and Divesh Srivastava},
  att_authors = {gc2602,ds8961,bk1836},
  att_private = {false},
  year = 2009,
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  url = {../papers/ganon.pdf},
  abstract = {
The recent rise in popularity of social networks, such as Facebook and
MySpace, has created large quantities of data about interactions
within these networks. Such data contains many private details about
individuals so anonymization is required prior to attempts to make
the data more widely available for scientific research. 
Prior work has considered simple graph data to be anonymized 
by removing all non-graph information and adding or deleting some edges. 
Since
social network data is richer in details about the users and their
interactions, loss of details due to anonymization limits the
possibility for analysis.  
%%
We present a new set of techniques for
anonymizing social network data based on grouping the entities into
classes, and masking the mapping between entities and the nodes that
represent them in the anonymized graph. Our techniques allow queries
over the rich data to be evaluated with high accuracy while
guaranteeing resilience to certain types of attack. 
To prevent inference of interactions, we rely on a critical 
``safety condition'' when forming these classes. 
We demonstrate utility via empirical data from social networking
settings. We give examples of complex
queries that may be posed and show that they can be answered over the
anonymized data efficiently and accurately.}
}
@inproceedings{CormodeDeligiannakisGarofalakisMcGregor09,
  title = {Probabilistic Histograms for Probabilistic Data},
  author = {Graham Cormode and Antonios Deligiannakis and Minos Garofalakis and Andrew McGregor},
  att_authors = {gc2602},
  att_private = {false},
  year = 2009,
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  abstract = {
There is a growing realization that 
modern database management systems (DBMSs) must be able 
to manage data that contains {\em uncertainties} that are represented in the
form of probabilistic relations. 
Consequently, the design of each 
core DBMS component must be revisited in
the presence of uncertain and probabilistic information.  
In this paper, we study how to build histogram synopses for probabilistic
relations, for the purposes of enabling both  DBMS-internal decisions (such as
indexing and query planning), and (possibly, user-facing) approximate query
processing tools. 
In contrast to initial work in this area, our {\em probabilistic histograms\/}
retain the key
{\em possible-worlds semantics\/} of probabilistic data, allowing for
more accurate, yet concise,  representation of the uncertainty 
characteristics of data and query results.
We present a variety of techniques for building optimal 
probabilistic histograms, each one tuned to a different choice 
of approximation-error metric. 
We show that these can be incorporated into a general Dynamic Programming (DP)
framework, which generalizes that used for existing histogram constructions. 
The end result is a histogram where each ``bucket'' is approximately represented by a
compact probability distribution function (PDF), which can 
be used as the basis for query planning and approximate query answering. 
We present novel, polynomial-time algorithms to find optimal probabilistic 
histograms for a variety of PDF-error metrics (including 
variation distance, sum squared error, max error, 
and variants of EMD).  
Our experimental study shows that our probabilistic histogram synopses can 
accurately capture the key statistical properties of uncertain data, while being
much more compact to store and work with than the original uncertain
relations. }
}
@inproceedings{CormodeChakrabartiMcGregor09,
  title = {Annotations in Data Streams},
  author = {Amit Chakrabarti and Graham Cormode and Andrew McGregor},
  booktitle = {International Colloquium on Automata, 
                  Languages and Programming ({ICALP})},
  year = 2009,
  att_authors = {gc2602},
  att_private = {false},
  url = {../papers/annotations.pdf},
  abstract = {
The central goal of data stream algorithms is to process massive streams
of data using {\em sublinear} storage space. 
Motivated by work in the database community on outsourcing database and data stream processing, we ask whether the space usage of such algorithms be further
reduced by enlisting a more powerful ``helper'' who can {\em annotate}
the stream as it is read. We do not wish to blindly trust the helper, so we
require that the algorithm be convinced of having computed a correct
answer. 
We show
upper bounds that achieve a non-trivial tradeoff between the
amount of annotation used and the space required to verify it.  We also
prove lower bounds on such tradeoffs, often nearly matching the upper
bounds, via notions related to Merlin-Arthur communication complexity.
Our results cover the classic data stream problems of selection,
frequency moments, and fundamental graph problems such as
triangle-freeness and connectivity. 
  Our work is also part of a
growing trend --- including recent studies of multi-pass streaming,
read/write streams and randomly ordered streams --- of asking more
complexity-theoretic questions about data stream processing. 
It is a recognition that, in addition to practical relevance, the
data stream model
raises many interesting theoretical questions in its own right. }
}
@inproceedings{CormodeGolabKornMcGregorSrivastavaZhang09,
  title = {Estimating the Confidence of Conditional Functional Dependencies},
  author = {Graham Cormode and Lukasz Golab and Flip Korn and Andrew McGregor
   			and Divesh Srivastava and Xi Zhang},
  att_authors = {gc2602,ds8961,pk1785,lg1173},
  att_private = {false},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  year = 2009,
  url = {../papers/cfdest.pdf},
  abstract = {
  Conditional functional dependencies (CFDs) have recently been proposed as
  extensions of classical functional dependencies that apply to a certain
  subset of the relation, as specified by a \emph{pattern tableau}.  
  Calculating the support and confidence of a CFD (i.e., the size of the 
  applicable subset and the extent to which it satisfies the CFD)
  gives valuable information about data semantics and data quality.  
  While computing the support is easier, computing the confidence 
  exactly is expensive if the relation is large, and estimating it from a 
  random sample of the relation is unreliable unless the sample is large.  
  
  We study how to efficiently estimate the confidence of a CFD 
  with a small number of passes (one or two) over the input using small space. 
  Our solutions are based on a variety of sampling and sketching techniques,
  and apply when the pattern tableau is known in advance, and also the harder 
  case when this is given after the data have been seen. 
  We analyze our algorithms, and show that they can guarantee 
  a small additive error; we also show that relative errors guarantees are not possible. 
  We demonstrate the power of these methods empirically, with a detailed study over
  a mixture of real and synthetic data. 
  These experiments show that it is possible to estimates the CFD confidence
  very accurately with summaries which are much smaller than the size of the data they represent.}
}
@inproceedings{BerindeCormodeIndykStrauss09,
  title = {Space-optimal Heavy Hitters with Strong Error Bounds},
  author = {Radu Berinde and Graham Cormode and Piotr Indyk and Martin Strauss},
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  year = 2009,
  url = {../papers/counters.pdf},
  slides = {../slides/counters.pdf},
  abstract = {
  The problem of finding heavy-hitters and approximating the frequencies
  of items is at the heart of many problems in data stream analysis. It
  has been observed that several proposed solutions to this problem can
  outperform their worst-case guarantees on real data. This leads to the
  question of whether some stronger bounds can be guaranteed. We answer
  this in the positive by showing that a class of ``counter-based
  algorithms'' (including the popular and very space-efficient Frequent
  and Spacesaving algorithms) provide much stronger approximation
  guarantees than previously known. Specifically, we show that errors in
  the approximation of individual elements do not depend on the
  frequencies of the most frequent elements, but only on the frequency
  of the remaining ``tail''.
  This shows that counter-based methods are the
  most space-efficient (in fact, space-optimal) algorithms having this
  strong error bound. 
  
  This tail guarantee allows these algorithms to solve the ``sparse
  recovery'' problem. Here, the goal is to recover a faithful
  representation of the vector of frequencies, $f$. We prove that using
  space $O(k)$, the algorithms construct an approximation $f^*$ to the
  frequency vector $f$ so that the L1 error $||f-f^*||_1$ is close to the best
  possible error $\min_{f'} ||f' - f||_1$, where $f'$ ranges over all vectors
  with at most $k$ non-zero entries. This improves the previously best
  known space bound of about $O(k \log n)$ for streams without element
  deletions. Other consequences of the tail guarantees are results for skewed
  (Zipfian) data, and guarantees for accuracy of merging multiple
summarized streams.  }
}
@inproceedings{CormodeTirthapuraXu09,
  title = {Time-Decayed Correlated Aggregates over Data Streams},
  author = {Graham Cormode and Srikanta Tirthapura and Bojian Xu},
  booktitle = {{SIAM} Conference on Data Mining ({SDM})},
  att_authors = {gc2602},
  att_private = {false},
  year = 2009,
  url = {../papers/corag.pdf},
  abstract = {
Data stream analysis frequently relies on identifying correlations
and posing conditional queries on the data after it has been seen.
Correlated aggregates form an important example of such queries,
which ask for an aggregation over one dimension of stream elements
which satisfy a predicate on another dimension. Since recent
events are typically more important than older ones, time decay
should also be applied to downweight less significant values. We
present space-efficient algorithms as well as space lower bounds
for the time-decayed correlated sum, a problem at the heart of many
related aggregations. By considering different fundamental classes
of decay functions, we separate cases where efficient relative error
or additive error is possible, from other cases where linear space is
necessary to approximate. In particular, we show that no efficient
algorithms are possible for the popular sliding window and exponential
decay models, resolving an open problem. The results are
surprising, since efficient approximations are known for other data
stream problems under these decay models. This is a step towards
better understanding which sophisticated queries can be answered
on massive streams using limited memory and computation.  
}
}
@inproceedings{CormodeGarofalakis09,
  title = {Histograms and Wavelets on Probabilistic Data},
  author = {Graham Cormode and Minos Garofalakis},
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2009,
  slides = {../slides/phistslides.pdf},
  url = {../papers/phist.pdf},
  note = {Best paper award},
  abstract = {
There is a growing realization that uncertain information
is a first-class citizen in modern database management.
As such, we need techniques to correctly and efficiently
process uncertain data in database systems. In particular, data
reduction techniques that can produce concise, accurate synopses
of large probabilistic relations are crucial. Similar to their
deterministic relation counterparts, such compact probabilistic
data synopses can form the foundation for human understanding
and interactive data exploration, probabilistic query planning
and optimization, and fast approximate query processing in
probabilistic database systems.
In this paper, we introduce definitions and algorithms for
building histogram- and Haar wavelet-based synopses on probabilistic
data. The core problem is to choose a set of histogram
bucket boundaries or wavelet coefficients to optimize the accuracy
of the approximate representation of a collection of probabilistic
tuples under a given error metric. For a variety of different
error metrics, we devise efficient algorithms that construct
optimal or near optimal size B histogram and wavelet synopses.
This requires careful analysis of the structure of the probability
distributions, and novel extensions of known dynamicprogramming-
based techniques for the deterministic domain.
Our experiments show that this approach clearly outperforms
simple ideas, such as building summaries for samples drawn 
from the data distribution, while taking equal or less time.}
}
@inproceedings{CormodeLiYi09,
  title = {Semantics of Ranking Queries for Probabilistic Data and Expected Ranks},
  author = {Graham Cormode and Feifei Li and Ke Yi},
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2009,
  url = {../papers/exprank.pdf},
  abstract = {
When dealing with massive quantities of data, topk
queries are a powerful technique for returning only the k
most relevant tuples for inspection, based on a scoring function.
The problem of efficiently answering such ranking queries has
been studied and analyzed extensively within traditional database
settings. The importance of the top-k is perhaps even greater in
probabilistic databases, where a relation can encode exponentially
many possible worlds. There have been several recent attempts
to propose definitions and algorithms for ranking queries over
probabilistic data. However, these all lack many of the intuitive
properties of a top-k over deterministic data. Specifically, we
define a number of fundamental properties, including exact-k,
containment, unique-rank, value-invariance, and stability, which
are all satisfied by ranking queries on certain data. We argue
that all these conditions should also be fulfilled by any reasonable
definition for ranking uncertain data. Unfortunately, none of the
existing definitions is able to achieve this.
To remedy this shortcoming, this work proposes an intuitive
new approach of expected rank. This uses the well-founded notion
of the expected rank of each tuple across all possible worlds
as the basis of the ranking. We are able to prove that, in
contrast to all existing approaches, the expected rank satisfies
all the required properties for a ranking query. We provide
efficient solutions to compute this ranking across the major
models of uncertain data, such as attribute-level and tuple-level
uncertainty. For an uncertain relation of N tuples, the processing
cost is O(N logN)no worse than simply sorting the relation.
In settings where there is a high cost for generating each tuple
in turn, we show pruning techniques based on probabilistic tail
bounds that can terminate the search early and guarantee that
the top-k has been found. Finally, a comprehensive experimental
study confirms the effectiveness of our approach.}
}
@inproceedings{CormodeShkapenyukSrivastavaXu09,
  author = {Graham Cormode and Vladislav Shkapenyuk and Divesh Srivastava and Bojian Xu},
  title = {Forward Decay: A Practical Time Decay Model for Streaming Systems},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2009,
  att_authors = {gc2602,ds8961,vs9593},
  att_private = {false},
  url = {../papers/fwddecay.pdf},
  abstract = {
Temporal data analysis in data warehouses and data
streaming systems often uses time decay to reduce the importance
of older tuples, without eliminating their influence, on the results
of the analysis. Exponential time decay is commonly used in
practice, even though other decay functions (e.g., polynomial
decay) have also been identified as useful. We argue that this
is because the usual definitions of time decay are backwards:
the decayed weight of a tuple is based on its age, measured
backward from the current time. Since this age is constantly
changing, such decay is too complex and unwieldy for scalable
implementation.
In this paper, we propose a new class of forward decay
functions based on measuring forward from a fixed point in
time. We show that this model captures the more practical
models already known, such as exponential decay and landmark
windows, but also includes a wide class of other types of time
decay. We provide efficient algorithms to compute a variety of
aggregates and draw samples under forward decay, and show
that these are easy to implement scalably. Further, we provide
empirical evidence that these can be executed in a production
data stream management system with little or no overhead
compared to the undecayed computations. Our implementation
required no extensions to the query language or the DSMS,
demonstrating that forward decay represents a practical model
of time decay for systems that deal with time-based data.}
}
@inproceedings{CormodeHadjieleftheriou08,
  author = {Graham Cormode and Marios Hadjieleftheriou},
  title = {Finding Frequent Items in Data Streams},
  att_authors = {gc2602,mh6516},
  att_private = {false},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  year = 2008,
  url = {../papers/freq.pdf},
  note = {Best paper award},
  pubsite = {http://www.research.att.com/~marioh/frequent-items/index.html},
  abstract = {
The frequent items problem is to process a stream of items and find
all items occurring more than a given fraction of the time. It is one
of the most heavily studied problems in data stream mining, dating
back to the 1980s. Many applications rely directly or indirectly
on finding the frequent items, and implementations are in use in
large scale industrial systems. However, there has not been much
comparison of the different methods under uniform experimental
conditions. It is common to find papers touching on this topic in
which important related work is mischaracterized, overlooked, or
reinvented.
In this paper, we aim to present the most important algorithms
for this problem in a common framework. We have created baseline
implementations of the algorithms, and used these to perform a
thorough experimental study of their properties. We give empirical
evidence that there is considerable variation in the performance of
frequent items algorithms. The best methods can be implemented
to find frequent items with high accuracy using only tens of kilobytes
of memory, at rates of millions of items per second on cheap
modern hardware.}
}
@inproceedings{CormodeSrivastavaYuZhang08,
  author = {Graham Cormode and Divesh Srivastava and Ting Yu and Qing Zhang},
  title = {Anonymizing Bipartite Graph Data using Safe Groupings},
  att_authors = {gc2602,ds8961},
  att_private = {false},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  year = 2008,
  url = {../papers/gpriv.pdf},
  abstract = {Private data often comes in the form of associations between entities, such
as customers and products bought from a pharmacy, which are naturally
represented in the form of a large, sparse bipartite graph. As with tabular
data, it is desirable to be able to publish anonymized versions of such data,
to allow others to perform ad hoc analysis of aggregate graph properties.
However, existing tabular anonymization techniques do not give useful or
meaningful results when applied to graphs: small changes or masking of
the edge structure can radically change aggregate graph properties.
We introduce a new family of anonymizations, for bipartite graph data,
called (k, l)-groupings. These groupings preserve the underlying graph
structure perfectly, and instead anonymize the mapping from entities to
nodes of the graph. We identify a class of safe (k, l)-groupings that have
provable guarantees to resist a variety of attacks, and show how to find such
safe groupings. We perform experiments on real bipartite graph data to
study the utility of the anonymized version, and the impact of publishing
alternate groupings of the same graph data. Our experiments demonstrate
that (k, l)-groupings offer strong tradeoffs between privacy and utility.}
}
@inproceedings{CormodeKornTirthapura08PODS,
  author = {Graham Cormode and Flip Korn and Srikanta Tirthapura},
  title = {Time-Decaying Aggregates in Out-of-order Streams},
  att_authors = {gc2602,pk1785},
  att_private = {false},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  year = 2008,
  url = {../papers/decaypods.pdf},
  slides = {../slides/decay-pods.pdf},
  abstract = {
  Processing large data streams is now a major topic in data management.
  The data involved can be truly massive, and the required
  analyses complex. In a stream of sequential events such as stock
  feeds, sensor readings, or IP traffic measurements, data tuples pertaining
  to recent events are typically more important than older
  ones. This can be formalized via time-decay functions, which assign
  weights to data based on the age of data. Decay functions
  such as sliding windows and exponential decay have been studied
  under the assumption of well-ordered arrivals, i.e., data arrives in
  non-decreasing order of time stamps. However, data quality issues
  are prevalent in massive streams (due to network asynchrony and
  delays etc.), and correct arrival order is not guaranteed.

  We focus on the computation of decayed aggregates such as
  range queries, quantiles, and heavy hitters on out-of-order streams,
  where elements do not necessarily arrive in increasing order of
  timestamps. Existing techniques such as Exponential Histograms
  and Waves are unable to handle out-of-order streams. We give the
  first deterministic algorithms for approximating these aggregates
  under popular decay functions such as sliding window and polynomial
  decay. We study the overhead of allowing out-of-order arrivals
  when compared to well-ordered arrivals, both analytically and experimentally.
  Our experiments confirm that these algorithms can be
  applied in practice, and compare the relative performance of different
approaches for handling out-of-order arrivals.}
}
@inproceedings{CormodeMcGregor08,
  author = {Graham Cormode and Andrew McGregor},
  title = {Approximation Algorithms for Clustering Uncertain Data},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  att_authors = {gc2602},
  att_private = {false},
  year = 2008,
  url = {../papers/pclust.pdf},
  slides = {../slides/pclust-pods.pdf},
  abstract = {
  There is an increasing quantity of data with uncertainty arising from
  applications such as sensor network measurements, record linkage, and
  as output of mining algorithms. 
  This uncertainty is typically formalized as probability density
  functions over tuple values. 
  Beyond storing and processing such data in a DBMS, it is necessary to
  perform other data analysis tasks such as data mining. 
  We study the core mining problem of {\em clustering} on uncertain
  data, and define appropriate natural generalizations of standard
  clustering optimization criteria. 
  Two variations arise, depending on whether a point is automatically
  associated with its optimal center, or whether it must be assigned to
  a fixed cluster no matter where it is actually located. 
  
  For uncertain versions of $k$-means and $k$-median, we show reductions
  to their corresponding weighted versions on data with no
  uncertainties. 
  These are simple in the unassigned case, but require some care for the
  assigned version. 
  Our most interesting results are for uncertain $k$-center, 
  which generalizes both traditional $k$-center and $k$-median objectives.
  We show a variety of bicriteria approximation algorithms. 
  One picks ${poly}(\log n)$ more centers and achieves a
  $(1+\epsilon)$ approximation to the best uncertain $k$-centers. 
  Another picks $2k$ centers and achieves a constant factor
  approximation. 
  Collectively, these results are the first known guaranteed
  approximation algorithms for the problems of clustering uncertain
data.  }
}
@inproceedings{CormodeKornMuthukrishnanSrivastava08BQ,
  author = {Graham Cormode and Flip Korn and S. Muthukrishnan and Divesh Srivastava},
  att_authors = {gc2602,ds8961,pk1785},
  att_private = {false},
  booktitle = {Scientific and Statistical Database Management ({SSDBM})},
  title = {Summarizing Two-Dimensional Data
                  with Skyline-based Statistical Descriptors},
  year = 2008,
  url = {../papers/2dquant.pdf},
  abstract = {
Much real data consists of more than one dimension,
such as financial transactions (eg, price $\times$ volume)
and IP network flows (eg, duration $\times$ numBytes),
and capture relationships between the variables.
For a single dimension, quantiles are intuitive and robust descriptors.
Processing and analyzing such data, particularly in data warehouse or data
streaming settings, requires similarly robust and informative
statistical descriptors that go beyond one-dimension.
Applying quantile methods to summarize a multidimensional
distribution along only singleton attributes
ignores the rich dependence amongst the variables.

In this paper, we present new skyline-based statistical
descriptors for capturing the distributions over pairs of dimensions.
They generalize the notion of quantiles in the individual
dimensions, and also incorporate properties of the joint distribution.
We introduce {\em $\phi$-quantours} and {\em $\alpha$-radials}, 
which are skyline points over subsets of the data, and propose
{\em ($\phi, \alpha$)-quantiles}, found from the union of these skylines,
as statistical descriptors of two-dimensional distributions.
We present efficient online algorithms for tracking
($\phi,\alpha$)-quantiles on two-dimensional streams
using guaranteed small space.
We identify the principal properties of the proposed descriptors
and perform extensive experiments with synthetic and real IP traffic data
to study the efficiency of our
proposed algorithms.}
}
@inproceedings{ChakrabartiCormodeMcGregor08,
  author = {Amit Chakrabarti and Graham Cormode and Andrew McGregor},
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {{ACM} Symposium on Theory of Computing ({STOC})},
  title = {Robust Lower Bounds for Communication and Stream Computation},
  year = 2008,
  url = {../papers/robustbounds.pdf},
  slides = {../slides/rcc-stoc.pdf},
  abstract = {
In this paper we study the communication complexity of evaluating
functions when the input data is distributed (according to some known
distribution) to two or more players. This naturally extends
previous models such as the best-case and worst-case partition models
to a model that captures random
partitions and distributions with redundancy  where bits of data may
be known to more than one player. 
This allows us to distinguish functions which require significant
communication for almost every allocation of input from those for
which only a few cases are hard. 

There is also a strong connection to proving lower-bounds in the data
stream model that are ``robust'' to the ordering of the data. 
That is, we prove lower-bounds for when
  order of the data-stream is not chosen adversarially
  but rather is determined uniformly (or near-uniformly) 
  from the set of all permuations. 
This streaming model has attracted recent interest, since lower bounds
  here give stronger evidence for the inherent hardness of streaming
  problems. 

Our results include the first average-partition communication
lower-bounds for problems including multi-party set-disjointness and
gap-hamming. Both are tight. We also improve extend and improve
previous results for a form of pointer-jumping
 that is relevant to the problem of selection. Collectively, 
these results yield lower-bounds for a
variety of problems in the random-order, data-stream model including
estimating the number of distinct elements, approximating frequency
moments, and quantile estimation.}
}
@inproceedings{CormodeKornMuthukrishnanWu08,
  author = {Graham Cormode and Flip Korn and S. Muthukrishnan and 
Yihua Wu},
  att_authors = {gc2602,pk1785},
  att_private = {false},
  title = {On Signatures for Communication Graphs},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  url = {../papers/graphsig.pdf},
  year = 2008,
  abstract = {
Communications between individuals can be represented by (weighted, 
multi-)
graphs. 
Many applications operate on communication graphs
associated with telephone calls, emails, Instant Messages (IM), 
blogs, web forums, e-business relationships and so on. 
These applications include identifying repetitive
fraudsters, message board aliases, multihoming of IP addresses, etc. 
Tracking such electronic
identities on communication network can be achieved if we have a
reliable ``signature'' for nodes and activities. While many examples
of adhoc signatures can be proposed for particular tasks, what is needed
is a systematic study of the principles behind the usage of signatures
in any task. 

We develop a formal framework for the use of signatures
on communication graphs and identify three fundamental properties
that are natural to signature schemes: persistence, uniqueness and
robustness.  We justify these properties by showing how they impact
a set of applications.  We then explore several signature
schemes --- previously defined or new --- in our framework 
and evaluate them on real data in terms
of these properties.  This provides insights into suitable
signature schemes for desired applications.
Finally, as case studies, we focus on the concrete application 
of enterprise network traffic.}
}
@inproceedings{CormodeKornTirthapura08,
  author = {Graham Cormode and Flip Korn and Srikanta Tirthapura},
  att_authors = {gc2602,pk1785},
  att_private = {false},
  title = {Exponentially Decayed Aggregates on Data Streams},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = 2008,
  slides = {../slides/expdecay.pdf},
  url = {../papers/expdecay.pdf},
  abstract = {In a massive stream of sequential events such as
stock feeds, sensor readings, or IP traffic measurements, tuples
pertaining to recent events are typically more important than
older ones. It is important to compute various aggregates over
such streams after applying a decay function which assigns
weights to tuples based on their age. We focus on the computation
of exponentially decayed aggregates in the form of quantiles and
heavy hitters. Our techniques are based on extending existing
data stream summaries, such as the q-digest and the``spacesaving''
algorithm. Our experiments confirm that our methods
can be applied in practice, and have similar space and time costs
to the non-decayed aggregate computation.}
}
@inproceedings{CormodeMuthukrishnanYi08,
  author = {G. Cormode and S. Muthukrishnan and K. Yi},
  att_authors = {gc2602},
  att_private = {false},
  title = {Algorithms for Distributed, Functional Monitoring},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  year = 2008,
  url = {../papers/soda08.pdf},
  abstract = {
We study what we call {\em functional monitoring} problems. 
We have 
$k$ players each tracking their inputs, say $A_j(t)$ up until time $t$, 
and communicating with a central coordinator. 
The coordinator's task is to
compute a given function $f$ over the union of the inputs
$\cup_{i} A_j(t)$, {\em continuously} at all times $t$. 
The goal is to minimize the number of bits communicated between the 
players and the coordinator.
A simple example is when $f$ is simply the sum, and the coordinator
is required to alert when the sum of a distributed 
set of values exceeds a given threshold. 
Of interest is the approximate version where the coordinator outputs
$1$ if $f \geq \tau$ and $0$ is $f\leq \tau-\epsilon$. 
This defines
the $(k,f,\tau,\epsilon)$ distributed, functional monitoring problem.

Functional monitoring problems are fundamental in distributed systems, 
in particular sensor networks, where we must minimize communication; they also
connect to problems in communication complexity, communication theory, 
and signal processing. Yet few formal bounds are known for 
functional monitoring. 

We give upper and lower bounds for 
the $(k,f,\tau,\epsilon)$ problem for foundational $f$'s. 
In particular, we study frequency moments ($F_0, F_1, F_2$). 
We show that for the simplest cases, the cost of monitoring a
function can be the same as the cost of its one-time computation. 
However, for $F_2$ we show a carefully constructed multi-round
algorithm which uses ``sketch summaries'' at multiple levels of detail
and solves the $(k,F_2,\tau,\epsilon)$ problem  
with communication ${O}$~$(k^2/\epsilon + k^{3/2}/\epsilon^3)$.
Since frequency moment estimation is central to other problems,
our results have immediate applications to histograms, wavelet computations,
and others. Our algorithmic techniques are likely to be useful for other 
functional monitoring problems. 
}
}
@inproceedings{BhagatCormodeRozenbaum07,
  author = {S. Bhagat and G. Cormode and I. Rozenbaum},
  att_authors = {gc2602},
  att_private = {false},
  title = {Applying Link-based Classification to Label Blogs},
  booktitle = {Joint WEBKDD and SNA-KDD Workshop},
  year = {2007},
  url = {../papers/sigkdd07.pdf},
  slides = {../slides/SNAKDD-blogs.pdf},
  abstract = {
In analyzing data from social and communication networks, we encounter
the problem of classifying objects where there is an explicit
link structure amongst the objects. We study the problem of inferring
the classification of all the objects from a labeled subset, using
only the link-based information amongst the objects.
We abstract the above as a labeling problem on multigraphs with
weighted edges. We present two classes of algorithms, based on
local and global similarities. Then we focus on multigraphs induced
by blog data, and carefully apply our general algorithms to
specifically infer labels such as age, gender and location associated
with the blog based only on the link-structure amongst them. We
perform a comprehensive set of experiments with real, large-scale
blog data sets and show that significant accuracy is possible from
little or no non-link information, and our methods scale to millions
of nodes and edges.}
}
@inproceedings{GangulyCormode07,
  author = {S. Ganguly and G. Cormode},
  att_authors = {gc2602},
  att_private = {false},
  title = {On Estimating Frequency Moments of Data Streams},
  booktitle = {Proceedings of {RANDOM}},
  year = {2007},
  url = {../papers/gc-l1.pdf},
  abstract = {Space-economical  estimation of the $p$th frequency moments, defined as
$F_p = \Sigma_{i=1}^n |{f_i}|^p$, for $p > 0$,  are of interest in
estimating all-pairs distances in a large data
matrix,
machine learning, and in data stream computation. 
Random sketches formed by the inner product  of the
frequency vector $f_1, \ldots, f_n$ with a suitably chosen random
vector were pioneered by Alon, Matias and Szegedy, and 
have since played a
central role in estimating $F_p$ and for data stream computations in
general. 
The concept of $p$-stable sketches formed by the inner product of the 
frequency
vector with a random vector whose components are drawn from a
$p$-stable distribution, was proposed by Indyk
for estimating $F_p$, for $0 < p < 2$, and has been further studied
by Li.

In this paper, we consider the problem of estimating $F_p$, for $0 <
p < 2$.  A disadvantage of the stable sketches technique and its
variants is that they require $O(\frac{1}{\epsilon^2})$ 
inner-products of the frequency
vector with dense vectors of stable 
(or nearly stable) random variables to be maintained.
This means that each stream update can be quite time-consuming. 
We present algorithms for estimating $F_p$, for $0 < p < 2$, 
that does not require the use of stable sketches or its
approximations.  
Our  technique is elementary in nature, in that, it
uses simple randomization in conjunction with well-known summary
structures for data streams, such as the CM sketch 
sketch and the Count sketch structure.
Our algorithms require space
${O}~(\frac{1}{\epsilon^{2+p}})$ to estimate $F_p$ to within $1 +/-
\epsilon$ factors and requires expected time $O(\log F_1 \log
\frac{1}{\delta})$ to process each update. Thus, our technique
trades an $O(\frac{1}{\epsilon^p})$ factor in space for
much more efficient processing of stream updates. 
We also present a stand-alone iterative estimator for $F_1$.
}
}
@inproceedings{CormodeTirthapuraXu07,
  author = {G. Cormode and S. Tirthapura and B. Xu},
  att_authors = {gc2602},
  att_private = {false},
  title = {Time-Decaying Sketches for Sensor Data Aggregation},
  booktitle = {{ACM} Principles of Distributed Computing ({PODC})},
  year = {2007},
  url = {../papers/timedecaypodc.pdf},
  abstract = {We present a new sketch for summarizing network data. 
The sketch has the following properties which make it useful in
communication-efficient aggregation in distributed streaming
scenarios, such as sensor networks: the sketch is duplicateinsensitive,
i.e. re-insertions of the same data will not affect
the sketch, and hence the estimates of aggregates. Unlike
previous duplicate-insensitive sketches for sensor data aggregation, 
it is also time-decaying, so that the weight
of a data item in the sketch can decrease with time according
to a user-specified decay function. The sketch can give
provably approximate guarantees for various aggregates of
data, including the sum, median, quantiles, and frequent elements.
The size of the sketch and the time taken to update
it are both polylogarithmic in the size of the relevant data.
Further, multiple sketches computed over distributed data
can be combined with the same guarantees. To our knowledge,
this is the first sketch that combines all the above
properties.}
}
@inproceedings{CormodeGarofalakis07,
  att_authors = {gc2602},
  att_private = {false},
  author = {G. Cormode and M. Garofalakis},
  title = {Sketching Probabilistic Data Streams},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  year = {2007},
  url = {../papers/probstreams.pdf},
  slides = {../slides/probstreams.pdf},
  abstract = {The management of uncertain, probabilistic data has 
recently emerged
as a useful paradigm for dealing with the inherent unreliabilities
of several real-world application domains, including data cleaning,
information integration, and pervasive, multi-sensor computing.
Unlike conventional data sets, a set of probabilistic tuples defines
a probability distribution over an exponential number of possible
worlds (i.e., grounded, deterministic databases). This possible
worlds interpretation allows for clean query semantics but
also raises hard computational problems for probabilistic database
query processors. To further complicate matters, in many scenarios
(e.g., large-scale process and environmental monitoring using multiple
sensor modalities), probabilistic data tuples arrive and need to
be processed in a streaming fashion; that is, using limited memory
and CPU resources and without the benefit of multiple passes over a
static probabilistic database. Such probabilistic data streams raise a
host of new research challenges for stream-processing engines that,
to date, remain largely unaddressed.

In this paper, we propose the first space- and time-efficient algorithms
for approximating complex aggregate queries (including,
the number of distinct values and join/self-join sizes) over probabilistic
data streams. Following the possible-worlds semantics,
such aggregates essentially define probability distributions over the
space of possible aggregation results, and our goal is to characterize
such distributions through efficient approximations of their key
moments (such as expectation and variance). Our algorithms offer
strong randomized estimation guarantees while using only sublinear
space in the size of the stream(s), and rely on novel, concise
streaming sketch synopses that extend conventional sketching ideas
to the probabilistic streams setting. Our experimental results verify
the effectiveness of our approach. }
}
@inproceedings{BhagatCormodeMuthukrishnanRozenbaumXue07,
  att_authors = {gc2602},
  att_private = {false},
  author = {S. Bhagat and G. Cormode and S. Muthukrishnan and 
                     I. Rozenbaum and H. Xue},
  title = {No Blog is an Island Analyzing Connections Across 
                     Information Networks},
  year = {2007},
  url = {../papers/icwsm07.pdf},
  booktitle = {International Conference on Weblogs and Social Media},
  abstract = {
There are multiple information and social networks among individuals,
from telephone to email, web, blog, instant message (IM) and
chat networks. Prior work has studied each of these individual networks
quite extensively, including telephone networks, postal
network, web communities, and so on. Still, what is of
great interest is how these networks collide and interact with each
other. Each individual has presence in multiple networks and it will
be of interest to understand how the references and presence in one
affects that in the others. Previously such studies would have been
difficult to do since each source of data was often owned by a specific
entity. Blogs now provide a unique, public source of data that
naturally provides visibility into multiple networks. For example,
blog entries cite web pages, blogs have friends and community, as
well as blog profiles that have links to email, IM and other networks.
In this paper, we use the blogs as a starting point to pull in data about
these multiple networks and study how these multiple networks interact
with each other. While much of the connections are within
specific networks, there is still a wealth of information, structure and
form to the connections across these networks as our analyses show.}
}
@inproceedings{CormodeMuthukrishnanZhuang07,
  att_authors = {gc2602},
  att_private = {false},
  author = {G. Cormode and S. Muthukrishnan and W. Zhuang},
  title = {Conquering the Divide: 
                   Continuous Clustering of Distributed Data Streams},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  year = {2007},
  url = {../papers/cdclust.pdf},
  abstract = {Data is often collected over a distributed network, but
in many cases, is so voluminous that it is impractical and
undesirable to collect it in a central location. Instead, we
must perform distributed computations over the data, guaranteeing
high quality answers even as new data arrives. In
this paper, we formalize and study the problem of maintaining
a clustering of such distributed data that is continuously
evolving. In particular, our goal is to minimize the communication
and computational cost, still providing guaranteed
accuracy of the clustering.

We focus on the k-center clustering, and provide a suite
of algorithms that vary based on which centralized algorithm
they derive from, and whether they maintain a single
global clustering or many local clusterings that can be
merged together. We show that these algorithms can be designed
to give accuracy guarantees that are close to the best
possible even in the centralized case. In our experiments,
we see clear trends among these algorithms, showing that
the choice of algorithm is crucial, and that we can achieve a
clustering that is as good as the best centralized clustering,
with only a small fraction of the communication required to
collect all the data in a single location.}
}
@inproceedings{ChakrabartiCormodeMcGregor07,
  author = {A. Chakrabarti and G. Cormode and A. McGregor},
  att_authors = {gc2602},
  att_private = {false},
  title = {A Near-Optimal Algorithm for Computing the Entropy of 
                     a Stream},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  year = {2007},
  url = {../papers/entropy-soda06.pdf},
  abstract = {
  We describe a simple algorithm for computing the empirical entropy of
  a stream of $m$ values in a single pass, using $O(\epsilon^{-2} \log
  (\delta^{-1}) \log^2 m)$ words of space.  Our algorithm is based upon
  a novel extension of a method introduced by Alon, Matias,  and Szegedy. 
  We show a space lower bound of $\Omega({\epsilon^{-2}/\log
  (1/\epsilon)})$, meaning that our algorithm is near optimal in terms of
  its dependency on $\epsilon$.  This improves over previous work on this
  problem.  We show that generalizing to
  $k$th order entropy requires close to linear space, and give additive
  approximations using our algorithm. 
}
}
@inproceedings{CormodeMuthukrishnan06CS,
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {Combinatorial Algorithms for Compressed Sensing},
  booktitle = {SIROCCO},
  year = {2006},
  url = {../papers/cs-sirocco.pdf},
  slides = {../slides/cs-siroccoslides.pdf},
  pubsite = {http://www.springerlink.com/openurl.asp?genre=article&id=doi:10.1007/11780823_22},
  note = {},
  abstract = {
In sparse approximation theory, the fundamental problem is to 
reconstruct 
a
signal A in R^n from linear measurements  with respect to a 
dictionary of psi_i's. Recently, there is focus on the novel direction 
of 
Compressed Sensing [Donoho:04] where the reconstruction can be done with 
very few---O(k log n)---linear measurements over a modified dictionary 
if 
the signal is compressible, that is, its information is concentrated in 
k 
coefficients with the original dictionary. In particular, these results 
prove that there exists a single O(k log n) x n measurement matrix such 
that any such signal can be reconstructed from these measurements, with 
error at most O(1) times the worst case error for the class of such 
signals.  Compressed sensing has generated tremendous excitement both 
because of the sophisticated underlying Mathematics and because of its 
potential applications.

In this paper, we address outstanding open problems in Compressed 
Sensing.  
Our main result is an explicit construction of a non-adaptive 
measurement 
matrix and the corresponding reconstruction algorithm
so that with a number of measurements polynomial in k, log n, 1/$epsilon$,
we can reconstruct compressible signals. This is the first known 
polynomial time explicit construction of any such measurement matrix. In 
addition,  our result improves the error guarantee from O(1) to 1 + 
$\epsilon$ and  improves the reconstruction time from poly{n} to poly{k log 
n}.

Our second result is a randomized construction of O(k polylog{n}) 
measurements that work for each signal with high probability and gives 
per-instance approximation guarantees rather than over the class of all 
signals. Previous work on Compressed Sensing does not provide such 
per-instance approximation guarantees; our result improves the best 
known 
number of measurements known from prior work in other areas including 
Learning Theory, Streaming algorithms and Complexity Theory for this 
case.

Our approach is combinatorial. In particular, we use two parallel sets 
of 
group tests, one to filter and the other to certify and estimate; the 
resulting algorithms are quite simple to implement.}
}
@inproceedings{CormodeKornMuthukrishnanSrivastava06,
  author = {G. Cormode and F. Korn and S. Muthukrishnan and D. Srivastava},
  att_authors = {gc2602,pk1785,ds8961},
  att_private = {false},
  title = {Space- and Time-Efficient Deterministic Algorithms for Biased Quantiles over Data Streams},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  year = {2006},
  note = {},
  url = {../papers/bq-pods.pdf},
  slides = {../slides/bquant-short.pdf},
  abstract = {
Skew is prevalent in data streams, and should be taken into account
by algorithms that analyze the data.
The problem of finding ``biased quantiles''--- that is, approximate
quantiles which must be more accurate for more extreme values ---
is a framework for summarizing such skewed data on data streams.
We present the first deterministic algorithms for answering biased
quantiles queries accurately with small---sublinear in the input size---
space and time bounds in one pass.
The space bound is near-optimal, and the amortized update cost is
close to constant,
making it practical for handling high speed network data streams.
We demonstrate that not only can we prove theoretical properties of the
algorithm, but that it also uses less space than existing methods in many
practical settings, and is fast to maintain.}
}
@inproceedings{CormodeKeralapuraRamimirtham06,
  author = {G. Cormode and R. Keralapura and J. Ramimirtham},
  att_authors = {gc2602},
  att_private = {false},
  title = {Communication-Efficient Distributed Monitoring of Thresholded Counts},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  url = {../papers/cdthresh.pdf},
  slides = {../slides/cdthresh.ppt},
  note = {},
  year = {2006},
  abstract = {
Monitoring is an issue of primary concern in current and
next generation networked systems. For example, the objective of
sensor networks is to monitor their surroundings for a variety of
different applications like atmospheric conditions, wildlife
behavior, and troop movements among others. Similarly, monitoring in
data networks is critical not only for accounting and management,
but also for detecting anomalies and attacks. Such monitoring
applications are inherently continuous and distributed, and must be
designed to minimize the communication overhead that they introduce.
In this context we introduce and study a
fundamental class of problems called ``thresholded counts'' where we
must return the {\em aggregate} frequency count of an event that is
continuously monitored by distributed nodes with a user-specified
accuracy whenever the actual count exceeds a given threshold value.

In this paper we propose to address the problem of thresholded
counts by setting local thresholds at each monitoring node and
initiating communication only when the locally observed data exceeds
these local thresholds. We explore algorithms in two categories:
{\em static thresholds} and {\em adaptive thresholds}.
In the static case, we consider
thresholds based on a linear combination of two alternate  
strategies, and show that there exists an optimal blend of the two
strategies that results in minimum communication overhead. We
further show that this optimal blend can be found using a steepest
descent search. In the adaptive case, we propose algorithms that
adjust the local thresholds based on the observed distributions of
updated information in the distributed monitoring system. We use
extensive simulations not only to verify the accuracy of our
algorithms and validate our theoretical results, but also to
evaluate the performance of the two approaches. We find that both  
approaches yield significant savings over the naive approach of
performing processing at a centralized location.}
}
@inproceedings{CormodeGarofalakisSacharidis06,
  author = {G. Cormode and M. Garofalakis and D. Sacharidis},
  att_authors = {gc2602},
  att_private = {false},
  title = {Fast Approximate Wavelet Tracking on Streams},
  booktitle = {Extending Database Technology},
  pages = {4--22},
  year = {2006},
  url = {../papers/gc-sketch.pdf},
  note = {},
  link = {http://dx.doi.org/10.1007/11687238_4},
  abstract = {
Recent years have seen growing interest in effective
algorithms for summarizing and querying massive, high-speed
data streams.
Randomized sketch synopses 
provide accurate approximations for general-purpose summaries of
the streaming data distribution
(e.g., wavelets). 
The focus of existing work has typically been on minimizing 
{\em space requirements} of the maintained synopsis 
--- however, to effectively support high-speed
data-stream analysis, a crucial practical requirement is to also
optimize: (1) the {\em update time} for incorporating a streaming
data element in the sketch, and (2) the {\em query time} for producing
an approximate summary (e.g., the top wavelet coefficients) from the
sketch.
Such time costs must be small enough to cope with rapid
stream-arrival rates and the real-time querying requirements of typical
streaming applications (e.g., ISP network monitoring).
With cheap and plentiful memory, space is often only a 
secondary concern after query/update time costs.  

In this paper, we propose the first fast solution to
the problem of tracking wavelet representations of
one-dimensional and multi-dimensional data streams, based on a
novel stream synopsis, the {\em Group-Count Sketch (GCS)}.
By imposing a hierarchical structure of groups over the data and
applying the GCS, our algorithms can quickly recover the most
important wavelet coefficients with guaranteed accuracy.
A tradeoff between query time and update time is established, by
varying the hierarchical structure of groups, allowing the right balance
to be found for specific data stream.
Experimental analysis confirms this tradeoff, and shows that all our
methods significantly outperform previously known methods 
in terms of both update time and query time, while
maintaining a high level of accuracy.}
}
@inproceedings{CormodeMuthukrishnanZhuang06,
  author = {G. Cormode and S. Muthukrishnan and W. Zhuang},
  title = {What's Different: Distributed, Continuous Monitoring
of Duplicate-Resilient Aggregates on Data Streams},
  att_authors = {gc2602},
  att_private = {false},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  pages = {20--31},
  year = 2006,
  url = {../papers/cdcdicde.pdf},
  link = {http://doi.ieeecomputersociety.org/10.1109/ICDE.2006.173},
  doi = {10.1109/ICDE.2006.173},
  note = {},
  abstract = {
Emerging applications in sensor systems and network-wide
IP traffic analysis present many technical challenges. They
need distributed monitoring and continuous tracking of events.
They have severe resource constraints not only at each site in
terms of per-update processing time and archival space for highspeed
streams of observations, but also crucially, communication
constraints for collaborating on the monitoring task. These
elements have been addressed in a series of recent works.
A fundamental issue that arises is that one cannot make the
uniqueness assumption on observed events which is present in
previous works, since widescale monitoring invariably encounters
the same events at different points. For example, within the network
of an Internet Service Provider packets of the same flow will
be observed in different routers; similarly, the same individual
will be observed by multiple mobile sensors in monitoring wild
animals. Aggregates of interest on such distributed environments
must be resilient to duplicate observations.

We study such duplicate-resilient aggregates that measure the
extent of the duplicationhow many unique observations are
there, how many observations are uniqueas well as standard
holistic aggregates such as quantiles and heavy hitters over
the unique items. We present accuracy guaranteed, highly
communication-efficient algorithms for these aggregates that
work within the time and space constraints of high speed streams.
We also present results of a detailed experimental study on both
real-life and synthetic data.}
}
@inproceedings{CormodeMuthukrishnanRozenbaum05VLDB,
  author = {G. Cormode and S. Muthukrishnan and I. Rozenbaum},
  att_authors = {gc2602},
  att_private = {false},
  title = {Summarizing and Mining Inverse Distributions on Data 
                    Streams via Dynamic Inverse Sampling},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  pages = {25--36},
  year = {2005},
  url = {../papers/inversedbnvldb.pdf},
  slides = {../slides/invdbnvldb.pdf},
  note = {},
  abstract = {
Database management systems face the challenge of
dealing with massive data distributions which
arrive at high speeds while there is only small storage
available for managing and mining them.
Emerging data stream management systems approach this
problem by summarizing and mining the distributions
using samples or sketches. However, data distributions
can be ``viewed'' in different ways. For example,
a data stream of integer values can be viewed either
as the {\em forward} distribution $f(x)$, ie., the
number of occurrences of $x$ in the
stream, or as its inverse, $f^{-1}(i)$, which is
the number of items that appear $i$ times.
While both such ``views'' are equivalent in stored
data systems, over data streams that entail approximations,
they may be significantly different. In other words,
samples and sketches developed for the forward distribution
may be ineffective for summarizing or mining the
inverse distribution. Yet, many applications such as
IP traffic monitoring naturally rely on mining inverse
distributions.

We formalize the problems of managing and mining inverse
distributions and show provable differences between summarizing
the forward distribution vs the inverse distribution.
We present methods for summarizing and mining
inverse distributions of data streams: they rely
on a novel technique to maintain a dynamic sample over the stream with
provable guarantees which can be used for variety of
summarization tasks (building quantiles or equidepth histograms)
and mining (anomaly detection: finding heavy hitters,
and measuring the number of rare items), all with provable
guarantees on quality of approximations and time/space used by
our streaming methods. We also complement our analytical
and algorithmic results by presenting an experimental
study of the methods over network data streams.}
}
@inproceedings{CormodeGarofalakis05VLDB,
  author = {G. Cormode and M. Garofalakis},
  att_authors = {gc2602},
  att_private = {false},
  title = {Sketching Streams Through the Net:
                  Distributed Approximate Query Tracking},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  pages = {13--24},
  year = {2005},
  url = {../papers/streamsnet.pdf},
  slides = {http://www.vldb2005.org/program/slides/tue/s13-cormode.ppt},
  abstract = {
While traditional database systems optimize for performance 
on one-shot query processing, emerging large-scale monitoring
applications require continuous tracking of complex data-analysis
queries over collections of physically-distributed streams. 
Thus, effective solutions have to be simultaneously space/time
efficient (at each remote monitor site), communication efficient
(across the underlying communication network), and provide
continuous, guaranteed-quality approximate query answers. 
In this paper, we propose novel algorithmic solutions for the
problem of continuously tracking a broad class of complex 
aggregate queries in such a distributed-streams setting.
Our tracking schemes maintain approximate query answers with
provable error guarantees, while simultaneously optimizing the
storage space and processing time at each remote site, as well
as the communication cost across the network.
In a nutshell, our algorithms rely on tracking general-purpose
randomized sketch summaries of local streams at remote sites
along with concise prediction models of local site behavior
in order to produce highly communication- and space/time-efficient 
solutions. 
The end result is a powerful approximate query tracking framework
that readily incorporates several complex analysis queries 
(including distributed join and multi-join aggregates, and
approximate wavelet representations), thus giving the first
known low-overhead tracking solution for such queries in the
distributed-streams model.
Experiments with real data validate our approach, revealing 
significant savings over naive solutions as well as our
analytical worst-case guarantees.}
}
@inproceedings{CormodeMuthukrishnan05PODS,
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {Space Efficient Mining of Multigraph Streams},
  pages = {271--282},
  year = {2005},
  url = {../papers/multigraph.pdf},
  slides = {../slides/multigraph-slides.pdf},
  pubsite = {http://www.sigmod.org/eproceedings/pods05/pdf/p271-cormode.pdf},
  note = {},
  abstract = {
  The challenge of monitoring massive amounts of data generated by
  communication networks has led to the interest in data stream
  processing. 
  We study streams of edges in massive communication multigraphs, defined by
  (source, destination) pairs. 
  The goal is to compute properties of the underlying graph while using small
  space (much smaller than the number of communicants), 
  and to avoid bias introduced because some edges may appear many times,
  while others are seen only once. 
  We give results for three fundamental problems on multigraph degree sequences: 
  estimating frequency moments of degrees, finding the heavy hitter
  degrees, and computing range sums of degree values. 
  In all cases we are able to show space bounds
  for our summarizing algorithms
  that are significantly smaller than storing complete information.
  We use a variety of data stream methods: 
  sketches, sampling, hashing and distinct
  counting, but a common feature is that we use the approach of
  {\em cascaded summaries}: nesting multiple estimation techniques within one
  another. 
  In our experimental study, we see that such summaries are highly
  effective, enabling massive multigraph streams to be effectively
  summarized to answer queries of interest with high accuracy using only
  a small amount of space. }
}
@inproceedings{CormodeGarofalakisMuthukrishnanRastogi05SIGMOD,
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  att_authors = {gc2602},
  att_private = {false},
  author = {G. Cormode and M. Garofalakis and S. Muthukrishnan 
                    and R. Rastogi},
  title = {Holistic Aggregates in a Networked World: 
                   Distributed Tracking of Approximate Quantiles},
  year = {2005},
  pages = {25--36},
  url = {../papers/cdquant.pdf},
  slides = {../slides/cdquant-slides.pdf},
  pubsite = {http://www.sigmod.org/eproceedings/sigmod05/pdf/p25-cormode.pdf},
  note = {},
  abstract = {
  While traditional database systems optimize for performance on
  one-shot queries, emerging large-scale monitoring applications
  require continuous tracking of  complex aggregates  and
  data-distribution summaries over collections of
  physically-distributed streams.
  Thus, effective solutions have to be simultaneously space efficient
  (at each remote site), communication efficient (across the underlying
  communication network), and provide continuous, guaranteed-quality
  estimates.
  In this paper, we propose novel algorithmic solutions
  for the problem of continuously tracking complex holistic
  aggregates in such a distributed-streams setting ---
  our primary focus is on approximate quantile summaries, but our
  approach is more broadly applicable and can handle other
  holistic-aggregate functions (e.g., ``heavy-hitters'' queries).
  We present the first known distributed-tracking schemes  for
  maintaining accurate quantile estimates  with provable
  approximation guarantees, while simultaneously optimizing
the storage space at each remote site as well as the
communication cost across the network.
In a nutshell, our algorithms employ a combination
of local tracking at remote sites and simple prediction
models for local site behavior in order to produce highly
communication- and space-efficient solutions.
We perform extensive experiments with real and synthetic data
to explore the various tradeoffs and understand the role of
prediction models in our schemes.
The results clearly validate our approach, revealing significant
savings over naive solutions as well as our analytical worst-case
guarantees.}
}
@inproceedings{CormodeMuthukrishnan05CM,
  booktitle = {{SIAM} Conference on Data Mining ({SDM})},
  att_authors = {gc2602},
  att_private = {false},
  author = {G. Cormode and S. Muthukrishnan},
  title = {Summarizing and Mining Skewed Data Streams},
  year = {2005},
  url = {../papers/cmz-sdm.pdf},
  slides = {../slides/cmz-slides.pdf},
  note = {},
  abstract = {
Many applications generate massive data streams.
Summarizing such massive data requires fast, small space algorithms to
support post-hoc queries and mining. 
An important observation is that such streams are rarely uniform, and
real data sources typically exhibit significant skewness. 
These are well modeled by Zipf distributions, which are characterized
by a parameter, $z$, that captures the amount of skew.

We present a data stream summary that can answer point queries with 
$\epsilon$ accuracy and show that the space needed is only
$O(\epsilon^{-\min\{1,1/z\}})$. 
This is the first $o(1/\epsilon)$ space algorithm for this problem,
and we show it is essentially tight for skewed distributions. 
We show that the same data structure can also estimate the $L_2$ norm
of the stream in $o(1/\epsilon^2)$ space for $z>1/2$, another
improvement over the existing $\Omega(1/\epsilon^2)$ methods. 

We support our theoretical results with an experimental study over a
large variety of real and synthetic data. 
We show that significant skew is present in both textual and
telecommunication data.
Our methods give strong accuracy, significantly better than other
methods, and behave exactly in line with their analytic bounds.}
}
@inproceedings{CormodeKornMuthukrishnanSrivastava05,
  author = {G. Cormode and F. Korn and S. Muthukrishnan and D. 
                  Srivastava},
  att_authors = {gc2602,ds8961,pk1785},
  att_private = {false},
  title = {Effective Computation of Biased Quantiles over Data 
                    Streams},
  year = {2005},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  pages = {20--31},
  url = {../papers/bquant-icde.pdf},
  slides = {../slides/bquant-slides.pdf},
  note = {},
  abstract = {
Skew is prevalent in many data sources such as IP traffic
streams.  To continually summarize the 
distribution of such data, a high-biased set of quantiles
(e.g., 50th, 90th and 99th percentiles) with finer error
guarantees at higher ranks (e.g., errors of 5, 1 and 0.1
percent, respectively) is more useful than uniformly 
distributed quantiles (e.g., 25th, 50th and 75th percentiles)
with uniform error guarantees.
In this paper, we address the following two problems.  
First, can we compute quantiles with finer 
error guarantees for the higher ranks of the data distribution 
effectively, using less space and computation time than 
computing all quantiles uniformly at the finest error?
Second, if {\em specific\/} quantiles and their error bounds 
are requested {\em a priori}, can the necessary space usage and 
computation time be reduced?

We answer both questions in the affirmative by formalizing them as the
``high-biased'' quantiles and the ``targeted'' quantiles problems,
respectively, and presenting algorithms with provable guarantees, that 
perform
significantly better than previously known solutions for these problems.  
We implemented our algorithms in the Gigascope data stream management
system, and evaluated alternate approaches for maintaining the
relevant summary structures.  Our experimental results on real and
synthetic IP data streams complement our theoretical analyses, and
highlight the importance of lightweight, non-blocking implementations
when maintaining summary structures over high-speed data streams.}
}
@inproceedings{CormodeMuthukrishnan05,
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {Substring Compression Problems},
  year = {2005},
  pages = {321--330},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  url = {../papers/substringsoda.pdf},
  slides = {../slides/substring-slides.pdf},
  note = {},
  abstract = {
We initiate a new class of string matching problems called 
{\em  Substring Compression Problems}.
Given a string $S$ that may be preprocessed, 
the problem is to quickly find the
compressed representation or the compressed size of any query substring of 
$S$ 
(Substring Compression Query or SCQ) or to find the length $l$ 
substring of $S$ 
whose compressed size is the least (Least Compressible Substring or LCS 
problem). 

Starting from the seminal paper of Lempel and Ziv over 25 years ago,
many different methods have emerged for compressing entire strings.
Determining substring compressibility is a natural variant that is
combinatorially and algorithmically challenging, yet surprisingly
has not been studied before. In addition, compressibility of strings
is emerging as a tool to compare biological sequences and analyze their
information content. However, typically, the compressibility 
of the entire sequence is not as informative as that of portions of
the sequences. Thus substring compressibility may be a more suitable 
basis for sequence analysis.

We present the first known, nearly optimal algorithms for substring 
compression problems---SCQ, LCS and their generaliztions---that 
are exact or provably approximate. 
Our exact algorithms exploit the structure in strings via suffix trees  
and 
our approximate algorithms rely on new relationships we find between 
Lempel-Ziv compression and string parsings. }
}
@inproceedings{CormodeKornMuthukrishnanSrivastava04,
  author = {G. Cormode and F. Korn and S. Muthukrishnan and D. Srivastava},
  att_authors = {gc2602,pk1785,ds8961},
  att_private = {false},
  year = {2004},
  title = {Diamond in the Rough: Finding Hierarchical Heavy Hitters in Multi-Dimensional Data},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  pages = {155--166},
  url = {../papers/ckmsh4.pdf},
  pubsite = {http://portal.acm.org/citation.cfm?id=1007588&coll=Portal&dl=ACM&CFID=40572378&CFTOKEN=78576472},
  note = {},
  abstract = {
  Data items archived in data warehouses or those that arrive online as
streams typically have attributes which take values from multiple
hierarchies (e.g., time and geographic location; source and
destination IP addresses). Providing an aggregate view of such data is
important to summarize, visualize, and analyze. We develop the
aggregate view based on certain hierarchically organized sets of
large-valued regions (``heavy hitters''). Such Hierarchical Heavy
Hitters (HHHs) were previously introduced as a crucial aggregation
technique in one dimension. In order to analyze the wider range of
data warehousing applications and realistic IP data streams, we
generalize this problem to multiple dimensions.

We illustrate and study two variants of HHHs for multi-dimensional
data. In particular, we identify ``overlap'' and ``split'' variants,
depending on how an aggregate computed for a child node in the
multi-dimensional hierarchy is propagated to its parent
element(s). For data warehousing applications, we present offline
algorithms that take multiple passes over the data and produce the
exact HHHs.  For data stream applications, we present online
algorithms that find approximate HHHs in one pass, with proven
accuracy guarantees.

We show experimentally, using real and synthetic data, that our
proposed online algorithms yield outputs which are very similar
(virtually identical, in many cases) to their offline
counterparts. The lattice property of the product of hierarchical
dimensions (``diamond'') is crucially exploited in our online
algorithms to track approximate HHHs using only a small, fixed number
of statistics per candidate node, regardless of the number of
dimensions.}
}
@inproceedings{CormodeKornMuthukrishnanJohnsonSpatscheckSrivastava04,
  author = {G. Cormode and F. Korn and S. Muthukrishnan and T. Johnson and O. Spatscheck and D. Srivastava},
  year = {2004},
  title = {Holistic {UDAF}s at streaming speeds},
  pages = {35--46},
  url = {../papers/cjkmss-streamudaf.pdf},
  pubsite = {http://portal.acm.org/citation.cfm?id=1007575&coll=Portal&dl=ACM&CFID=40572378&CFTOKEN=78576472},
  note = {},
  booktitle = {{ACM} {SIGMOD} International 
                    Conference on Management of Data ({SIGMOD})},
  abstract = {
Many algorithms have been proposed to approximate holistic ag-
gregates, such as quantiles and heavy hitters, over data streams.
However, little work has been done to explore what techniques are
required to incorporate these algorithms in a data stream query pro-
cessor, and to make them useful in practice.
In this paper, we study the performance implications of using
user-dened aggregate functions (UDAFs) to incorporate selection-
based and sketch-based algorithms for holistic aggregates into a
data stream management system's query processing architecture.
We identify key performance bottlenecks and tradeoffs, and pro-
pose novel techniques to make these holistic UDAFs fast and space-
efcient for use in high-speed data stream applications. We evalu-
ate performance using generated and actual IP packet data, focus-
ing on approximating quantiles and heavy hitters. The best of our
current implementations can process streaming queries at OC48
speeds (2x 2.4Gbps).}
}
@inproceedings{Cormode04,
  author = {G. Cormode},
  att_authors = {gc2602},
  att_private = {false},
  title = {The Hardness of the Lemmings Game, or {Oh} no, more {NP}-Completeness Proofs},
  url = {../papers/cormodelemmings.pdf},
  link = {Cormode04LemsTR.html},
  pages = {65--76},
  booktitle = {Proceedings of Third International Conference on Fun with Algorithms},
  year = {2004},
  abstract = {
In the computer game `Lemmings', the player must guide a tribe of green-haired
lemming creatures to safety, and save them from an untimely demise. 
We formulate the decision problem which is, given a level of the game, to
decide whether it is possible to complete the level (and hence to find 
a solution to that level).  
Under certain limitations, this can be decided in polynomial
time, but in general the problem is shown to be NP-Hard.  
This can hold even if there
is only a single lemming to save, thus showing that it is hard to
approximate the number of saved lemmings to any factor.}
}
@inproceedings{CormodeCzumajMuthukrishnan04,
  author = {G. Cormode and A. Czumaj and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {How to {\em increase} the acceptance ratios of top conferences},
  url = {http://www.cs.rutgers.edu/~muthu/ccmfun.pdf},
  booktitle = {Proceedings of Third International Conference on Fun with Algorithms},
  year = {2004},
  pages = {262--273},
  link = {CormodeCzumajMuthukrishnan04TR.html},
  abstract = {
 In the beginning was the pub. This work was triggered by a pub conversation where the
authors observed that many resumes list acceptance ratios of conferences where their
papers appear, boasting the low acceptance ratio. The lower the ratio, better your paper
looks. The list might look equally impressive if one listed the rejection ratio of
conferences where ones paper was submitted and rejected. We decided to lampoon
rather than lament the effort the PC typically put in: wouldn't the world be better if we
could encourage only high quality submissions and so run top conferences with very
high acceptance ratios? This paper captures our thoughts, and it is best consumed in a
pub (and in color).}
}
@inproceedings{CormodeMuthukrishnan04Infocom,
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {What's New: Finding Significant Differences in Network
                     Data Streams},
  booktitle = {Proceedings of {IEEE} {Infocom}},
  note = {},
  year = {2004},
  pages = {1534--1545},
  url = {../papers/whatsnew.pdf},
  link = {CormodeMuthukrishnan05ToN.html},
  pubsite = {http://www.ieee-infocom.org/2004/Papers/33_1.PDF},
  slides = {../slides/changes-infocom.pdf},
  abstract = {Monitoring and analyzing network traffic usage patterns 
is vital for managing IP Networks. An important problem is to provide 
network managers with information about changes in traffic, informing them 
about ``what's new''. Specifically, we focus on the challenge of finding 
significantly large differences in traffic: over time, between interfaces 
and between routers. We introduce the idea of a deltoid: an item that has 
a large difference, whether the difference is absolute, relative or 
variational.

We present novel algorithms for finding the most significant deltoids in 
high speed traffic data, and prove that they use small space, very small 
time per update, and are guaranteed to find significant deltoids with 
pre-specified accuracy. In experimental evaluation with real network 
traffic, our algorithms perform well and recover almost all deltoids. This 
is the first work to provide solutions capable of working over the data 
with one pass, at network traffic speeds.}
}
@inproceedings{CormodeMuthukrishnan04CMLatin,
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {An Improved Data Stream Summary: The Count-Min Sketch 
                     and its Applications},
  year = {2004},
  booktitle = {Proceedings of Latin American Theoretical Informatics (LATIN)},
  pages = {29-38},
  url = {../papers/cm-latin.pdf},
  note = {},
  slides = {../slides/cmlatintalk.pdf},
  link = {CormodeMuthukrishnan04CMJalg.html},
  pubsite = {http://springerlink.metapress.com/openurl.asp?genre=article&issn=0302-9743&volume=2976&spage=29},
  abstract = {We introduce a new sublinear space data structure---the Count-Min Sketch--- for summarizing data streams. Our sketch allows fundamental queries in data stream summarization such as point, range, and inner product queries to be approximately answered very quickly; in addition, it can be applied to solve several important problems in data streams such as finding quantiles, frequent items, etc. The time and space bounds we show for using the CM sketch to solve these problems significantly improve those previously known --- typically from $1/\epsilon^2$ to $1/\epsilon$ in factor.

Journal version appeared in Journal of Algorithms.
}
}
@inproceedings{Cormode03Stable,
  author = {G. Cormode},
  att_authors = {gc2602},
  att_private = {false},
  title = {Stable distributions for stream computations: 
                    it's as easy as 0,1,2},
  booktitle = {Workshop on Management and Processing of Massive Data
                     Streams at FCRC},
  year = {2003},
  url = {../papers/stable.pdf},
  slides = {../slides/mpds-stable.pdf},
  abstract = {A surprising number of data stream problems are solved by methods involving computations with stable distributions. This paper will give a short summary of some of these problems, and how the best known solutions depend on use of stable distributions; it also lists some related open problems.}
}
@inproceedings{CormodeMuthukrishnan03PODS,
  title = {What's Hot and What's Not: Tracking Most Frequent Items
                Dynamically},
  att_authors = {gc2602},
  att_private = {false},
  year = {2003},
  pages = {296-306},
  author = {G. Cormode and S. Muthukrishnan},
  booktitle = {{ACM} Principles of Database Systems ({PODS})},
  note = {},
  url = {../papers/hotitems.pdf},
  slides = {../slides/whatshot-pods.pdf},
  pubsite = {http://doi.acm.org/10.1145/773153.773182},
  abstract = {Most database management systems maintain statistics on the underlying relation. One of the important statistics is that of the 'hot items' in the relation: those that appear many times (most frequently, or more than some threshold). For example, end-biased histograms keep the hot items as part of the histogram and are used in selectivity estimation. Hot items are used as simple outliers in data mining, and in anomaly detection in networking applications. We present a new algorithm for dynamically determining the hot items at any time in the relation that is undergoing deletion operations as well as inserts. Our algorithm maintains a small space data structure that monitors the transactions on the relation, and when required, quickly outputs all hot items, without rescanning the relation in the database. With user-specified probability, it is able to report all hot items. Our algorithm relies on the idea of ``group testing'', is simple to implement, and has provable quality, space and time guarantees. Previously known algorithms for this problem that make similar quality and performance guarantees can not handle deletions, and those that handle deletions can not make similar guarantees without rescanning the database. Our experiments with real and synthetic data shows that our algorithm is remarkably accurate in dynamically tracking the hot items independent of the rate of insertions and deletions.}
}
@inproceedings{CormodeKornMuthukrishnanSrivastava03,
  title = {Finding Hierarchical Heavy Hitters in Data Streams},
  att_authors = {gc2602,ds8961,pk1785},
  att_private = {false},
  year = {2003},
  pages = {464-475},
  author = {G. Cormode and F. Korn and S. Muthukrishnan and D. Srivastava},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  url = {../papers/ckms-hhh.pdf},
  note = {},
  link = {http://www.vldb.org/conf/2003/papers/S15P01.pdf},
  abstract = {Aggregation along hierarchies is a critical summary technique in a large variety of online applications including decision support, and network management (e.g., IP clustering, denial-of-service attack monitoring). Despite the amount of recent study that has been dedicated to online aggregation on sets (e.g., quantiles, hot items), surprisingly little attention has been paid to summarizing hierarchical structure in stream data.

The problem we study in this paper is that of finding Hierarchical Heavy Hitters (HHH): given a hierarchy and a fraction phi, we want to find all HHH nodes that have a total number of descendants in the data stream larger than phi of the total number of elements in the data stream, after discounting the descendant nodes that are HHH nodes. The resulting summary gives a topological 'cartogram' of the hierarchical data. We present deterministic and randomized algorithms for finding HHHs, which builds upon existing techniques by incorporating the hierarchy into the algorithms. Our experiments demonstrate several factors of improvement in accuracy over the straightforward approach, which is due to making algorithms hierarchy-aware}
}
@inproceedings{CormodeMuthukrishnan03DN,
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  year = {2003},
  booktitle = {European Symposium on Algorithms},
  title = {Estimating Dominance Norms of Multiple Data Streams},
  series = {LNCS},
  note = {},
  volume = {2838},
  url = {../papers/dominance.pdf},
  slides = {../slides/esadom.pdf},
  link = {CormodeMuthukrishnan02DomTR.html},
  pubsite = {http://springerlink.metapress.com/openurl.asp?genre=article&issn=0302-9743&volume=2832&spage=148},
  abstract = {
There is much focus in the algorithms and database communities on
designing tools to manage and mine data streams.
Typically, data streams consist of multiple signals.
Formally, a stream of multiple signals is $(i,a_{i,j})$
where $i$'s correspond to the domain, $j$'s index the different
signals and $a_{i,j}\geq 0$ give the value of the $j$th signal at
point $i$.
We study the problem of finding norms that are cumulative of
the multiple signals in the data stream.

For example, consider the max-dominance norm,
defined as $\Sigma_i \max_j \{a_{i,j}\}$.
It may be thought as estimating the norm of the  ``upper
envelope'' of the multiple signals, or alternatively, as
estimating the norm of the ``marginal'' distribution of tabular data streams.
It is used in applications to estimate the ``worst case influence'' of
multiple processes, for example in IP traffic analysis, electrical
grid monitoring and
financial domain.
In addition, it is a natural measure, generalizing the union of data
streams or counting distinct elements in data streams.

We present the first known data stream algorithms for estimating
max-dominance of multiple signals.
In particular, we use workspace and time-per-item that are both sublinear (in
fact, poly-logarithmic)  in the input size.  In contrast
other
notions of dominance on streams $a$, $b$ --- min-dominance
($\Sigma_i \min_j \{a_{i,j}\}$),
count-dominance ($|\{i | a_i > b_i\}|$) or relative-dominance
($\Sigma_i a_i/\max\{1,b_i\}$ ) --- are all impossible to estimate
accurately with sublinear space.}
}
@inproceedings{CormodeDatarIndykMuthukrishnan02,
  author = {G. Cormode and M. Datar and P. Indyk and 
                   S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {Comparing Data Streams Using {H}amming Norms},
  year = {2002},
  pages = {335--345},
  booktitle = {International Conference on 
                  Very Large Data Bases ({VLDB})},
  note = {},
  url = {../papers/hammingnorm.pdf},
  slides = {../slides/vldb02.pdf},
  link = {CormodeDatarIndykMuthukrishnan03.html},
  pubsite = {http://www.vldb.org/conf/2002/S10P02.pdf},
  abstract = {
Massive data streams are now fundamental to many data processing
applications.  For example, Internet routers produce large scale
diagnostic data streams.  Such streams are rarely stored in
traditional databases, and instead must be processed ``on the fly'' as
they are produced.  Similarly, sensor networks produce multiple data
streams of observations from their sensors.  There is growing focus on
manipulating  data streams, and hence, there is a need to identify
basic operations of interest in managing data streams, and to support
them efficiently.
We propose computation of the Hamming norm as a basic operation of
interest.  The Hamming norm formalizes ideas
that are used throughout data processing.
When applied to a single stream, the Hamming norm gives the number of distinct 
items that are
present in that data stream, which is a statistic of great interest in
databases.  When applied to a pair of streams, the Hamming norm gives
an important measure of (dis)similarity:  the number of unequal item
counts in the two streams. Hamming norms have many uses
in comparing data streams.

We present a novel approximation technique for estimating the Hamming norm
for massive data streams; this relies on what we call the ``$l_0$
{\it sketch}''
and we prove its accuracy. We test our approximation method on a large
quantity of synthetic and real stream data, and show that the
estimation is accurate to within a few percentage points. 

  Journal version 
                 in {\em {IEEE} Transactions on Knowledge and Data
                 Engineering} 15(3):529--541, 2003}
}
@inproceedings{CormodeIndykKoudasMuthukrishnan02,
  author = {G. Cormode and P. Indyk and N. Koudas and 
                  S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {Fast Mining of Tabular Data via 
                  Approximate Distance Computations},
  year = {2002},
  pages = {605--616},
  booktitle = {International Conference on Data Engineering ({ICDE})},
  note = {},
  url = {../papers/tabular.pdf},
  slides = {../slides/icdeslides.pdf},
  pubsite = {http://csdl.computer.org/comp/proceedings/icde/2002/1531/00/15310605abs.htm},
  abstract = {
Tabular data abound in many  data  stores: traditional
relational databases store tables, and
new applications also generate massive tabular
datasets.
For example, consider the geographic distribution of cell
phone traffic at different  base stations across the
country
(which can be  represented by a table indexed by
  latitude and longitude),
or the evolution of traffic at Internet
routers over time
(which can be represented by a table
indexed by time, source and destination IP addresses).
Detecting similarity patterns in such data
sets (e.g., which geographic regions have similar cell phone
usage  distribution,  which  IP subnet traffic distributions
over time intervals are similar,  etc) is of great importance.
Identification   of  such  patterns  poses  many
conceptual challenges (what is a  suitable  similarity
distance  function  for two ``regions'') as well as
technical challenges (how to perform similarity computations
efficiently as massive tables get accumulated over time) that
we address.

We  present methods  for  determining
similar regions in massive tabular data.
Our methods are for computing the ``distance'' between
any two subregions of a tabular data: they are approximate,
but  highly  accurate as we prove mathematically, and they are
fast, running in time nearly linear in the table size. Our methods
are general since these distance computations can be
applied to any  mining or
similarity   algorithms that use $L_p$ norms.
A novelty of our distance computation procedures is that they
work for any $L_p$ norms --- not only the traditional
$p=2$  or  $p=1$,  but for all $p\leq 2$; the choice of $p$,
say {\em fractional} $p$,
provides an interesting alternative similarity behavior!
We use  our algorithms in  a  detailed
experimental study of the clustering patterns in  real  tabular
data obtained  from  one  of  AT\&T's  data stores and show that our
methods  are  substantially  faster   than   straightforward
methods  while remaining highly accurate, and able to detect
interesting patterns by varying the value of $p$.}
}
@inproceedings{CormodeMuthukrishnan02,
  author = {G. Cormode and S. Muthukrishnan},
  att_authors = {gc2602},
  att_private = {false},
  title = {The String Edit Distance Matching Problem with Moves},
  year = {2002},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  pages = {667--676},
  note = {},
  url = {../papers/editmoves.pdf},
  slides = {../slides/soda2k2.pdf},
  link = {CormodeMuthukrishnan01EditTR.html},
  pubsite = {http://doi.acm.org/10.1145/545381.545470},
  abstract = {
The edit distance between two strings $S$ and $R$
is defined to be the minimum number of character inserts,
deletes and changes needed to convert $R$ to $S$.
Given a text string $t$ of length $n$, and a pattern string $p$
of length $m$, informally, the string edit distance matching problem is to
compute the smallest edit distance between $p$ and
substrings of $t$. A well known dynamic programming algorithm
takes time $O(nm)$ to solve this problem, and it is an important
open problem in Combinatorial Pattern Matching to
significantly improve this bound.

We relax the problem
so that (a) we allow an additional operation, namely,
{\em substring moves}, and (b) we approximate the string edit distance
upto a factor of $O(\log n \log^* n)$. ($\log^* n$ is the
number of times $\log$ function is applied to $n$ to produce a constant.)
Our result is a near linear time deterministic
algorithm for this version of the problem.
This is the first
known significantly subquadratic algorithm for a string
edit distance problem in which the distance involves
nontrivial alignments.
Our results are obtained by embedding strings into $L_1$
vector space using a simplified parsing technique we call
{\em Edit Sensitive Parsing} (ESP). This embedding is
approximately distance preserving, and we show many applications
of this embedding to string proximity problems including nearest neighbors,
outliers, and streaming computations with strings.
}
}
@inproceedings{CormodeMuthukrishnanSahinalp01,
  author = {G. Cormode and S Muthukrishnan 
                   and S. C. {{S}}ahinalp},
  att_authors = {gc2602},
  att_private = {false},
  title = {Permutation Editing and Matching via Embeddings},
  booktitle = {International Colloquium on Automata, 
                  Languages and Programming ({ICALP})},
  note = {},
  volume = {2076},
  year = {2001},
  pages = {481--492},
  url = {../papers/cms-perm.pdf},
  slides = {../slides/icalp.pdf},
  pubsite = {http://link.springer.de/link/service/series/0558/bibs/2076/20760481.htm},
  abstract = {
If the genetic maps of two species are modelled as permutations of
 (homologous) genes, the number of chromosomal rearrangements
 in the form of deletions, block moves, inversions etc. to transform one
 such permutation to another can be used as a measure of their evolutionary
 distance. Motivated by such scenarios in Computational Biology and
elsewhere, we study problems of
computing distances between permutations
 as well as matching permutations in
 sequences, and finding most similar permutation from a collection
 (``nearest neighbor'').

We adopt a
 general approach: embed permutation distances of relevance into well-known
 vector spaces in an approximately distance-preserving manner,
 and solve the resulting problems on the well-known spaces.
Our results are as follows:
\begin{itemize}
\item
We present the first known approximately distance preserving
embeddings of these permutation distances into well-known spaces.
\item
Using these embeddings, we obtain several results, including
the first known efficient
 solution for approximately solving nearest neighbor
 problems with permutations
 and
the first known
 algorithms for finding permutation distances in the ``data stream''
 model.
\item
We consider a novel
 class of problems called {\em permutation matching} problems
which are similar to string matching problems,
except that the
pattern is a permutation (rather than a string)
and present linear or near-linear time
algorithms for approximately solving permutation matching problems;
in contrast, the corresponding string problems take significantly
longer.
}
}
@inproceedings{CormodePatersonSahinalpVishkin00,
  author = {G. Cormode and M. Paterson and S. C.
                   {{S}}ahinalp and U. Vishkin},
  att_authors = {gc2602},
  att_private = {false},
  title = {Communication Complexity of Document Exchange},
  booktitle = {{ACM}-{SIAM} Symposium on Discrete Algorithms ({SODA})},
  note = {},
  year = {2000},
  pages = {197--206},
  url = {../papers/docexchange.pdf},
  link = {http://doi.acm.org/10.1145/338219.338252},
  slides = {../slides/soda.pdf},
  abstract = {
We have two users, $A$ and $B$, who hold documents $x$ and $y$ respectively.
Neither of the users has any information about the other's document.
They exchange messages
 so that $B$ computes $x$; it may be required that $A$ compute
 $y$ as well.
Our goal is to design communication protocols with
 the main objective of minimizing the total number of bits they exchange;
 other objectives are
 minimizing the number of rounds and the complexity of internal
 computations.
An important notion which determines the efficiency of the
 protocols is how one measures the distance between $x$ and $y$.
We consider several metrics for measuring this distance,
 namely the Hamming metric, the Levenshtein metric (edit distance),
 and a new LZ metric, which is introduced in this paper.
We show how to estimate the distance between $x$ and $y$ using a
 single message of logarithmic size.
For each metric, we present the first
 communication-efficient protocols, which often
 match the corresponding lower bounds.
A consequence of these are error-correcting codes for these error
 models which correct up to $d$ errors in $n$ characters using
 $O(d \log n)$ bits.
Our most interesting methods use a new
 {\em histogram transformation} that we introduce to convert
 edit distance to $L_1$ distance.}
}
@inproceedings{OzsoyogluCormodeBalkirOzsoyoglu00,
  author = {G. Ozsoyoglu and N. H. Balkir and
                  G. Cormode and  Z. M. Ozsoyoglu},
  att_authors = {gc2602},
  att_private = {false},
  title = {Electronic Books in Digital Libraries},
  year = {2000},
  booktitle = {Proceedings of IEEE Advances in Digital Libraries (ADL)},
  note = {},
  pages = {5--14},
  url = {../papers/electronicbooks.pdf},
  link = {OzsoyogluCormodeBalkirOzsoyoglu04.html},
  pubsite = {http://csdl.computer.org/comp/proceedings/adl/2000/0659/00/0659toc.htm},
  abstract = {
Electronic Book is an application with a multimedia database of instructional resources, which include hyperlinked text, instructor's audio/video clips, slides, animation, still images, etc. as well as content-based infomration about these data, and metadata such as annotations, tags, and cross-referencing information. Electronic books in the Internet or on CDs today are not easy to learn from. We propose the use of a multimedia database of instructional resources in constructing and delivering multimedia lessons about topics in an electronic book.

We introduce an electronic book data model containing (a) topic objects and and (b) instructional resources, called instruction modules which are multimedia presentations possibly capturing real-life lectures of instructors. We use the notion of topic prerequisites for topics at different detail levels, to allow electronic book users to request / compose multimedia lessons about topics on the electronic book. We present automated construction of the ''best'' user-tailored lesson (as a multimedia presentation). }
}

This file was generated by bibtex2html 1.92.