"Big data is ubiquitous but heterogeneous. Big data can be used to tally clicks and traffic on web pages, find patterns in stock trades, track consumer preferences, identify linguistic correlations in large corpuses of texts. This book examines big data not as an undifferentiated whole but contextually, investigating the varied challenges posed by big data for health, science, law, commerce, an…
We live in the era of Big Data, with storage and transmission capacity measured not just in terabytes but in petabytes (where peta- denotes a quadrillion, or a thousand trillion). Data collection is constant and even insidious, with every click and every "like" stored somewhere for something. This book reminds us that data is anything but "raw," that we shouldn't think of data as a natural reso…
In this book, the author argues that in embarking on an unprecedented effort to build surveillance capabilities deeply into communications infrastructure, the U.S. government is opting for short-term security and creating dangerous long-term risks. Landau describes what makes communications security hard, warrantless wiretapping and the role of electronic surveillance in the war on terror, the …
A concise introduction to fundamental methods for finding and extracting relevant information from the ever-increasing amounts of biomedical text available.OCLC-licensed vendor bibliographic record.
Collected papers based on talks presented at two Neural Information Processing Systems workshops.State-of-the-art algorithms and theory in a novel domain of machine learning, prediction when the output has structure.Machine learning develops intelligent computer systems that are able to generalize from previously seen examples. A new domain of machine learning, in which the prediction must sati…
Pervasive and networked computers have dramatically reduced the cost of collecting and distributing large datasets. In this context, machine learning algorithms that scale poorly could simply become irrelevant. We need learning algorithms that scale linearly with the volume of the data while maintaining enough statistical efficiency to outperform algorithms that simply process a random subset o…
Borders in Cyberspace investigates issues arising from national differences in law, public policy, and social and cultural values in light of the emerging global information infrastructure. The contributions include detailed analyses of some of the most visible issues, including intellectual property, security, privacy, and censorship.OCLC-licensed vendor bibliographic record.
"A Bradford book."OCLC-licensed vendor bibliographic record.
A guide for assessing an organization's data quality practice and a roadmap for implementing a viable data and information quality management program, based on rigorous research and drawing on real-world examples.OCLC-licensed vendor bibliographic record.
"When an application is built, an underlying data model is chosen to make that application effective. Frequently, other applications need the same data, only modeled differently. The solution of copying the underlying data and modeling is costly in terms of storage and makes data maintenance and evolution impossible. View mechanisms are a technique to model data differently for various applicat…