Workshop On Algorithms For Modern Massive Datasets

2y ago
14 Views
2 Downloads
929.80 KB
38 Pages
Last View : 17d ago
Last Download : 3m ago
Upload by : Pierre Damon
Transcription

Workshop onAlgorithms for ModernMassive DatasetsStanford University and Yahoo! ResearchThe MMDS (modern massive datasets) workshop provides aforum for discussions on massive, high-dimensional, andnonlinear-structured data between computer scientists,computational and applied mathematicians, statisticians, andpractitioners to promote cross-fertilization of ideas.June 21-24, 2006Stanford, CA

2

Wednesday: June 21, 2006Linear Algebra 00-11:00Tutorial: Ravi KannanSampling in large matrices11:00-11:30Santosh VempalaSampling methods for low rank approximation11:30-12:00Petros DrineasSubspace sampling and relative error matrix approximation12:00-1:30Lunch (on your own)1:30-2:30Tutorial:D ianne O ’LearyMatrix factorizations for information retrieval2:30-3:00Pete StewartSparse reduced rank approximations to sparse matrices3:00-3:30Haesun ParkAdaptive discriminant analysis by regularized minimum squared errors3:30-4:00Break4:00-4:30Michael MahoneyCUR matrix decompositions for improved data analysis4:30-5:00Daniel SpielmanFast algorithms for graph partitioning, sparsification, and solving SDD systems5:00-5:30Anna Gilbert/Martin StraussList decoding of noisy Reed-Muller-like codes5:30-6:00Bob PlemmonsLow-rank nonnegative factorizations for spectral imaging applications6:00-6:30Art OwenA hybrid of multivariate regression and factor analysis6:30-8:00Reception (at New Guinea Garden)3

Thursday: June 22, 2006Industrial Applications and Sampling MethodsTimeEvent9:00-10:00Tutorial: Prabhakar RaghavanThe changing face of web search10:00-10:30Tong ZhangStatistical ranking problem10:30-11:00Break11:00-11:30Michael BerryText mining approaches for email surveillance11:30-12:00Hongyaun ZhaIncorporating query difference for learning retrieval functions12:00-12:30Trevor Hastie/Ping LiEfficient L2 and L1 dimension reduction in massive databases12:30-2:00Lunch (on your own)2:00-3:00Tutorial: Muthu MuthukrishnanAn algorithm icist’s look at com pressed sensing problem s3:00-3:30Inderjit DhillonKernel learning with Bregman matrix divergences3:30-4:00Bruce HendricksonLatent semantic analysis and Fiedler retrieval4:00-4:30Break4:30-5:00Piotr IndykNear-optimal hashing for approximate nearest neighbor problems5:00-5:30Moses CharikarCompact representations for data5:30-6:00Sudipto GuhaAt the confluence of streams, order, information, and signals6:00-6:30Frank McSherryPreserving privacy in large-scale data analysis4

Friday: June 23, 2006Kernel Learning and ApplicationsTimeEvent9:00-10:00Tutorial: Dimitris AchlioptasApplications of random matrices in spectral computations and machine learning10:00-10:30Tomaso PoggioLearning from data: Theory, engineering applications and (a bit of) neuroscience10:30-11:00Break11:00-11:30Stephen SmaleGeometry and topology of data11:30-12:00Gunnar CarlssonAlgebraic topology and analysis of high dimensional data12:00-12:30Vin de SilvaPoint-cloud topology via harmonic forms12:30-2:00Lunch (on your own)2:00-2:30Dan BoleyFast clustering leads to fast support vector machine training and more2:30-3:00Chris DingOn the equivalence of (semi-)nonnegative matrix factorization and k-means3:00-3:30Al InselbergParallel coordinates: visualization & data mining for high dimensional datasets3:30-4:00Joel TroppOne sketch for all: a sublinear approximation scheme for heavy hitters4:00-4:30Break4:30-5:00David DonohoNeedles in haystacks: Finding relevant variables among many useless ones5:00-5:30Rob TibshiraniPrediction by supervised principal components5:30-6:00Tao Yang/Apostolos GerasoulisPage ranking for large-scale internet search:Ask.com ’s experiences6:00-8:00Poster Session/Banquet (at Wallenberg Hall)5

Saturday: June 24, 2006Tensor-Based Data 0Tutorial: Lek-Heng LimTensors, symmetric tensors and nonnegative tensors in data analysis11:00-11:30Eugene TyrtyshnikovTensor compression of petabyte-size data11:30-12:00Lieven De LathauwerThe decomposition of a tensor in a sum of rank (R1, R2, R3) terms12:00-1:30Lunch1:30-2:00Orly AlterMatrix and tensor computations for reconstructing cellular pathways2:00-2:30Shmuel FriedlandTensors: Ranks and approximations2:30-3:00Tammy KoldaMultilinear algebra for analyzing data with multiple linkages3:00-3:30Lars EldénComputing the best rank-(R1, R2, R3) approximation of a tensor3:30-4:00Break4:00-4:30Liqun QiEigenvalues of tensors and their applications4:30-5:00Brett BaderAnalysis of latent relationships in semantic graphs using DEDICOM5:00-5:30Alex VasilescuMultilinear (tensor) algebraic framework for computer vision and graphics5:30-6:00Rasmus BroMulti-way analysis of bioinformatic data6:00-6:30Pierre ComonIndependent component analysis viewed as a tensor decomposition6:30-8:00Closing Reception6

Wednesday: June 21, 2006 (Linear Algebra Basics)to compute such an approximation in time linear inthe number of non-zero entries of the matrix.10:00-11:00SAMPLING IN LARGE MATRICESIn this talk, we discuss this approach and presenttwo ways of generalizing it, called adaptive samplingand volume sampling. Together, they show that asample of O(k/c k log k) rows of any matrix cangenerate a rank-k matrix whose error is at most (1 c)times that of the optimum rank-k matrix. Thealgorithm based on this computes such amultiplicative low-rank approximation, also in lineartime.In many application areas, the input data matrix istoo large to be handled by traditional Linear Algebraalgorithms. A number of recent results address this.One proves that a small random sample of columnsand a small random sample of rows are sufficient toapproximate any matrix provided the samplingprobabilities are judiciously chosen. Also, from agood low-rank approximation (LRA) to the sampledsub-matrices, one can derive a good LRA to thewhole matrix. These approximations are suitable fora host of numerical applications which go under thename of Principal Component Analysis.Santosh VempalaDepartment of MathematicsMassachusetts Institute of Technology11:30-12:00We also discuss applications of these methods to abroad class of combinatorial problems of which atypical one is to maximize a low-degree n-variablepolynomial over the corners of the unit n-cube. Wedescribe an efficient algorithm for finding a roughanalog of LRA to a tensor which then helps usestimate the maximum.SUBSPACE SAMPLING AND RELATIVEERROR MATRIX APPROXIMATIONIn this talk we will focus on low-rank matrixdecompositions that are explicitly expressed interms of a small number of actual columns and/orrows of a matrix, as opposed to, e.g., linearcombinations of up to all the columns and/or rowsof the matrix, such as provided by truncating thesingular value decomposition (SVD). Motivations forstudying such matrix decompositions include (i) theefficient decomposition of large low-rank matricesthat posses additional structure such as sparsity ornon-negativity, (ii) expressing Gram matrices instatistical learning theory in terms of a small numberof actual data points, (iii) the improved datainterpretability that such decompositions provide fordatasets in the Internet domain, the social sciences,biology, chemistry, medicine, etc., and (iv) ons in space-constrained settings.Ravi KannanDepartment of Computer ScienceYale University11:00-11:30SAMPLING METHODS FOR LOW RANKAPPROXIMATIONIt is well-known that the sum of the first k terms ofthe Singular Value Decomposition of a matrix givesits optimal rank-k approximation (under the 2-normor the Frobenius norm). Is computing the SVDessential for low-rank approximation?It was shown in 1998 (FKV) that a small sample ofrows chosen according to their squared lengths canbe used to compute a low-rank approximationwhose error is at most the best possible plus a termthat depends on the sample size and the norm of theoriginal matrix. This leads to a randomized algorithmWe shall discuss two such decompositions. Given anm-by-n matrix A, the first one approximates A by theproduct CX, where C consists of a few columns of A,and X is a coefficient matrix. The second one is of theform CUR, where C consists of a few columns of A, Rconsists of a few rows of A, and U is a carefully7

Wednesday: June 21, 2006constructed, constant-sized matrix. Previous suchmatrix decompositions only guaranteed additiveerror approximations. Our algorithms forconstructing C, X, U, and R take low polynomial timeand return approximations that are almost the bestpossible in a relative error sense. They employ thesubspace sampling technique; in particular, rows andcolumns of A are sampled with probabilities thatdepend on the lengths of the rows of the top fewsingular vectors of A.2:30-3:00SPARSE REDUCED-RANKAPPROXIMATIONS TO SPARSE MATRICESThe purpose of this talk is to describe an algorithmfor producing a reduced rank approximation to asparse m x n matrix A in which the factors are eithersmall or sparse. When the sparseness of thefactorization is not an issue, there are s: the singular value decompositionand the pivoted QR decomposition. This talk focuseson the latter.Petros DrineasDepartment of Computer ScienceRensselaer Polytechnic Institute1:30-2:30We will show how a variant of the pivoted Gram-Schmidt algorithm can produce approximations to Aof the form XS, where X is an m x p matrix consistingof columns of X and S is a p x n dense matrix. Thus ifp is sufficiently small, this factorization is suitable forthe case where m is much larger than n. The samealgorithm applied to the transpose of A anapproximation of the form TY, where Y is an p x nmatrix consisting of rows of A and T is an dense m xp matrix. If both approximations have beencomputed, they may be combined into anapproximation of the form XUY, where U is a p x pmatrix--an approximation suitable for the casewhere both m and n are large.MATRIX FACTORIZATIONS FORINFORMATION RETRIEVALThis tutorial will survey a wide variety of matrixdecompositions that have been proposed for use ininformation retrieval. Comparisons among themethods and their applicability to massive data setswill be emphasized.Dianne O'LearyDepartment of Computer Science and Institute forAdvanced Computer StudiesUniversity of Maryland, College ParkThe approximations can be computed a column (orrow) at a time. At each stage the error in theapproximation can be computed essentially for free,thus giving a rigorous criterion for stopping theprocess.G.W. StewartDepartment of Computer ScienceUniversity of Maryland, College Park8

Linear Algebra Basicsdecompositions of the following form: given an mby-n matrix A, decompose it as a product of threematrices, C, U, and R, where C consists of a fewcolumns of A, R consists of a few rows of A, and U isa small carefully constructed matrix that guaranteesthat the product CUR is "close," either provably orempirically, to A.Applications of suchdecompositions include the computation ofcompressed “sketches” for large matrices in a passefficient manner, matrix reconstruction, speeding upkernel-based statistical learning computations,sparsity-preservation in low-rank approximations,and improved interpretability of data analysismethods. We shall review the motivation for thesedecompositions, discuss various choices for thematrices C, U, and R that are appropriate in differentapplication domains, and then discuss theapplication of CUR decompositions to three diverseapplication domains: human genetics, medicalimaging, and internet recommendation systems. Ineach of these applications, the columns and rowshave a natural interpretation in terms of theprocesses generating the data, and CURdecompositions can be used to solve reconstruction,clustering, classification, or prediction problems ineach of these areas.3:00-4:00ADAPTIVE DISCRIMINANT ANALYSIS BYREGULARIZED MINIMUM SQUAREDERRORSFisher's discriminant analysis for binary classdimension reduction and the binary classifier designbased on the minimum squared error formulation(MSE) have been widely utilized for handling highdimensional clustered data sets. As the data sets getmodified from incorporating new data points anddeleting obsolete data points, there is a need todevelop efficient updating and downdatingalgorithms for these methods to avoid expensiverecomputation of the solution from scratch.In this talk, an efficient algorithm for adaptive linearand nonlinear kernel discriminant analysis based onregularized MSE, called KDA/RMSE, is introduced. Inadaptive KDA/RMSE, updating and downdating on(EVD)orsingularvaluedecomposition (SVD) is approximated by updatingand downdating of the QR decomposition achievingan order of magnitude speed up. This fast algorithmfor adaptive kernelized discriminant analysis isdesigned by utilizing regularization and somerelationship between linear and nonlineardiscriminant analysis for dimension reduction andthe MSE for classifier design. An efficient algorithmfor computing leave-one-out cross validation is alsointroduced by utilizing downdating of KDA/RMSE.Michael MahoneyYahoo! Research4:30-5:00FAST ALGORITHMS FOR GRAPHPARTITIONING AND SPARSIFICATION,AND THE SOLUTION OF SYMMETRICDIAGONALLY-DOMINANT LINEARSYSTEMSJoint work with Hyunsoo Kim and Barry Drake.Haesun ParkCollege of ComputingGeorgia Institute of TechnologyMotivated by the problem of solving systems oflinear equations, we develop nearly-linear timealgorithms for partitioning graphs and for buildingstrong sparsifiers. The graph partitioning algorithm isthe first provably fast graph partitioning algorithmthat finds sparse cuts of nearly optimal balance. It isbased on an algorithm for finding small clusters inmassive graphs that runs in time proportional to thesize of the cluster it outputs. Using the graph4:00-4:30CUR MATRIX DECOMPOSITIONS FORIMPROVED DATA ANALYSISMuch recent work in theoretical computer science,numerical linear algebra, machine learning, and dataanalysishasconsideredlow-rankmatrix9

Wednesday: June 21, 2006partitioning algorithm and random sampling, weshow how to spectrally approximate any graph by asparse subgraph.Joint work with Robert Calderbank.Anna Gilbert/Martin StraussDepartment of MathematicsUniversity of Michigan, Ann ArborWe then apply these algorithms, along withconstructions of low-stretch spanning trees, tooptimize a preconditioner construction introducedby Vaidya. We thereby obtain a nearly-linear timealgorithm for approximately solving systems of linearequations of the form Ax b, where A is symmetricand diagonally dominant.5:30-6:00LOW-RANK NONNEGATIVEFACTORIZATIONS FOR SPECTRALIMAGING APPLICATIONSNonnegative matrix factorization (NMF) is atechnique with numerous applications that has beenused extensively in recent years for approximatinghigh dimensional data where the data are composedof nonnegative entries. The process involves lowrank nonnegative approximate matrix factorizationsto allow the user to work with reduced dimensionalmodels and to facilitate more efficient statisticalclassification of data. In spectral imaging, we willdescribe our use of NMF to obtain spectralsignatures for space object identification using apoint source, imaged at different spectralfrequencies. In other applications, we are alsoconcerned with multispectral data to be processedfor extended object identification and analysis, andfor processing stacks of correlated 2D images. Wewish to find a nonnegative factorization of amultiway array, i.e., a tensor. The non-negativetensor factorization (NTF) problem has a unique setof difficulties for our applications that we will discuss.Joint work with Shang-Hua Teng.Daniel SpielmanDepartment of Computer ScienceYale University5:00-5:30LIST DECODING OF NOISY REED-MULLERLIKE CODESCoding theory has played a central role in thedevelopment of computer science. One critical pointof interaction is decoding error-correcting codes.First- and second-order Reed-Muller (RM(1) andRM(2), respectively) codes are two fundamentalerror-correcting codes which arise in communicationas well as in probabilistically checkable proofs andlearning. (The theory of first-order Reed-Mullercodes is also known as Fourier analysis on theBoolean cube.)Joint work with Christos Boutsidis, Misha Kilmer, andPaul Pauca.In this paper, we take the first steps towardextending the decoding tools of RM(1) into therealm of quadratic binary codes. We show how torecover a substantial subcode of RM(2) called aKerdock code, in the presence of significant noise.The Kerdock codes are a well-studied family of codesfor coding theory, radar signaling, and spreadspectrum communication. Our result is a listdecoding result for Kerdock codes which is roughlyanalogous to that of RM(1). In addition, we presenta new algorithmic characterization of Kerdock codesthat we hope will be more useful to the algorithmic(and coding theory) community than the classicdescriptions.Bob PlemmonsDepartment of Mathematics andComputer ScienceWake Forest University10

Linear Algebra Basics6:00-6:30A HYBRID OF MULTIVARIATE REGRESSIONAND FACTOR ANALYSISMultivariate regression is a useful tool for identifyingage related genes from microarray data. Introducinglatent variables into the regression provides adiagnostic tool for spotting observations that areoutliers and for identifying missing variables. Anexample of the former is a patient whose kidneyappears, from gene expression, to be that of a muchyounger patient. The latent variables alone yield afactor analysis model while the measured variablesalone are a multivariate regression. The criterion forthe combined model is a Frobenious norm on theresidual. The set of minimizers can be characterizedin terms of a singular value decomposition. A poweriteration appears to be faster the SVD in some realdata.Joint work with Stuart Kim and Jacob Zahn.Art B. OwenDepartment of StatisticsStanford University11

Thursday: June 22, 2006 (Industrial Applications and Sampling Methods)9:00-10:0011:00-11:30THE CHANGING FACE OF WEB SEARCHTEXT MINING APPROACHES FOR EMAILSURVEILLANCEWeb search has come to dominate ourconsciousness as a convenience we take for granted,as a medium for connecting advertisers and buyers,and as a fast-growing revenue source for thecompanies that provide this service. Following abrief overview of the state of the art and how we gotthere, this talk covers a spectrum of technicalchallenges arising in web search - ranging from socialsearch to auction design and incentive mechanisms.Prabhakar RaghavanYahoo! Research10:00-10:30STATISTICAL RANKING PROBLEMRanking has important applications in web-search,advertising, user recommender systems, etc, and isessential for internet companies such as Yahoo.In this talk, I will first go over some formulations andmethods of ranking in the statistical literature. ThenI will focus on a formulation suitable for web search,and talk about training relevance models based onDCG (discounted cumulated gain) optimization.Under this metric, the system output quality isnaturally determined by the performance near thetop of its rank-list. I will mainly discuss varioustheoretical issues in this learning problem. Theyreflect real issues encountered at Yahoo whenbuilding a machine learning based web-searchranking system.Joint work with David Cossock at Yahoo.Tong ZhangYahoo! ResearchAutomated approaches for the identification andclustering of semantic features or topics are highlydesired for text mining applications. Using a lowrank non-negative matrix factorization algorithm toretain natural data non-negativity, we eliminate theneed to use subtractive basis vector and encodingcalculations present in techniques such as principalcomponent analysis for semantic feature abstraction.Existing techniques for non-negative matrixfactorization are briefly reviewed and a newapproach for non-negative matrix factorization ispresented. A demonstration of the use of thisapproach for topic (or discussion) detection andtracking is presented using the Enron EmailCollection.Joint work with Murray Browne.Michael BerryDepartment of Computer ScienceUniversity of Tennessee, Knoxville11:30-12:00INCORPORATING QUERY DIFFERENCE FORLEARNING RETRIEVAL FUNCTIONSIn this talk we discuss information retrieval methodsthat aim at serving a diverse stream of user queries.We propose methods that emphasize theimportance of taking into consideration of querydifference in learning effective retrieval functions.We formulate the problem as a multi-task learningproblem using a risk minimization framework. Inparticular, we show how to calibrate the empiricalrisk to incorporate query difference in terms ofintroducing nuisance parameters in the statisticalmodels, and we also propose an alternatingoptimization method to simultaneously learn theretrieval function and the nuisance parameters. We12work out the details for both L and L regularization

Industrial Applications and Sampling Methodscases. We illustrate the effectiveness of theproposed methods using modeling data extractedfrom a commercial search engine.result that there exists a set of roughly k log nvectors of dimension n each such that any ndimensional vector can be recovered to nearly bestaccuracy possible using k terms, from only the innerproducts of the vector with the k log n vectors. Thisresult has partial precursors in algorithms researchand many postcursors. I will provide an overview ofCompressed Sensing, its pre- and postcursors, froman algorithmicist's point of view.Joint work with Zhaohui Zheng, Haoying Fu andGordon Sun.Hongyuan ZhaDepartment of Computer Science and EngineeringPennsylvania State University12:00-12:30Muthu MuthukrishnanGoogle Inc.EFFICIENT L2 AND L1 DIMENSIONREDUCTION IN MASSIVE DATABASES3:00-3:30Sampling and sketching (including randomprojections) are two basic strategies for dimension2reduction. For dimension reduction in L usingrandom projections, we show that the accuracy canbe improved by taking advantage of the marginalinformation; and the efficiency can be improveddramatically using a very sparse random projection1matrix. For dimension reduction in L , we prove ananalog of the JL lemma using nonlinear estimators.Previous known results have proved that no analog1of the JL lemma exists for L if restricted to linearestimators. As a competitor to random projections, asketch-based sampling algorithm is very efficient andhighly flexible in sparse data. This samplingalgorithm generates random samples online fromsketches. These various methods are useful formining huge databases (e.g., association rching, as well as efficiently computing the SVMkernels.KERNEL LEARNING WITH BREGMANMATRIX DIVERGENCESBregman divergences offer a rich variety ofdistortion measures that are applicable to variedapplications in data mining and machine learning.Popular Bregman vector divergences are the squaredEuclidean distance, relative entropy and the ItakuraSaito distance. These divergence measures can beextended to Hermitian matrices, yielding as specialcases, the squared Frobenius distance, vonNeumann divergence and the Burg divergence. Inthis talk, I will talk about the kernel learning problemthat uses these Bregman matrix divergences in anatural way. A key advantage is in obtaining efficientupdate algorithms for learning low-rank kernelmatrices.Joint work with Brian Kulis, Matyas Sustik and JoelTropp.Inderjit DhillonDepartment of Computer SciencesUniversity of Texas, AustinJoint work with Kenneth Church.Trevor Hastie/Ping LiDepartment of StatisticsStanford University2:00-3:00AN ALGORITHMICIST'S LOOK ATCOMPRESSED SENSING PROBLEMSRecently Donoho introduced the problem ofCompressed Sensing and proved an interesting13

Thursday: June 22, 2006In recent years, there has been extensive researchdone on approximate variants of the nearestneighbor problem. One of the algorithms, based onthe idea of Locality-Sensitive Hashing (LSH), has beensuccessfully used in several applied scenarios.3:30-4:00LATENT SEMANTIC ANALYSIS ANDFIEDLER RETRIEVALLatent semantic analysis (LSA) is a method forinformation retrieval and processing which is basedupon the singular value decomposition. It has ageometric interpretation in which objects (e.g.documents and keywords) are placed in a lowdimensional geometric space. In this talk, we derivea new algebraic/geometric method for placingobjects in space to facilitate information analysis.Our approach uses an alternative motivation, andreduces to the computation of eigenvectors ofLaplacian matrices. We show that our method,which we call Fiedler retrieval, is closely related toLSA, and essentially equivalent for particular choicesof scaling parameters. We then show that Fiedlerretrieval supports a number of generalizations andextensions that existing LSA approaches cannothandle, including unified text and link analysis.In this talk I will review that work, as well as describerecent developments in this area. In particular, I willshow a new LSH-based algorithm, whose runningtime significantly improves over the earlier bounds.The running time of the new algorithm is provablyclose to the best possible in the class of hashingbased algorithms.Piotr IndykComputer Science and Artificial Intelligence LabMassachusetts Institute of Technology5:00-5:30COMPACT REPRESENTATIONS FOR DATASeveral algorithmic techniques have been devisedrecently to deal with large volumes of data. At theheart of many of these techniques are ingeniousschemes to represent data compactly. This talk willpresent several examples of such compactrepresentation schemes. Some of these are inspiredby techniques devised in the field of approximationalgorithms for “rounding” solutions of LP and SDPrelaxations. We will also see how such compactrepresentation schemes lead to efficient, one-passalgorithms for processing large volumes of data(streaming algorithms).Bruce HendricksonSandia National Laboratories4:30-5:00NEAR-OPTIMAL HASHING ALGORITHMFOR THE APPROXIMATE NEARESTNEIGHBOR PROBLEMThe nearest neighbor problem is defined as follows:given a set of n data points in a d-dimensional space,construct a data structure which, given a query pointq, returns the data point closest to the query.Moses CharikarDepartment of Computer SciencePrinceton UniversityThe problem is of importance in multiple areas ofcomputer science. Nearest neighbor algorithms canbe used for: classification (in machine learning),similarity search (in biological, text or visualdatabases), data quantization (in signal processingand compression), etc. Unfortunately, classicalgorithms for geometric search problems, such askd-trees, do not scale well as the dimensionincreases.14

Industrial Applications and Sampling Methods5:30-6:006:00-6:30AT THE CONFLUENCE OF STREAMS;ORDER, INFORMATION AND SIGNALSPRESERVING PRIVACY IN LARGE-SCALEDATA ANALYSISIn this talk we will explore several new directions indata streams, particularly at the intersection ofstreams, information and signal processing. One ofthe most important features of a data stream is themeaning or the semantics of the elements streamingby. They typically either define a distribution over adomain or represent a signal. We will investigate afew problems corresponding to these two scenarios.The first set of problems will consider modeling adistribution as a stream of updates, and investigatespace efficient algorithms for computing variousinformation theoretic properties, divergences etc.The second set of problems would consider waveletprocessing on data streams.We discuss privacy issues associated with theanalysis of sensitive data sets, motivated by thegeneral inaccessibility of sensitive data due toprivacy concerns. We present a definition of“differential privacy,” in which each user is assuredthat no conclusion reached by the data analyst ismuch more likely than if than user had notparticipated in the study. The unequivocal nature ofthe statement, quantified over all possibleconclusions and all prior knowledge of the analyst,gives very strong privacy guarantees. Additionally,we show detail how many common analyses can befaithfully implemented in a framework that yieldsdifferential privacy, and discuss the opportunities forincorporating new analyses.Sudipto GuhaDepartment of Computer andInformation ScienceUniversity of PennsylvaniaJoint work with Cynthia Dwork.Frank McSherryMicrosoft Research15

Thursday: June 22, 2006Friday: June 23, 2006 (Kernel Learning and Applications)and ultimately for science itself. I will then show afew examples of our efforts in developing machinesthat learn for applications such as visual recognitionand graphics and speech synthesis. Finally, I willsketch a new theory of the ventral stream of thevisual cortex, describing how the brain may learn torecognize objects, and show that the resulting modelis capable of performing recognition on datasets ofcomplex images at the level of human performancein rapid categorization tasks. The model performs aswell or better than most state-of-art computer visionsystems in recognition of complex images.9:00-10:00APPLICATIONS OF RANDOM MATRICES INSPECTRAL COMPUTATIONS AND MACHINELEARNINGWe will see how carefully crafted random matricescan be used to enhance a wide variety ofcomputational tasks, including:dimensionalityreduction, spectral computations, and kernelmethods in machine learning. Several examples willbe considered, including the following two.Imagine that we want to compute the top feweigenvectors of a matrix A, hoping to extract the“important” features in A. We will prove that eitherthis is not worth doing, or that we can begin byrandomly throwing away a large fraction of A'sentries.Tomaso PoggioCenter for Biological and Computational Learning,Computer Science and Artificial IntelligenceLaboratory, and McGovern Institute for BrainResearchMassachusetts Institute of Technology11:00-11:30A famous result of Johnson and Lindenstraussas

Massive Datasets Stanford University and Yahoo! Research The MMDS (modern massive datasets) workshop provides a forum for discussions on massive, high-dimensional, and nonlinear-structured data between computer scientists, computational and applied mathematicians, statisticians, a

Related Documents:

Bruksanvisning för bilstereo . Bruksanvisning for bilstereo . Instrukcja obsługi samochodowego odtwarzacza stereo . Operating Instructions for Car Stereo . 610-104 . SV . Bruksanvisning i original

10 tips och tricks för att lyckas med ert sap-projekt 20 SAPSANYTT 2/2015 De flesta projektledare känner säkert till Cobb’s paradox. Martin Cobb verkade som CIO för sekretariatet för Treasury Board of Canada 1995 då han ställde frågan

service i Norge och Finland drivs inom ramen för ett enskilt företag (NRK. 1 och Yleisradio), fin ns det i Sverige tre: Ett för tv (Sveriges Television , SVT ), ett för radio (Sveriges Radio , SR ) och ett för utbildnings program (Sveriges Utbildningsradio, UR, vilket till följd av sin begränsade storlek inte återfinns bland de 25 största

Hotell För hotell anges de tre klasserna A/B, C och D. Det betyder att den "normala" standarden C är acceptabel men att motiven för en högre standard är starka. Ljudklass C motsvarar de tidigare normkraven för hotell, ljudklass A/B motsvarar kraven för moderna hotell med hög standard och ljudklass D kan användas vid

LÄS NOGGRANT FÖLJANDE VILLKOR FÖR APPLE DEVELOPER PROGRAM LICENCE . Apple Developer Program License Agreement Syfte Du vill använda Apple-mjukvara (enligt definitionen nedan) för att utveckla en eller flera Applikationer (enligt definitionen nedan) för Apple-märkta produkter. . Applikationer som utvecklas för iOS-produkter, Apple .

XSEDE HPC Monthly Workshop Schedule January 21 HPC Monthly Workshop: OpenMP February 19-20 HPC Monthly Workshop: Big Data March 3 HPC Monthly Workshop: OpenACC April 7-8 HPC Monthly Workshop: Big Data May 5-6 HPC Monthly Workshop: MPI June 2-5 Summer Boot Camp August 4-5 HPC Monthly Workshop: Big Data September 1-2 HPC Monthly Workshop: MPI October 6-7 HPC Monthly Workshop: Big Data

och krav. Maskinerna skriver ut upp till fyra tum breda etiketter med direkt termoteknik och termotransferteknik och är lämpliga för en lång rad användningsområden på vertikala marknader. TD-seriens professionella etikettskrivare för . skrivbordet. Brothers nya avancerade 4-tums etikettskrivare för skrivbordet är effektiva och enkla att

Den kanadensiska språkvetaren Jim Cummins har visat i sin forskning från år 1979 att det kan ta 1 till 3 år för att lära sig ett vardagsspråk och mellan 5 till 7 år för att behärska ett akademiskt språk.4 Han införde två begrepp för att beskriva elevernas språkliga kompetens: BI