Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

Similar documents
Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

Multimedia Databases 1/29/ Indexes for Multimedia Data Indexes for Multimedia Data Indexes for Multimedia Data

Dimensionality Reduction

Multimedia Databases. Previous Lecture. 4.1 Multiresolution Analysis. 4 Shape-based Features. 4.1 Multiresolution Analysis

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

Multimedia Databases. 4 Shape-based Features. 4.1 Multiresolution Analysis. 4.1 Multiresolution Analysis. 4.1 Multiresolution Analysis

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Multimedia Databases. Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Data Warehousing & Data Mining

Lecture: Face Recognition and Feature Reduction

Notes on Latent Semantic Analysis

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Lecture: Face Recognition and Feature Reduction

Multimedia Databases. Previous Lecture Video Abstraction Video Abstraction Example 6/20/2013

Machine learning for pervasive systems Classification in high-dimensional spaces

CS47300: Web Information Search and Management

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

1 Singular Value Decomposition and Principal Component

Singular Value Decompsition

Statistical Pattern Recognition

Pavel Zezula, Giuseppe Amato,

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Multimedia Databases. Wolf-Tilo Balke Younès Ghammad Institut für Informationssysteme Technische Universität Braunschweig

EE731 Lecture Notes: Matrix Computations for Signal Processing

Functional Maps ( ) Dr. Emanuele Rodolà Room , Informatik IX

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II

Preprocessing & dimensionality reduction

Principal Components Theory Notes

Multimedia Databases Video Abstraction Video Abstraction Example Example 1/8/ Video Retrieval Shot Detection

Linear Algebra (Review) Volker Tresp 2017

Nearest Neighbor Search for Relevance Feedback

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Robustness of Principal Components

Linear Algebra (Review) Volker Tresp 2018

Basic Concepts in Matrix Algebra

5 Spatial Access Methods

Manning & Schuetze, FSNLP, (c)

Review of Linear Algebra

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Numerical Methods I Singular Value Decomposition

Robotics 2 Data Association. Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard

AMS526: Numerical Analysis I (Numerical Linear Algebra)

7. Variable extraction and dimensionality reduction

Homework 2. Solutions T =

Data Warehousing & Data Mining

Matrix Decomposition and Latent Semantic Indexing (LSI) Introduction to Information Retrieval INF 141/ CS 121 Donald J. Patterson

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

Information Retrieval. Lecture 6

Chapter XII: Data Pre and Post Processing

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

EE731 Lecture Notes: Matrix Computations for Signal Processing

Multivariate Statistical Analysis

Random projection trees and low dimensional manifolds. Sanjoy Dasgupta and Yoav Freund University of California, San Diego

7. Symmetric Matrices and Quadratic Forms

Introduction to Matrix Algebra

5 Spatial Access Methods

Image Data Compression

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Clustering VS Classification

L26: Advanced dimensionality reduction

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1

Principal components analysis COMS 4771

An Algorithmist s Toolkit Nov. 10, Lecture 17

15 Singular Value Decomposition

Data Warehousing & Data Mining

Data Mining Techniques

Parallel Singular Value Decomposition. Jiaxing Tan

Lecture 7. Econ August 18

Carnegie Mellon Univ. Dept. of Computer Science Database Applications. SAMs - Detailed outline. Spatial Access Methods - problem

Clusters. Unsupervised Learning. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Advanced data analysis

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

Matrices, Vector Spaces, and Information Retrieval

USING SINGULAR VALUE DECOMPOSITION (SVD) AS A SOLUTION FOR SEARCH RESULT CLUSTERING

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Lecture 3: Review of Linear Algebra

Issues and Techniques in Pattern Classification

Motivating the Covariance Matrix

Lecture 15: Random Projections

Approximation Algorithms

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Singular Value Decomposition

CS 3750 Advanced Machine Learning. Applications of SVD and PCA (LSA and Link analysis) Cem Akkaya

Maximum variance formulation

Multimedia Networking ECE 599

Apprentissage non supervisée

The Gram Schmidt Process

Noise & Data Reduction

The Gram Schmidt Process

Review (Probability & Linear Algebra)

Maths for Signals and Systems Linear Algebra for Engineering Applications

PCA Review. CS 510 February 25 th, 2013

Issues Concerning Dimensionality and Similarity Search

Assignment 3. Latent Semantic Indexing

Unsupervised learning: beyond simple clustering and PCA

Latent Semantic Indexing (LSI) CE-324: Modern Information Retrieval Sharif University of Technology

Transcription:

Multimedia Databases Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de

14 Indexes for Multimedia Data 14 Indexes for Multimedia Data 14.1 Curse of Dimensionality 14.2 Dimension Reduction 14.3 GEMINI Indexing Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 2

14.1 Curse of Dimensionality Curse of Dimensionality Why are traditional index structures useless in multidimensional spaces? For (approx.) uniformly distributed data: all known index trees start failing at about15-20 dimensions Their use leads to higher costs than a linear scan Is it possible to create an efficient high-dimensional tree structure? What structure do high-dimensional spaces have anyway? Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 3

14.1 Basic Geometry Relationship between high-dimensional cubes and spheres in the R d (Hyper-) cube: Edge length 1 Center c = (½, ½,, ½) (Hyper-) sphere: Radius1 Center o = (0, 0,..., 0) Is the center of the cube always inside the sphere? o c Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 4

14.1 Basic Geometry Calculating the Euclidean distance between the two centers: For d = 4 the center of the cube is on the edge of the sphere For d 5 the center of the cube is outside the sphere Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 5

14.1 Basic Geometry Where are points in high-dimensional space located? Inside or on the surface of a cube? Consider cube A = [0,1] d Cut out smaller cube B with edge length (1-2ε) The volume of A is 1, the volume of B is (1 2ε) d A B ε (1-2ε) ε Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 6

14.1 Basic Geometry In high-dimensional spaces, almost all points lie near the surface of A What is the volume of the inner cube? ε / d 2 50 100 500 1000 0,1 0,64 1,4 10 5 2 10 10 3,5 10 49 1,2 10 97 0,05 0,81 0,01 2,7 10 5 1,3 10 23 1,8 10 46 0,01 0,96 0,36 0,13 4,1 10 5 1,7 10 9 If a point is positioned randomly (uniformly) in the outer cube, for large d it is with very low probability in the inner cube Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 7

14.1 Basic Geometry How big is the volume of spheres inscribed in cubes? Again cube A = [0,1] d and an inscribed sphere S For even d, S has volume 0.5 Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 8

14.1 Basic Geometry How big is the volume of S? How many randomly distributed points in the cube are needed, to ensure on average that at least one lies within the sphere? The number of points grows exponentially in d d Volume Nr. Points 2 0,79 2 4 0,31 4 10 0,002 402 20 2,46 10 8 4063127 40 3,28 10 21 3,05 10 20 100 1,87 10 70 5,35 10 69 Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 9

14.1 Basic Geometry How far away is the nearest neighbor from the center of the sphere? 1,000,000 uniformly distributed points in the cube Even the closest neighbors are far away! The maximum distance is d 1/2 (diagonal) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 10

14.1 Basic Geometry How many points are there at exactly distance s from the center? d s: As the distance increases, the variance is lower! For large d, almost all points have the same distance from the query (Beyer et al, 1998) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 11

14.1 Conclusion High-dimensional spaces are different In high-dimensional spaces, the sequential search through the objects is often better than using some index structure On the other hand, our analysis was focused on uniformly distributed points in Euclidean spaces Real-world data may have an intrinsic lower dimensionality Example: dimensions price and maximal speed in a vehicle database Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 12

14.1 Speeding up Sequential Search Vector Approximation Files (Weber et al, 1998) Partition each dimension into intervals Dimension i is divided into 2 b i intervals Represented by b i bits E.g., splitting some dimension in 4 intervals Representation of these intervals by 00, 01, 10 and 11 The i-th coordinate of each data point can thus approximately be represented by b i bits Thus, each point can be approximated by b = b 1 + b 2 + + b d bits Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 13

14.1 VA Files Points and their encoding in a VA-File Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 14

14.1 VA Files Advantages of VA Files If b is large enough, there are significantly more partitions of space into hyper-cubes as there are data points Thus, collisions are nearly impossible, and every bit vector represents just one point It is much faster to perform bit-wise operations on fixed-length bit vectors, than performing calculations on the original representation of the data Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 15

14.1 VA Files Query processing: filter & refine e.g., region queries: Sequentially scan over all partitions For partitions that intersect with the search region Check all contained points using their exact coordinates Actual results False drops Excluded points Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 16

14.2 Dimensionality Reduction Indexing of high dimensional data is problematic, but does every dimension contain essential information? Strongly correlated dimensions can be combined e.g., price in Euro and price in Dollars If in some dimensions the objects differ more, then these dimensions also carry more information than others Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 17

14.2 Principal Component Analysis Principal Component Analysis (PCA), also known as Karhunen-Loève transform Detection of linear dependencies between features, so-called axes The most pronounced axis is called the main axis The correlation is always subject to a certain variation Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 18

14.2 Principal Component Analysis Linear dependence is a sign of redundancy A dimension may be represented as a linear combination of other dimensions Idea Rotate (and shift) the axes such that there is no linear dependence between them Remove all axes with low variance, to keep the error introduced by omission of information minimal Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 19

14.2 Principal Component Analysis Example: rotate the basis of a coordinate system Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 20

14.2 Principal Component Analysis The covariance matrix determines linear dependencies between different data dimensions Let X = (X 1,..., X d ) be a random vector which is uniformly distributed on the set of (d-dimensional) data points Center the coordinate system around the mean:, with Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 21

14.2 Principal Component Analysis The covariance between X i and X j is The covariance is positive, if X j always tends to have large values whenever X i does (and vice versa) The covariance is negative, if X j always tends to have large values whenever X i has small values (and vice versa) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 22

14.2 Principal Component Analysis The covariance matrix contains all pair-wise covariances between the dimensions Dimensions with nonzero covariance are interdependent and thus carry redundant information Idea: rotate the centered data around the origin such that all covariances have value 0 The dimensions are then linearly independent from each other Distances between data points are preserved, since it is only a rotation Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 23

14.2 Principal Component Analysis Linear Algebra: Any symmetric matrix A can be diagonalized This means that there are matrices Q and D with A = Q D Q 1 Where Q is orthogonal, therefore Q T = Q 1 D is diagonal, i.e., besides the main diagonal it contains only 0 s The orthogonal matrices belonging to linear mappings are always just reflections and rotations Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 24

14.2 Principal Component Analysis The covariance matrix is symmetric If the covariance matrix is diagonalized and the transformation corresponding to matrix Q is applied to the data, then the covariance between the new data dimensions is always 0 The properties of the transformation are transferred to the covariance! Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 25

14.2 Principal Component Analysis The diagonal matrix D contains the eigenvalues of the covariance matrix eigenvalue decomposition Dimensions with low variance of the data (small corresponding eigenvalues) can be removed after the transformations The resulting error is small and can be generally neglected Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 26

14.2 Example Data points: Center m= Covariance matrix Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 27

14.2 Example Transformation of the data: Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 28

14.2 Example x-axis is the main axis (since it has the largest eigenvalue) Variance on the y axis is small compared with variance on the x-axis (0.8 vs. 2.83) y-axis can therefore be omitted without much loss of information (dimensionality reduction) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 29

14.2 Latent Semantic Indexing A similar method to principal component analysis is Latent Semantic Indexing (LSI) In principal component analysis, the covariance matrix is decomposed, while in LSI the feature matrix F is decomposed The feature matrix contains the feature vectors as columns Note, this matrix is usually not quadratic and symmetric In this case: F = U D V T with diagonal matrix D and matrices U and V with orthonormal column vectors Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 30

14.2 Latent Semantic Indexing The decomposition of the feature matrix is a transformation on minimal concepts that are scaled by the eigenvalues in D Each concept is an artificial dimension of a vector space, which expresses the latent semantics of the feature vectors LSI is used in information retrieval, to represent synonyms by a common concept and to split ambiguous terms into different concepts Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 31

14.2 Latent Semantic Indexing Through the decomposition of the feature matrix, the semantics of feature vectors (the columns of the feature matrix) is contained in matrix V T feature dimensions (the rows of the feature matrix) is contained in matrix U They can be used as in the principal component analysis for extracting dependencies Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 32

14.2 Latent Semantic Indexing According to the eigenvalue decomposition, D indicates the importance of the new dimensions Irrelevant dimensions (particularly those with eigenvalue 0) can be deleted with minimal error Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 33

14.2 Latent Semantic Indexing The decomposition of the feature matrix for the previous example yields: = Two latent dimensions with weights 7.95 and 1.8 Weights of these dimensions in point 1: 0.17 and -0.14 U and V transform the original axes, respectively the data points in the new space Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 34

14.2 Dimension Reduction Principle component analysis and LSI, however, are difficult to use when there are frequent changes in the data Data changes could drastically affect the feature values and their respective variances and covariances For this reason, they are usually only calculated for a representative portion of the data and used for all objects Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 35

14.3 GEMINI Indexing GEMINI - GEneric Multimedia Object INdexIng (Faloutsos, 1996) Can be used for hard to calculate distance functions between objects (e.g., edit distances) Only useful, if there are no directly usable efficient search algorithms Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 36

14.3 GEMINI Indexing Idea: use a distance function which is easy to compute as an estimate for the complex distance function and immediately prune some objects, i.e. a quick and dirty test E.g., use the distance regarding average colors for the exclusion of certain histograms from the comparison A histogram with avg. color red will be more similar to a histogram with avg. color orange, than to one with avg. color blue Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 37

14.3 GEMINI Indexing Assume we have to calculate a complex distance function d(a, B) on the set of objects O Choose a function (transformation) f on O and a simple distance function δ with the property A, B O : δ(f(a), f(b)) d(a, B) ( lower bounding property ) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 38

14.3 Example Comparison of time series for stock quotes The comparison of two curves with the Euclidean distance (d) on a discrete grid (e.g., closing prices) is very complicated Transforming each curve with DFT and comparing the coefficients yields the same result, but is just as expensive (Parseval's Theorem) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 39

14.3 Example But comparing only the first few coefficients corresponds to a dimension reduction (f) The Euclidean distance of only the first coefficient is faster to compute (δ) Since the Euclidean distance is calculated as a sum of non-negative terms (square), the distance δ of the first term always underestimates the distance d of all the terms (lower bounding) Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 40

14.3 Algorithm Preparation Transform all the database objects with function f Pruning To answer a region query (Q, r) exclude all objects A with δ(a, Q)> r from the search Query processing Compute d(b, Q) for all remaining objects B and exclude all objects with d(b, Q) > r Return all the remaining elements Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 41

14.3 Algorithm By using estimations false positives may occur, but never false negatives Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 42

14.3 Nearest Neighbors Nearest-neighbor queries can be performed using the index, too Attention: the nearest neighbor according to δ is not always the nearest neighbor according to d Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 43

14.3 Nearest Neighbors However, if A is the nearest neighbor of Q regarding δ, then the nearest neighbor of Q regarding d can have at most distance d(a, Q) Therefore: Find the nearest neighbors A of Q regarding δ Determine d(a, Q) Create an area query with radius d(a, Q) The nearest neighbor of Q regarding d is guaranteed to be among the results Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 44

14.3 Fast Map How can complicated distance functions be simplified (respecting the lower bounding)? Fast Map represents all database objects as points in R k, on the basis of some random distance function d (Faloutsos and Lin, 1995) The Euclidean distance then allows an approximation of the original distance Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 45

14.3 Fast Map Fast Map represents the objects in the k-dimensional space, where k is a pre-selected constant The higher the k, the more accurate is the respective approximation The lower the k, the more efficient is the search process Solution: find a suitable compromise depending on the application Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 46

14.3 Fast Map Requirements Efficient mapping (if possible only linear effort) Good distance approximation when regarding the original distances Efficient mapping of new objects independent of all other objects Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 47

14.3 Fast Map Idea: consider all the objects as already in k- dimensional Euclidean space, however without knowing their coordinates Find k orthogonal coordinate axes based on the distances between the points Project the points on the axes and compute the feature values in Euclidean space Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 48

14.3 Fast Map Projection (of a point C) on an axis given by pivot points A and B C A Considering the law of cosines: X B Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 49

14.3 Fast Map In this way, each point C can be projected onto the coordinate axis given by A and B No coordinates of the involved points are required for this, but only the distances from each other On this axis C has coordinate value d(a, X) C A X B Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 50

14.3 Fast Map After a new axis was added: Project all data points on the (k-1)-dimensional hyperplane, whose normal vector is the new axis We again have a data set with a total of k dimensions including the newly created one The coordinates of the remaining (k-1) dimensions are unknown We can however calculate the distances of the points in the hyperplane Arbitrary points A and B in the hyperplane have distance d'(a, B) with where x(a) and x(b) are the coordinates of A and B, on the newly added axis Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 51

14.3 Fast Map Repeat this steps k times In this way, we create k new (orthogonal) dimensions, which describe the data This allows us to approximate objects with complex distance function by points in the standard space For this, only the distance matrix is needed Afterwards, we can also apply standard algorithms Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 52

14.3 Fast Map Problem: which objects should we choose as pivot elements? The two objects that are farthest apart provide the least loss of accuracy They can be found with quadratic cost and with some heuristics, even within linear time Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 53

14.3 Example Calculate the pairwise dissimilarities in a set of words based on the edit distance To allow for indexing, all words should be mapped as multidimensional points in Euclidean space The Euclidean distance can then be used instead of the more complicated editing distance Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 54

14.3 Example Edit distance between words O = {Medium, Database, Multimedia, System, Object} =: {w 1,, w 5 } Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 55

14.3 Fast Map Mapping on 4-dimensional points: Original: Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 56

Next Semester Lectures Relational Database Systems II Information Retrieval and Web Search Engines Distributed Database Systems and Peer-to-Peer Data Management Seminar Information Extraction - How to Read the Web Multimedia Databases Wolf-Tilo Balke Institut für Informationssysteme TU Braunschweig 57