Difference between revisions of "Private:progress-dameh"

From NMSL
Line 2: Line 2:
 
* Research: Gram Matrix approximation
 
* Research: Gram Matrix approximation
 
** Recap:
 
** Recap:
*** Problem: kerenl methods gram matrix is $O(N^2)$, which is computational and memory usage infiseable for large-scale datasets
+
*** Problem: kernel methods gram matrix is O(N^2), which is computational and memory usage infeasible for large-scale datasets
*** Prevoius Work:
+
*** Previous Work:
 
**** Low rank matrix approximation:
 
**** Low rank matrix approximation:
***** using EginDecomposition: Optimal solution but reqiures <math>O(N^3)</math> computation to constuct the low rank matix
+
***** using EigneDecomposition: Optimal solution but requires O(N^3) computation to construct the low rank matrix
***** Solution: use Nysrom therorem to approximate the eigen decomposition of the matrix which reqiures <math>O(m^2N)</math> where m << N.
+
***** Solution: use Nytsrom theorem to approximate the EigenDecomposition of the matrix which requires O(m^2N) where m << N.
  
***** Problem: In this methods in despite of reduce the computation complexity still we need <math>O(N^2)</math> to store the matrix, so the next method reduce both computation and memory.
+
***** Problem: In this methods in despite of reducing the computation complexity still we need O(N^2) to store the matrix, so the next methods reduce both computation and memory.
  
 
**** Efficient implementation for computing the kernel matrix:
 
**** Efficient implementation for computing the kernel matrix:
 
***** Using Spatial Indexing:
 
***** Using Spatial Indexing:
using Z-order, H-order, Td-tress and other spatial indexing techniques to order the points in the space and then use sliding window where the kernel function computed only between points in that wiondow.
+
using Z-order, H-order, kd-tress and other spatial indexing techniques to order the points in the space and then using sliding window where the kernel function computed only between points in that window.
****** Problems: reqiures <math>O(NlogN)</math> as preprocessing to order the points, fails for very hight dimentional Data and hard to parallel
+
****** Problems: requires O(NlogN) as preprocessing to order the points, fails for very high dimensional Data and hard to parallel
So we proposed a method that works fine foy large-scale and highly dimentinla data, overcome or pervoius work drwobbacks, easy to parallel, reqiures linear time preprocessing and reduce the computation and memory usae down form quadratic to subquadratic:
+
So we proposed a method that works fine for  large-scale and highly dimensional data that overcomes all pervious works drawbacks, easy to parallel, requires linear time preprocessing and reduce both computation and memory usage down form quadratic to subquadratic:
 
***** Using Spatial Hashing or Locality sensitive hashing:
 
***** Using Spatial Hashing or Locality sensitive hashing:
where the kerenl function and the cluster methods works on the buckets indicusally.
+
where the kernel function and the cluster methods works on the buckets individually.
 
****** LSH families :
 
****** LSH families :
******* Random projection
+
******* Random projection: Hamming distance, Euclidean Distance, Cosine Distance
******** Hamming distance
+
Problem:  They assume data uniformly distributed solution:
******** Eclidean Distance
+
******* Spectral Hashing: overcomes previous families problmes
******** Cosine Distance
 
********* Problem:  assumes data uniforly distubuted solution:
 
******** Spectral Hasing
 
  
General Problem for random projection, it' doesnt take care of data comes in a kerenel space, Solution
+
General Problem for random projection, it' doesn’t take in account if data are in a kernel space, Solution
******* Kerenalized LSH
+
******* Kernelized LSH : Problem: more complex to compute
 +
 
 +
** In progress: Comparing different LSH families
  
** Recap: Comparing differnt LSH families
 
  
 
= Spring 2011 (TA/RA) =
 
= Spring 2011 (TA/RA) =

Revision as of 12:25, 6 May 2011

Summer 2011 (TA)

  • Research: Gram Matrix approximation
    • Recap:
      • Problem: kernel methods gram matrix is O(N^2), which is computational and memory usage infeasible for large-scale datasets
      • Previous Work:
        • Low rank matrix approximation:
          • using EigneDecomposition: Optimal solution but requires O(N^3) computation to construct the low rank matrix
          • Solution: use Nytsrom theorem to approximate the EigenDecomposition of the matrix which requires O(m^2N) where m << N.
          • Problem: In this methods in despite of reducing the computation complexity still we need O(N^2) to store the matrix, so the next methods reduce both computation and memory.
        • Efficient implementation for computing the kernel matrix:
          • Using Spatial Indexing:

using Z-order, H-order, kd-tress and other spatial indexing techniques to order the points in the space and then using sliding window where the kernel function computed only between points in that window.

            • Problems: requires O(NlogN) as preprocessing to order the points, fails for very high dimensional Data and hard to parallel

So we proposed a method that works fine for large-scale and highly dimensional data that overcomes all pervious works drawbacks, easy to parallel, requires linear time preprocessing and reduce both computation and memory usage down form quadratic to subquadratic:

          • Using Spatial Hashing or Locality sensitive hashing:

where the kernel function and the cluster methods works on the buckets individually.

            • LSH families :
              • Random projection: Hamming distance, Euclidean Distance, Cosine Distance

Problem: They assume data uniformly distributed solution:

              • Spectral Hashing: overcomes previous families problmes

General Problem for random projection, it' doesn’t take in account if data are in a kernel space, Solution

              • Kernelized LSH : Problem: more complex to compute
    • In progress: Comparing different LSH families


Spring 2011 (TA/RA)

  • Courses:
    • CMPT 885: Special Topics in Computer Architecture
    • CMPT 886: Special Topics in Operating Systems


Fall 2010 (FAS-GF/RA)

  • Courses:
    • CMPT 705: Design/Analysis Algorithms
    • CMPT 820: Multimedia Systems
  • Worked on Gram matrix approximation, High Performance/large scale application case studies


Summer 2010 (TA)

  • Worked on large Scale Kernel methods


Spring 2010 (RA)

  • Courses:
    • CMPT 768: Sound Synthesis
  • Worked on Locality sensitive hashing families and their applications and implementations