Difference between revisions of "Private:progress-neshat"
From NMSL
Line 2: | Line 2: | ||
* '''Courses:''' None | * '''Courses:''' None | ||
'''working on: Large Scale data processing with MapReduce on GPU/CPU hybrid systems ''' | '''working on: Large Scale data processing with MapReduce on GPU/CPU hybrid systems ''' | ||
+ | |||
+ | === Jan 25 === | ||
+ | *Explored Hadoop and its properties like HDFS | ||
+ | *Explored Architecture of NVIDIA GPU cluster's arch and specs | ||
+ | *(On Going)Read two examples of large scale data proc. with map reduce | ||
+ | *Read papers about GPU clusters for HPC | ||
+ | *(On Going)Designed high-level architecture of proposed approach for using Hadoop over Hybrid CPU/GPU systems | ||
+ | |||
=== Jan 17 === | === Jan 17 === |
Revision as of 16:12, 21 January 2011
Spring 2011 (GF)
- Courses: None
working on: Large Scale data processing with MapReduce on GPU/CPU hybrid systems
Jan 25
- Explored Hadoop and its properties like HDFS
- Explored Architecture of NVIDIA GPU cluster's arch and specs
- (On Going)Read two examples of large scale data proc. with map reduce
- Read papers about GPU clusters for HPC
- (On Going)Designed high-level architecture of proposed approach for using Hadoop over Hybrid CPU/GPU systems
Jan 17
- read two papers about Phoenix, a mapreduce implementation for multi-core processors.
- spent some days to figure out how to use Mark framework and run some samples, but couldn't fully understand. These works has been done:
- Configured system (windows) to run Mars, including cuda and SDK installation as well as VS9 configuring.
- Corrected some typos in the code (library mismatching)
- Asking authors about problems, and got this answer: "I must apologize that mars_v2 is buggy and complex, and we don't maintain the code base any more, I strongly recommend you to try the latest version on linux"
- tried to install mars_v2 on Linux, but it is still buggy and complex. It seems this frame work could run only with certaing configuration, and with older versions of CUDA.
- Explored Mars to find its algorithm, and found in co-processing mode (Hybrid) they partition input data into two parts, one for CPU processing, the other for GPU processing. After the map stage, they merge data on CPU side, then dispatch data again to CPU workers and GPU workers.
- Looked at phonix, another System for MapReduce Programming from Stanford. It was the comparison base for Mars.
- Spent 2 days for writing resume and being prepared for YouTube interview.
Jan 10
- Explored related works and potential ideas
Fall 2010 (TA)
- Courses:
- CMPT-820: Multimedia Systems
- CMPT-825: NLP
- worked on:
- effective advertising in video
- Submissions:
- SmartAd: a smart autonomous system for effective advertising in video (ICME 11)
Summer 2010 (RA)
- Writing for publication
- worked on:
- Estimating the click-through rate for new ads with semantic and feature based similarity
algorithms
Spring 2010 (RA)
- Courses:
- CMPT-886: Special topics in operation systems
- worked on:
- Accelarting online auction using GPU
- Estimating the click-through rate for new ads with semantic and feature based similarity
algorithms
- submitted
- Accelerating online auctions with Optimized Parallel GPU based algorithms: Accelerating Vickrey-Clarke-Groves (VCG) Mechanism (proposal for GPU Gem book)
Fall 2009 (TA)
- Courses:
- CMPT-705: Algorithm
- CMPT-771: Internet Architecture and Protocols
- worked on:
- implementing FEC on mobile tv testbed