Tags:
tag this topic
create new tag
view all tags
---+Discussion on May 19th 2006 reading group meeting Paper presented: Arikan, Okan. _Compression of Motion Capture Databases_, to appear in Siggraph 2006 proceedings * Project page (with paper and video): http://www.cs.utexas.edu/~okan/papers/s2006/compression.html * [[#PaperOverview][Paper Overview]] * [[#PaperDiscussion][Paper Discussion]] ----------------------------------------------------------------- #PaperOverview ---++Paper Overview They present a method to compress a large database of skeletal motion data. Their method is separated in two parts: 1. Global motion compression 2. Specific precise compression of joints in contact with the environment (the feet in their case) __Global Motion Compression__ * Each joint is associated to 3 virtual markers which are tracked through time. The marker positions are said to be more _linear_ than the DOF angles. The DOF angles can easily be re-extracted from these markers, even after compression (using least-square fit). * Separate the motion in 16-32 frames clips * For each clip, fit a 3D Bezier curve through the moving markers * Each clip is therefore a vector in a space of dimension _d_ = 12 x 3 x _number of joints_ . * Reduce the dimension using Clustered PCA over the whole database * Spectral clustering using Nystrom Approximation (ref given) * Typically, 1 to 20 clusters * Randomly draw 10000 frames of the database before performing the CPCA * A parameter is used to decide how many dimensions are kept * Quantize elements of the reduced vector to 16 bits. __Specific Joint Compression__ * Ground reaction force is quite significant and applies over a veryshort time ==> High frequencies in the motion * Sliding feet are a perceptually important artifact * Consider the _x,y,z_ coordinates of the virtual markers on the feet (or other contact joints) as separate 1D signals * For each clip, apply DCT on these signals, then quantize, then entropy-encode (Huffman codes) * During decompression, use IK ([[ftp://ftp.cis.upenn.edu/pub/badler/public_html/gmod/0528a.pdf][Tolani et al. 2006]]) to plant foot at reconstructed position. __Features and results__ * To access an individual frame, one has to decompresse a 16-32 frame clip * Compresses at 1 ms/frame, decompress at 1.2 ms/frame (7 times real-time) * Random access any clip for decompression * CPCA performed offline on a random 10000 frames of animation, clips can be processed independantly * After CPCA, clips can be compressed independantly and incrementally. If statistical distribution changes, can perform CPCA again. * Clip-to-clip transition can be discontinuous. Fix this by solving a sparse linear system over the clip, called Continuous Merge (???) -------------------- #PaperDiscussion ---++Paper Discussion Here's what we think is missing in the paper : * Progressive compression / Generating various animation LOD * The technique doesn't take into account that the database is made of multiple sequences. * Good results only if the database is large enough * The frame-rates are not realistic for usage such as real-time games * Incrementally compressing motion is efficient as long as the _statistical properties_ do not change. When does that happen? * Baseline comparison methods are probably too simple * Decompression could exhibit cache issues since large chunk of data (PCA matrices) must be randomly accessed * Compression is faster than decompression? Sounds weird... * Justification for not using angular data is kind of weak Here are some ideas that came out : * Check how receptive field weighted regression would perform for temporal compression * Use progressive compression technique for large mocap database exploration over a slow channel (see Main.PhilippeBeaudoin) Some links to papers that are not referred to but are related : * A. Ahmed, A. Hilton, and F. Mokhtarian. _Adaptive compression of human animation data_. In Eurograhics - Short Paper, September 2002. * Naka, T., Mochizuki, Y., Hijiri, T., Cornish, T., and Asahara, S. _A compression/decompression method for streaming based humanoid animation_. In Proc. of the 4th Symp. on VRML, 1999. http://doi.acm.org/10.1145/299246.299264 * T.K. Capin, E. Petajan, J. Ostermann. _Very low bit rate coding of virtual human animation in MPEG-4_, Proceedings of ICME 2000. http://ieeexplore.ieee.org/xpl/abs_free.jsp?arNumber=871554 * S. Chattopadhyay, S.M. Bhandarkar, K. Li. _BAP sparsing: a novel approach to MPEG-4 body animation parameter compression_, Proceedings Systems Communications, 2005. http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=1515510 * S. Chattopadhyay, S.M. Bhandarkar, K. Li. _Compression by indexing: an improvement over MPEG-4 body animation parameter compression_, Proceedings of SPIE, 2006. http://spiedl.aip.org/getabs/servlet/GetabsServlet?prog=normal&id=PSISDG00607100000160710K000001&idtype=cvips&gifs=yes -- Main.PhilippeBeaudoin - 19 May 2006
E
dit
|
A
ttach
|
Watch
|
P
rint version
|
H
istory
: r4
<
r3
<
r2
<
r1
|
B
acklinks
|
V
iew topic
|
Ra
w
edit
|
M
ore topic actions
Topic revision: r4 - 2006-05-26
-
PhilippeBeaudoin
Home
Site map
BETA web
Communications web
Faculty web
Imager web
LCI web
Main web
SPL web
Sandbox web
TWiki web
TestCases web
Imager Web
Create New Topic
Index
Search
Changes
Notifications
RSS Feed
Statistics
Preferences
P
View
Raw View
Print version
Find backlinks
History
More topic actions
Edit
Raw edit
Attach file or image
Edit topic preference settings
Set new parent
More topic actions
Account
Log In
Register User
E
dit
A
ttach
Copyright © 2008-2025 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki?
Send feedback