File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Deep Learning Techniques for Sketch-Based Face Modeling and Automatic Image/Video Stylization
Title | Deep Learning Techniques for Sketch-Based Face Modeling and Automatic Image/Video Stylization |
---|---|
Authors | |
Issue Date | 2017 |
Publisher | Visual Computing Research Center. |
Citation | Visual Computing Summer School at Shenzhen University, Shenzhen, China, 16-19 July 2017 How to Cite? |
Abstract | Face modeling has been paid much attention in the field of visual computing. There exist many scenarios, including cartoon characters, avatars for social media, 3D face caricatures as well as face-related art and design, where low-cost interactive face modeling is a popular approach especially among amateur users. In this lecture, I present a deep learning based sketching system for 3D face and caricature modeling. This system has a labor-efficient sketching interface, that allows the user to draw freehand imprecise yet expressive 2D lines representing the contours of facial features. A novel CNN based deep regression network is designed for inferring 3D face models from 2D sketches. Our system also supports gesture based interactions for users to further manipulate initial face models. Both user studies and numerical results indicate that our sketching system can help users create face models quickly and effectively. A significantly expanded face database with diverse identities, expressions and levels of exaggeration is constructed to promote further research and evaluation of face modeling techniques. Color and tone stylization in images and videos strives to enhance unique themes with artistic color and tone adjustments. It has a broad range of applications from professional image postprocessing to photo sharing over social networks. I also present a novel deep learning architecture for exemplar-based image stylization, which learns local enhancement styles from image pairs. Image stylization can be efficiently accomplished with a single forward pass through our deep network. To extend our deep network from image stylization to video stylization, we exploit temporal superpixels (TSPs) to facilitate the transfer of artistic styles from image exemplars to videos. Experiments on a number of datasets for image stylization as well as a diverse set of video clips demonstrate the effectiveness of our deep learning architecture. |
Persistent Identifier | http://hdl.handle.net/10722/295660 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Yu, Y | - |
dc.date.accessioned | 2021-02-03T08:47:24Z | - |
dc.date.available | 2021-02-03T08:47:24Z | - |
dc.date.issued | 2017 | - |
dc.identifier.citation | Visual Computing Summer School at Shenzhen University, Shenzhen, China, 16-19 July 2017 | - |
dc.identifier.uri | http://hdl.handle.net/10722/295660 | - |
dc.description.abstract | Face modeling has been paid much attention in the field of visual computing. There exist many scenarios, including cartoon characters, avatars for social media, 3D face caricatures as well as face-related art and design, where low-cost interactive face modeling is a popular approach especially among amateur users. In this lecture, I present a deep learning based sketching system for 3D face and caricature modeling. This system has a labor-efficient sketching interface, that allows the user to draw freehand imprecise yet expressive 2D lines representing the contours of facial features. A novel CNN based deep regression network is designed for inferring 3D face models from 2D sketches. Our system also supports gesture based interactions for users to further manipulate initial face models. Both user studies and numerical results indicate that our sketching system can help users create face models quickly and effectively. A significantly expanded face database with diverse identities, expressions and levels of exaggeration is constructed to promote further research and evaluation of face modeling techniques. Color and tone stylization in images and videos strives to enhance unique themes with artistic color and tone adjustments. It has a broad range of applications from professional image postprocessing to photo sharing over social networks. I also present a novel deep learning architecture for exemplar-based image stylization, which learns local enhancement styles from image pairs. Image stylization can be efficiently accomplished with a single forward pass through our deep network. To extend our deep network from image stylization to video stylization, we exploit temporal superpixels (TSPs) to facilitate the transfer of artistic styles from image exemplars to videos. Experiments on a number of datasets for image stylization as well as a diverse set of video clips demonstrate the effectiveness of our deep learning architecture. | - |
dc.language | eng | - |
dc.publisher | Visual Computing Research Center. | - |
dc.relation.ispartof | Visual Computing Summer School at Shenzhen University | - |
dc.title | Deep Learning Techniques for Sketch-Based Face Modeling and Automatic Image/Video Stylization | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Yu, Y: yzyu@cs.hku.hk | - |
dc.identifier.authority | Yu, Y=rp01415 | - |
dc.identifier.hkuros | 276553 | - |
dc.publisher.place | Shenzhen | - |