Figure 3 - uploaded by Peter Eisert
Content may be subject to copyright.
Examples of texture maps. Left: a fully con- structed texture map. Right: a partial texture map ex- tracted from a single image. 

Examples of texture maps. Left: a fully con- structed texture map. Right: a partial texture map ex- tracted from a single image. 

Source publication
Conference Paper
Full-text available
In this paper a model-based multi-view image generation system for video conferencing is presented. The system assumes that a 3-D model of the person in front of the camera is available. It extracts texture from speaking person sequence images and maps it to the static 3-D model during the videoconference session. Since only the incrementally updat...

Context in source publication

Context 1
... blending to combine textures from dif- ferent camera images. A texture map (see Figure 3) of the model is stored in the system. In the beginning of a video conference session, this texture map is empty. ...

Similar publications

Article
Full-text available
This paper discusses the use of videoconferencing as a tool to enhance collaboration amongst online learners in an open learning context. We present quantitative data from a longitudinal study of naturalistic videoconferencing interactions held via the FM videoconference tool, indicating its global use. The case study presented is analysed accordin...
Article
Full-text available
Background Paediatric ENT services in regional areas can be provided through telemedicine (tele-ENT) using videoconferencing or with a conventional outpatient department ENT service (OPD-ENT) in which patients travel to see the specialist. The objective of this study was to identify the least-cost approach to providing ENT services for paediatric o...
Poster
Full-text available
This poster presents preliminary descriptive and qualitative results from an in-progress study. We are testing the feasibility of providing a perspective-taking intervention a) via videoconference b) to groups of nursing students.
Article
Full-text available
Specialist care consultations were identified by two research nurses using documentation in patient records, appointment diaries, electronic billing services and on-site observations at a 441-bed long term care facility. Over a six-month period there were 3333 consultations (a rate of 1511 consultations per year per 100 beds). Most consultations we...
Article
Full-text available
Objective: to assess the effects of an educational intervention on smoking cessation aimed at the nursing team. Method: this is a quasi-experimental study with 37 nursing professionals from a Brazilian hospital from May/2019 to December/2020. The intervention consisted of training nursing professionals on approaches to hospitalized smokers divid...

Citations

... Optical flow algorithms can help when more detailed information about the behavior of borders [7,[11][12][13][14][15][16][17][18][19]. Deformation models may also aid describing the evolution of grains [26,27]. ...
Article
The foremost by-product of this paper is the automation of geological undertakings, for instance, dealing with exceptionally thin sections of rocks that were subjected to deformation alongside finite steps of time which can be recorded in video for later analysis using image processing and numerical analysis procedures. Markers are used in order to trace gradients of deformation over a sample and study other mechanical properties. Image processing and video sequence analysis can be a very powerful investigation tool and this paper shows preliminary results from its use on microtectonics. The proposed algorithm is a combination of two well-known approaches: feature extraction and block matching.
... These polygonal surfaces typically connect a few hundred 3D vertices (see Figure 2). Such articulated meshes are often used as generic models in model-based movement tracking systems [15,42] (see Figure 3). Candide [37] and Eisert's MPEG4 compliant articulated head [42]. ...
... Such articulated meshes are often used as generic models in model-based movement tracking systems [15,42] (see Figure 3). Candide [37] and Eisert's MPEG4 compliant articulated head [42]. ...
Article
Full-text available
This paper presents the main approaches used to synthesize talking faces, and provides greater detail on a handful of these approaches. An attempt is made to distinguish between facial synthesis itself (i.e. the manner in which facial movements are rendered on a computer screen), and the way these movements may be controlled and predicted using phonetic input. The two main synthesis techniques (model-based vs. image-based) are contrasted and presented by a brief description of the most illustrative existing systems. The challenging issues—evaluation, data acquisition and modeling—that may drive future models are also discussed and illustrated by our current work at ICP.
... These polygonal surfaces typically connect a few hundred 3D vertices (see Figure 2). Such articulated meshes are often used as generic models in model-based movement tracking systems [15,42] (see Figure 3). Candide [37] and Eisert's MPEG4 compliant articulated head [42]. ...
... Such articulated meshes are often used as generic models in model-based movement tracking systems [15,42] (see Figure 3). Candide [37] and Eisert's MPEG4 compliant articulated head [42]. ...
Article
Full-text available
This paper presents the main approaches used to synthesize talking faces, and provides greater detail on a handful of these approaches. No system is described exhaustively, however, and, for purposes of conciseness, not all existing systems are reviewed. An attempt is made to distinguish between facial synthesis itself (i.e the manner in which facial movements are rendered on a computer screen), and the way these movements may be controlled and predicted using phonetic input.
... Pattern matching has been widely used for estimating head motion [20]. Few projects [3] apply an analysisby-synthesis technique for recovering also facial movements because of the complexity of the forward model both in terms of geometry and texture. ...
Article
Full-text available
We present here the "labiophone", a virtual system for audio-visual speech communication. A clone of the speaker is animated at distance by articulatory movements extracted from the speaker's image and captured thanks to a video-camera centered on the speaker's face. The clone consists of a mesh driven by a few articulatory parameters and clothed by blended textures. The characteristics of the articulatory model and the textures blending are transmitted at the initiation of the dialog. Then only articulatory parameters are transmitted at a very low bit rate through the telecommunication or web network. Preliminary evaluation of such a system is presented below. Keywords: speech, facial animation, articulatory modelling, movement estimation, texture mapping. 1. INTRODUCTION Speech communication is multi-modal: if auditory and visual perception provide complementary information about the speaker and its emotional state, they collaborate intimately to enhance the intelligibility of the ...
Thesis
In teleconferencing applications, animated 3D heads can replace the usual video channels. This offers high compression opportunities, as well as the freedom of virtual spaces : one can compose a virtual place on screen, where 3D representations are debating. This thesis introduces an ad-hoc rendering algorithm, that can be applied to photorealistic 3D heads, at the expense of slightly limited viewing angles. Fast renderings are achieved on simple computers as well as 2D virtual machines. An automatic control architecture of the virtual cameras and the broadcasted view is also proposed. Cameras produce synthetic views, and react to speaking events. Switching between various (partial) camera views intends to let the debate look more attractive and intelligible. More simultaneous participants can take part, without their image size being lowered too much. The manual-interface-free automatic scheme enables the user to naturally talk and concentrate on the discussions. The previous parts have been implemented in a prototype. Several sound-scene scenarios have been experimented, playing with the image and sound association in a spatial-sound environment, where events from or outside the viewed area can be simulated. An hybrid video-based and 3D-based solution to the face-animating problem is defended as well. Partial images of eyes, mouth and eyebrows regions from live performance are inlaid on the clone surface texture. That way, it's up to the spectator to interpret the broadcasted video-like expressions. Another prototype has been built to test the real-time visual empathy.
Article
Abstract Inlooking forward to more natural telecollaboration, we can anticipate that the teleconferencing system,of the future will enable participants at distant locations to share the same virtual space. The visual object of each ,participant can be transmitted ,to the
Article
This paper presents a bibliography of nearly 1700 references related to computer vision and image analysis, arranged by subject matter. The topics covered include computational techniques; feature detection and segmentation; image and scene analysis; two-dimensional shape; pattern; color and texture; matching and stereo; three-dimensional recovery and analysis; three-dimensional shape; and motion. A few references are also given on related topics, including geometry and graphics, compression and processing, sensors and optics, visual perception, neural networks, artificial intelligence and pattern recognition, as well as on applications.
Article
Full-text available
Pour la téléconférence, on peut remplacer l'image des correspondants distants par des modèles 3D animés de leurs visages. En plus de taux de compression avantageux, cette approche offre les libertés du virtuel : on peut par exemple composer à l'écran l'impression d'un lieu unique, virtuel, où débattent les représentants 3D. Cette thèse présente un algorithme de rendu spécifique, applicable à des clones 3D photo-réalistes de visages. En restreignant les angles de vue autorisés, il permet un rendu simple et rapide, même avec des ordinateurs peu puissants ou sur des machines virtuelles. On propose aussi une architecture de régie automatique, reliée à des caméras virtuelles qui réagissent aux interventions des participants et en proposent une image synthétique. En alternant plusieurs vues (éventuellement partielles) de la scène, on autorise plus de participants simultanés, sans compromettre la qualité de l'image proposée ni l'intelligibilité du débat restitué. Automatique, cette approche libère l'utilisateur du système, qui peut se concentrer sur un débat rendu plus attractif, à la fois comme spectateur et comme participant. Cette thèse rend aussi compte de la réalisation d'un prototype de communication qui intègre les éléments précédents et permet de juger la qualité de la communication obtenue. À cette occasion, l'utilisation d'un environnement sonore qui intègre les interventions distantes et leur localisation (dans ou hors de l'image) est discutée, avec plusieurs expérimentations sur l'association entre l'image et le son spatialisé. Enfin, on introduit une solution hybride (3D et vidéo) pour animer les clones des visages. En incrustant à la surface d'un clone statique l'image des yeux, des sourcils et de la bouche vues par une caméra, on laisse aux spectateurs la responsabilité d'interpréter les expressions originales, dans toute leur dimension vidéo (forte résolution spatiale et temporelle). Un second prototype permet de juger de l'empathie visuelle.
Article
This paper addresses the issue of the reconstruction of intermediate views from a pair of stereoscopic images. Such a reconstruction is needed for the enhancement of depth perception in stereoscopic systems, e.g., "continuous look around" or adjustment of virtual camera baseline. The algorithm proposed here addresses the issue of blur; unlike typical reconstruction algorithms that perform averaging between disparity-compensated left and right images the new algorithm uses non-linear filtering via a winner-takes-all strategy. The image under reconstruction is assumed to be a tiling by fixed-size blocks that come from various positions of either the left or right images using disparity compensation. The tiling map is modeled by a binary decision field while the disparity model is based on a smoothness constraint. The models are combined through a maximum a posteriori probability (MAP) criterion. The intermediate intensities, disparities and the binary decision field are estimated jointly...