標(biāo)題: Titlebook: Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction; COST Action 2102 Int Anna Esposito,Nikolaos G. Bourbakis,Ioanni [打印本頁(yè)] 作者: Spring 時(shí)間: 2025-3-21 18:00
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction影響因子(影響力)
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction影響因子(影響力)學(xué)科排名
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction網(wǎng)絡(luò)公開(kāi)度
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction網(wǎng)絡(luò)公開(kāi)度學(xué)科排名
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction被引頻次
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction被引頻次學(xué)科排名
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction年度引用
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction年度引用學(xué)科排名
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction讀者反饋
書目名稱Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction讀者反饋學(xué)科排名
作者: 發(fā)酵劑 時(shí)間: 2025-3-21 21:08
Ekfrasis: A Formal Language for Representing and Generating Sequences of Facial Patterns for Studyinsis) as a software methodology that synthesizes (or generates) automatically various facial expressions by appropriately combining facial features. The main objective here is to use this methodology to generate various combinations of facial expressions and study if these combinations efficiently represent emotional behavioral patterns.作者: commensurate 時(shí)間: 2025-3-22 03:34 作者: Presbycusis 時(shí)間: 2025-3-22 05:34
Study on Speaker-Independent Emotion Recognition from Speech on Real-World Data classifiers on utterance level is applied, in attempt to improve the performance of the emotion recognizer. Experimental results demonstrate significant differences on recognizing emotions on acted/real-world speech.作者: onlooker 時(shí)間: 2025-3-22 09:48 作者: MORT 時(shí)間: 2025-3-22 13:43 作者: 巧辦法 時(shí)間: 2025-3-22 18:11
Towards Slovak Broadcast News Automatic Recording and Transcribing Servicelso all automatically extracted metadata (verbal and nonverbal), and also to select incorrectly automatically identified data. The architecture of the present system is linear, which means every module starts after the previous has finished the data processing.作者: Gyrate 時(shí)間: 2025-3-22 21:31 作者: TEM 時(shí)間: 2025-3-23 01:36
Combining Features for Recognizing Emotional Facial Expressions in Static Images set was obtained combining PCA and LDA features (93% of correct recognition rate), whereas, combining PCA, LDA and Gabor filter features the net gave 94% of correct classification on facial expressions of subjects not included in the training set.作者: Bombast 時(shí)間: 2025-3-23 09:26
Expressive Speech Synthesis Using Emotion-Specific Speech Inventoriesl for 99% of the logatoms and for all natural sentences. Recognition rates significantly above chance level were obtained for each emotion. The recognition rate for some synthetic sentences exceeded that of natural ones.作者: inhibit 時(shí)間: 2025-3-23 10:35 作者: 嘲笑 時(shí)間: 2025-3-23 16:09 作者: endocardium 時(shí)間: 2025-3-23 19:30
On the Relevance of Facial Expressions for Biometric Recognition use the Japanese Female Facial Expression Database (JAFFE) in order to evaluate the influence of facial expression in biometric recognition rates. In our experiments we used a nearest neighbor classifier with different number of training samples, different error criteria, and several feature extrac作者: Regurgitation 時(shí)間: 2025-3-24 00:09 作者: 釘牢 時(shí)間: 2025-3-24 05:23 作者: incredulity 時(shí)間: 2025-3-24 07:51 作者: 抵消 時(shí)間: 2025-3-24 14:13 作者: 擁擠前 時(shí)間: 2025-3-24 16:03 作者: 羞辱 時(shí)間: 2025-3-24 21:07
The Organization of a Neurocomputational Control Model for Articulatory Speech Synthesis of neurophysiology and cognitive psychology. Thus it is based on such neural control circuits, neural maps and mappings as are hypothesized to exist in the human brain, and the model is based on learning or training mechanisms similar to those occurring during the human process of speech acquisitio作者: Soliloquy 時(shí)間: 2025-3-24 23:50
Automatic Speech Recognition Used for Intelligibility Assessment of Text-to-Speech Systemst of general speech processing algorithms is proposed. It is based on automatic recognition methods developed for discrete and fluent speech processing. The idea is illustrated on two case studies: a) comparison of listening evaluation of Czech rhyme tests with automatic discrete speech recognition 作者: Grasping 時(shí)間: 2025-3-25 06:50 作者: 橡子 時(shí)間: 2025-3-25 09:07 作者: 持久 時(shí)間: 2025-3-25 12:29 作者: 重畫只能放棄 時(shí)間: 2025-3-25 16:53 作者: 亂砍 時(shí)間: 2025-3-25 21:22
Application of Expressive Speech in TTS System with Cepstral Descriptionts with storytelling speaking style have been performed. This particular speaking style is suitable for applications aimed at children as well as special applications aimed at blind people. Analyzing human storytellers’ speech, we designed a set of prosodic parameters prototypes for converting speec作者: 連累 時(shí)間: 2025-3-26 02:19 作者: Abbreviate 時(shí)間: 2025-3-26 04:56
Expressive Speech Synthesis Using Emotion-Specific Speech Inventoriesnd 26 logatoms containing all the diphones and CVC triphones necessary to synthesize the same sentence. The speech material was produced by a professional actress expressing all logatoms and the sentence with the six basic emotions and in neutral tone. 7 emotion-dependent inventories were constructe作者: 用手捏 時(shí)間: 2025-3-26 09:09
Study on Speaker-Independent Emotion Recognition from Speech on Real-World Dataformed towards examining the behavior of a detector of negative emotional states over non-acted/acted speech. Furthermore, a score-level fusion of two classifiers on utterance level is applied, in attempt to improve the performance of the emotion recognizer. Experimental results demonstrate signific作者: 壓倒 時(shí)間: 2025-3-26 14:38 作者: 咯咯笑 時(shí)間: 2025-3-26 18:25
Conference proceedings 2008wed contributions of the p- ticipants at the COST 2102 International Conference on Verbal and Nonverbal F- tures of Human–Human and Human–Machine Interaction, held in Patras, Greece, October 29–31, 2007, hosted by the 19th IEEE International Conference on Tools with Artificial Intelligence (ICTAI 20作者: Employee 時(shí)間: 2025-3-26 22:17 作者: Astigmatism 時(shí)間: 2025-3-27 01:44 作者: mosque 時(shí)間: 2025-3-27 06:01 作者: 代理人 時(shí)間: 2025-3-27 11:57 作者: admission 時(shí)間: 2025-3-27 14:14
Lecture Notes in Computer Sciencehttp://image.papertrans.cn/v/image/981106.jpg作者: Trabeculoplasty 時(shí)間: 2025-3-27 20:17
https://doi.org/10.1007/978-3-540-70872-8biometric; data mining; emotion recognition; facial expressions; facial patterns; gestures; hci; multimodal作者: 走路左晃右晃 時(shí)間: 2025-3-27 22:11
Computational Stylometry: Who’s in a Play?e works of four sample playwrights that are freely available in machine-readable form. Strong characters are those whose speeches constitute homogeneous categories in comparison with other characters—their speeches are more attributable to themselves than to their play or their author.作者: Circumscribe 時(shí)間: 2025-3-28 03:21
978-3-540-70871-1Springer-Verlag Berlin Heidelberg 2008作者: 控制 時(shí)間: 2025-3-28 07:14
Verbal and Nonverbal Features of Human-Human and Human-Machine Interaction978-3-540-70872-8Series ISSN 0302-9743 Series E-ISSN 1611-3349 作者: 冬眠 時(shí)間: 2025-3-28 13:54 作者: ENACT 時(shí)間: 2025-3-28 17:15 作者: 尖酸一點(diǎn) 時(shí)間: 2025-3-28 21:49 作者: 高談闊論 時(shí)間: 2025-3-29 02:35
Robert Vích,Jan Nouza,Martin Vondra Untersuchungen, die speziell auf die Probleme des Stra?enverkehrs abgestimmt sind, abgerundet. Im Rahmen des vorliegenden Forschungsauftrages werden diese Untersuchungen auf Nordrhein-Westfalen begrenzt. Es dürfte keine Schwierigkeiten bereiten, das Verfahren auch auf das übrige Bundesgebiet anzuwe作者: 致敬 時(shí)間: 2025-3-29 03:14 作者: 有幫助 時(shí)間: 2025-3-29 08:46 作者: GIDDY 時(shí)間: 2025-3-29 12:28 作者: 按等級(jí) 時(shí)間: 2025-3-29 17:42
Szabolcs Levente Tóth,David Sztahó,Klára Vicsieit macht (z.B.: Die Erkenntnis, da? nicht alles mit einem einzigen Knopfdruck automatisch abl?uft und das System erheblichen Aufwand erfordert), schwindet das Interesse bei den Anwendern. Die zuvor gepriesene Innovation gilt nur noch als ?nützliches Spielzeug“: Die Systeme werden nur noch sporadisc作者: 良心 時(shí)間: 2025-3-29 20:27 作者: COMMA 時(shí)間: 2025-3-30 02:44
Marian Bartlett,Gwen Littlewort,Esra Vural,Kang Lee,Mujdat Cetin,Aytul Ercil,Javier Movellan作者: CRAB 時(shí)間: 2025-3-30 07:02 作者: Collision 時(shí)間: 2025-3-30 11:58 作者: 熒光 時(shí)間: 2025-3-30 14:58
Verbal and Nonverbal Features of Human-Human and Human-Machine InteractionCOST Action 2102 Int作者: 該得 時(shí)間: 2025-3-30 17:25
Anna Esposito,Nikolaos G. Bourbakis,Ioannis Hatzil作者: Esalate 時(shí)間: 2025-3-31 00:08 作者: 提煉 時(shí)間: 2025-3-31 04:08
Data Mining Spontaneous Facial Behavior with Automatic Expression Codingmovements. Automated classifiers were able to differentiate real from fake pain significantly better than na?ve human subjects, and to detect critical drowsiness above 98% accuracy. ?Issues for application of machine learning systems to facial expression analysis are discussed.作者: conceal 時(shí)間: 2025-3-31 05:38 作者: GNAW 時(shí)間: 2025-3-31 11:18
The Organization of a Neurocomputational Control Model for Articulatory Speech Synthesisntly capable of generating acoustic speech signals by controlling an articu latory-acoustic vocal tract model. The module developed thus far is capable of producing single sounds (vowels and consonants), simple CV- and VC-syllables, and first sample words. In addition, processes of human-human inter作者: Canary 時(shí)間: 2025-3-31 13:59
ECESS Platform for Web Based TTS Modules and Systems Evaluationessary modules of other partners. By using the RES client they could also build-up a complete TTS system via web, without using any of their own modules. Several partners can contribute their modules, even with the same functionality, and it is easy to add a new module to the whole web-based distrib作者: Gleason-score 時(shí)間: 2025-3-31 19:29 作者: Promotion 時(shí)間: 2025-3-31 21:54 作者: AMBI 時(shí)間: 2025-4-1 02:32 作者: ALLAY 時(shí)間: 2025-4-1 06:06 作者: 恃強(qiáng)凌弱的人 時(shí)間: 2025-4-1 13:26
Zsófia Ruttkay,Rieks op den Akkerund Nachteile der Publizit?tspflicht kaum analysiert. Man unterstellt weitgehend bestimmte unterneh- merische Reaktionen auf die Ver?ffentlichung betrieblicher Sachverhalte als mehr oder weniger offensichtlich und diskutiert im wesentlichen nur darüber, inwieweit die Folgen dieser Reaktionen unter e作者: coalition 時(shí)間: 2025-4-1 15:47
Robert Vích,Jan Nouza,Martin Vondrande, die z.B. dem Klima-Atlas [62] entnommen werden k?nnte, sondern auch die jeweils zugeh?rige Windst?rke von Belang. Es liegt nun nahe, neben den üblichen Angaben von St?rke und Richtung Windrosen aufzustellen, in denen nicht die Windrichtung an sich, sondern gewogen nach der jeweiligen Windst?rke