標(biāo)題: Titlebook: Computer Vision, Imaging and Computer Graphics Theory and Applications; 15th International J Kadi Bouatouch,A. Augusto de Sousa,Jose Braz C [打印本頁(yè)] 作者: 不同 時(shí)間: 2025-3-21 17:17
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications影響因子(影響力)
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications影響因子(影響力)學(xué)科排名
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications網(wǎng)絡(luò)公開(kāi)度
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications網(wǎng)絡(luò)公開(kāi)度學(xué)科排名
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications被引頻次
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications被引頻次學(xué)科排名
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications年度引用
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications年度引用學(xué)科排名
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications讀者反饋
書目名稱Computer Vision, Imaging and Computer Graphics Theory and Applications讀者反饋學(xué)科排名
作者: Priapism 時(shí)間: 2025-3-21 21:59 作者: Onerous 時(shí)間: 2025-3-22 00:46 作者: JOT 時(shí)間: 2025-3-22 05:33 作者: 公社 時(shí)間: 2025-3-22 11:43 作者: glacial 時(shí)間: 2025-3-22 15:22
Scene Text Localization Using Lightweight Convolutional Networksing the fusion of two lightweight neural network architectures, MobileNetV2 and Single Shot Detector (SSD), into our approach named MobText. As experimental results in the ICDAR’11 and ICDAR’13 datasets demonstrates that our solution yields the best trade-off between effectiveness and efficiency in 作者: glacial 時(shí)間: 2025-3-22 19:26
Early Stopping for Two-Stream Fusion Applied to Action Recognitionual networks. In addition to reducing overfitting, this method also reduces the training cost, since it usually requires fewer epochs to complete the classification process. Experiments are conducted on UCF101 and HMDB51 datasets, which are two challenging benchmarks in the context of action recogni作者: 脾氣暴躁的人 時(shí)間: 2025-3-22 23:52 作者: 哎呦 時(shí)間: 2025-3-23 03:39
https://doi.org/10.1057/9781137325129rning to automatically construct projections from the extracted features. We show that our three variations of deep learning, which jointly treat feature extraction, selection, and projection, allow efficient creation of high-quality visual overviews of large shape collections, require minimal user 作者: Gum-Disease 時(shí)間: 2025-3-23 08:11
https://doi.org/10.1057/9781403983213 humanities may improve and support the daily workflows of humanists. Despite the generally low level of digitization, multiple projects already build a basis to help in digitizing the field. As an example, the . project collected a vast amount of musicological data throughout the last 16 years and 作者: GLIDE 時(shí)間: 2025-3-23 13:06 作者: 帶來(lái)的感覺(jué) 時(shí)間: 2025-3-23 14:18
https://doi.org/10.1057/9780230599116tion performance only drops from . to . while . of the label information is removed. The flexibility of the detector with respect to the different levels of details in label definitions is investigated for a traffic surveillance application, involving public and proprietary datasets with non-overlap作者: GOAD 時(shí)間: 2025-3-23 18:15
https://doi.org/10.1057/9780230599116ing the fusion of two lightweight neural network architectures, MobileNetV2 and Single Shot Detector (SSD), into our approach named MobText. As experimental results in the ICDAR’11 and ICDAR’13 datasets demonstrates that our solution yields the best trade-off between effectiveness and efficiency in 作者: 詩(shī)集 時(shí)間: 2025-3-24 00:40
A Changing Focus: The Invalided Men,ual networks. In addition to reducing overfitting, this method also reduces the training cost, since it usually requires fewer epochs to complete the classification process. Experiments are conducted on UCF101 and HMDB51 datasets, which are two challenging benchmarks in the context of action recogni作者: Complement 時(shí)間: 2025-3-24 03:50 作者: Narcissist 時(shí)間: 2025-3-24 08:24 作者: GROSS 時(shí)間: 2025-3-24 10:56 作者: 碎石頭 時(shí)間: 2025-3-24 17:09
Computer Vision, Imaging and Computer Graphics Theory and Applications15th International J作者: 行乞 時(shí)間: 2025-3-24 21:17 作者: CROW 時(shí)間: 2025-3-25 01:54
Conference proceedings 2022phics Theory and Applications, VISIGRAPP 2020, held in Valletta, Malta, in February 2020..The 25 thoroughly revised and extended papers presented in this volume were carefully reviewed and selected from 455 submissions. The papers contribute to the understanding of relevant trends of current researc作者: Generic-Drug 時(shí)間: 2025-3-25 04:17 作者: 宏偉 時(shí)間: 2025-3-25 07:42 作者: 使熄滅 時(shí)間: 2025-3-25 12:36 作者: PLAYS 時(shí)間: 2025-3-25 19:18
The Belief-Desire Theory of Emotionskeleton-based rotation reduces task completion times and increases user satisfaction, while not introducing additional costs, being thus an interesting addition to the palette of 3D manipulation tools.作者: 寬度 時(shí)間: 2025-3-25 22:17 作者: OPINE 時(shí)間: 2025-3-26 00:35 作者: 有權(quán)威 時(shí)間: 2025-3-26 06:50 作者: 相反放置 時(shí)間: 2025-3-26 09:45 作者: 不能平靜 時(shí)間: 2025-3-26 15:46
RGB-D Images Based 3D Plant Growth Prediction by Sequential Images-to-Images Translation with Plant dule in the network. We verified how the prediction accuracy was affected by the internal structure of the network. In addition, the extension of our network with plant priors was further investigated to evaluate the impact for plant growth prediction tasks.作者: 口音在加重 時(shí)間: 2025-3-26 17:08
The Belief-Desire Theory of Emotionsrithm (GA) for convex polytope generation. It directly transforms 3D point clouds or triangle meshes into solid primitives. The filtered primitive set is then used as input for a GA-based CSG extraction stage. We evaluate two different CSG extraction methodologies and furthermore compare our pipeline to current state-of-the-art methods.作者: 表被動(dòng) 時(shí)間: 2025-3-26 23:42
CSG Tree Extraction from 3D Point Clouds and Meshes Using a Hybrid Approachrithm (GA) for convex polytope generation. It directly transforms 3D point clouds or triangle meshes into solid primitives. The filtered primitive set is then used as input for a GA-based CSG extraction stage. We evaluate two different CSG extraction methodologies and furthermore compare our pipeline to current state-of-the-art methods.作者: 未開(kāi)化 時(shí)間: 2025-3-27 03:45
Intention Understanding for Human-Aware Mobile Robots: Comparing Cues and the Effect of Demographicshree lighting schemes and tested them out in an online experiment. We found that signals resembling automotive signaling work the best also for logistic mobile robots. We further find that people’s opinion of these signaling methods will be influenced by their demographic background (gender, age).作者: AVID 時(shí)間: 2025-3-27 06:00 作者: HERE 時(shí)間: 2025-3-27 13:02 作者: SMART 時(shí)間: 2025-3-27 16:14 作者: ANN 時(shí)間: 2025-3-27 21:05
Intention Understanding for Human-Aware Mobile Robots: Comparing Cues and the Effect of Demographicsret what the robot’s intentions are. This is especially important when a robot is driving down a crowded corridor. It is essential for people in its vicinity to understand which way the robot wants to go next. To explore what signals are the best for conveying its intention to turn, we implemented t作者: 分解 時(shí)間: 2025-3-28 00:58 作者: MARS 時(shí)間: 2025-3-28 04:41
Scalable Visual Exploration of?3D Shape Databases via?Feature Synthesis and?Selectionensionality reduction of feature vectors extracted from shape descriptions. We address the problem of feature extraction by exploring both combinations of hand-engineered geometric features and using the latent feature vectors generated by a deep learning classification method, and discuss the compa作者: 廢止 時(shí)間: 2025-3-28 09:25 作者: 為寵愛(ài) 時(shí)間: 2025-3-28 14:30 作者: metropolitan 時(shí)間: 2025-3-28 18:35 作者: 牽索 時(shí)間: 2025-3-28 20:16 作者: 用手捏 時(shí)間: 2025-3-29 02:50 作者: 轎車 時(shí)間: 2025-3-29 05:46
RGB-D Images Based 3D Plant Growth Prediction by Sequential Images-to-Images Translation with Plant nd an existing image-to-image translation technique based on U-Net to images-to-images translation by incorporating convLSTM into skip connections in U-Net. With this architecture, we can achieve sequential image prediction tasks such that future images are predicted from several past ones. Since de作者: 鑲嵌細(xì)工 時(shí)間: 2025-3-29 08:24
Computer Vision, Imaging and Computer Graphics Theory and Applications978-3-030-94893-1Series ISSN 1865-0929 Series E-ISSN 1865-0937 作者: 臭了生氣 時(shí)間: 2025-3-29 13:32 作者: 美學(xué) 時(shí)間: 2025-3-29 18:01 作者: recede 時(shí)間: 2025-3-29 22:22 作者: 輪流 時(shí)間: 2025-3-30 03:11
Evaluative Theories of Emotionsorms of interactivity, allowing intuitive user interactions with ubiquitous services in mixed environments. However, they require to synchronize multiple platforms and various technologies. Their heterogeneity makes them complex, and sparsely interoperable or extensible. Therefore, designers and dev作者: 大包裹 時(shí)間: 2025-3-30 06:13 作者: HACK 時(shí)間: 2025-3-30 10:34 作者: antecedence 時(shí)間: 2025-3-30 16:27
The Emotions in Hellenistic Philosophyret what the robot’s intentions are. This is especially important when a robot is driving down a crowded corridor. It is essential for people in its vicinity to understand which way the robot wants to go next. To explore what signals are the best for conveying its intention to turn, we implemented t作者: 大喘氣 時(shí)間: 2025-3-30 20:35
The Emotions in Hellenistic Philosophyver, since the creation and sharing of images is increasing exponentially, images as a media form suffer from being devalued, as the quality of single images are getting less and less important, and the frequency of the shared content turns to be the focus. In this work, an interactive system which 作者: 青少年 時(shí)間: 2025-3-30 23:47 作者: 根除 時(shí)間: 2025-3-31 02:12