Analysis and QoE evaluation of cloud gaming service adaptation under different network conditions: The case of NVIDIA GeForce NOW

M Suznjevic, I Slivar… - 2016 Eighth International …, 2016 - ieeexplore.ieee.org
… In this paper, we conduct an analysis of the commercial NVIDIA GeForce NOW game
streaming platform adaptation mechanisms in light of variable network conditions. We further …

Building a Remote Laboratory Based on NVIDIA GeForce Experience and Moonlight Streaming

TK Linh, PD Hung - … : 18th International Conference, CDVE 2021, Virtual …, 2021 - Springer
… combination of labs and NVIDIA GeForce Experience and Moonlight Streaming. On the
server-side connected to the practical KITs, NVIDIA GeForce Experience enhances video quality. …

[PDF][PDF] HUMAN RE-IDENTIFICATION USING SIAMESE CONVOLUTIONAL NEURAL NETWORK ON NVIDIA GEFORCE RTX 2060

A ELAVARASAN - 2021 - eprints.utm.my
… orang dalam beberapa kamera pada platform NVIDIA® GeForce RTX ™ 2060, termasuk …
semula seseorang menggunakan SCNN pada platform NVIDIA® GeForce RTX ™ 2060. …

Faster matrix-vector multiplication on GeForce 8800GTX

N Fujimoto - 2008 IEEE International Symposium on Parallel …, 2008 - ieeexplore.ieee.org
Recently a GPU has acquired programmability to perform general purpose computation fast
by running ten thousands of threads concurrently. This paper presents a new algorithm for …

[PDF][PDF] Simulation Acceleration of Index Modulation using NVIDIA GeForce GTX 960M

U Yaqub, M Al-Mouhamed - To be Submitted, 2016 - researchgate.net
1Abstract—Massively parallel computing has been applied extensively in various scientific
and engineering domains. In this project we apply this concept to accelerating simulation of a …

[PDF][PDF] General-purpose sparse matrix building blocks using the NVIDIA CUDA technology platform

M Christen, O Schenk, H Burkhart - First workshop on general purpose …, 2007 - Citeseer
… We investigate the performance on the NVIDIA GeForce 8800 multicore chip. We exploit
the architectural features of the GeForce 8800 GPU to design an efficient GPU-parallel sparse …

A GPU‐based streaming algorithm for high‐resolution cloth simulation

M Tang, R Tong, R Narain, C Meng… - Computer Graphics …, 2013 - Wiley Online Library
… We have implemented our algorithm on three different commodity GPUs: a NVIDIA GeForce
GTX 580, a NVIDIA GeForce GTX 680, and a NVIDIA Tesla K20c. Their parameters are …

A GPU-based numerical model coupling hydrodynamical and morphological processes

J Hou, Y Kang, C Hu, Y Tong, B Pan, J Xia - International Journal of …, 2020 - Elsevier
… At the same time, a desktop with an NVIDIA GeForce GTX 1080Ti GPU was used to … NVIDIA
GeForce GTX 2080Ti GPU graphics computing was more efficient than the NVIDIA GeForce

GPU acceleration for FEM-based structural analysis

S Georgescu, P Chow, H Okuda - Archives of Computational Methods in …, 2013 - Springer
… in [54] is very limited, with a NVIDIA GeForce 8800 GTX being at best 55 % faster than an
Intel … generation NVIDIA GeForce GTX 580 GPU over an Intel Core i7 2600K CPU at 3.4 GHz. …

[HTML][HTML] Optimized GPU implementation of merck molecular force field and universal force field

Á Jász, Á Rák, I Ladjánszki, G Cserey - Journal of Molecular Structure, 2019 - Elsevier
… shows measurements with newer GPU hardware as well, an NVIDIA GTX Titan being
roughly 6.8 times faster than an NVIDIA GeForce 9800 GTX device, however, no recent …

Tiling for performance tuning on different models of GPUs

C Xu, SR Kirk, S Jenkins - 2009 Second International …, 2009 - ieeexplore.ieee.org
… Table I lists some differences between the NVidia GTX 260 and the NVidia GeForce 8800
GTS. Assume the scenario when the programmer optimizes the algorithm on the GTX 260; he …

Fast GPU-based CT reconstruction using the common unified device architecture (CUDA)

H Scherl, B Keck, M Kowarschik… - 2007 IEEE Nuclear …, 2007 - ieeexplore.ieee.org
… In comparison to the nine-way coherent Cell Broadband Engine Architecture, for which we
have previously demonstrated a substantial reconstruction speed [1], the NVIDIA GeForce

[PDF][PDF] GPU accelerated acoustic likelihood computations.

P Cardinal, P Dumouchel, G Boulianne… - …, 2008 - isca-archive.org
… In image processing, Erra [7] implemented fractal image compression algorithms on a
NVidia GeForce FX 6800 and obtained a speed-up of 280x over the equivalent CPUbased …

Fast face recognition approach using a graphical processing unit “GPU”

Y Ouerhani, M Jridi, A Alfalou - 2010 IEEE International …, 2010 - ieeexplore.ieee.org
… We can notice that the GPU Nvidia Geforce 8400 GS execution is 3 times faster than an
Intel Core 2 CPU 2.00 GHZ and 2.5 times faster than a Pentium Dual Core CPU 2.50 GHZ. An …

[PDF][PDF] Statistical power consumption analysis and modeling for GPU-based computing

X Ma, M Dong, L Zhong, Z Deng - Proceeding of ACM SOSP Workshop on …, 2009 - yecl.org
… In our experiment, the test computer was equipped with a NVidia GeForce 8800gt graphics
card with a 200 Watt power specification, AMD Athlon 64x2 3.0GHz Dual-Core Processor, …

Harvesting graphics power for MD simulations

JA Van Meel, A Arnold, D Frenkel… - Molecular …, 2008 - Taylor & Francis
… We tested our code on a modern GPU, the NVIDIA GeForce 8800 GTX. Results for two MD
algorithms suitable for short-ranged and long-ranged interactions, and a congruential shift …

An Implementation of LASER Beam Welding Simulation on Graphics Processing Unit Using CUDA

E Nascimento, E Magalhães, A Azevedo, LES Paes… - Computation, 2024 - mdpi.com
… Computational times were compared between GPU implementations using an NVIDIA
GeForce ® GTX ™ 1080 Ti and CPU implementations using the Intel ® Core ™ i7 8750H and the …

Porting a high-order finite-element earthquake modeling application to NVIDIA graphics cards using CUDA

D Komatitsch, D Michéa, G Erlebacher - Journal of Parallel and Distributed …, 2009 - Elsevier
… Our experimental setup is composed of an NVIDIA GeForce 8800 GTX card installed on the
… Linux kernel 2.6.23; and of an NVIDIA GeForce GTX 280 card installed on the PCI Express 1 …

[PDF][PDF] Real-Time Ray Tracing Using Nvidia OptiX.

H Ludvigsen, AC Elster - Eurographics (Short Papers), 2010 - diglib.eg.org
… However, the GPU used in [SCCC09] is a Nvidia Geforce 8800 GTS that is several generations
older and substantially slower than the Quadro FX 5800 used in our test bench. Again, …

Towards acceleration of fault simulation using graphics processing units

K Gulati, SP Khatri - Proceedings of the 45th Annual Design Automation …, 2008 - dl.acm.org
… Our results, implemented on a NVIDIA GeForce GTX 8800 GPU card, indicate that our
approach is on average 35× faster when compared to a commercial fault simulation engine. With …