Results 11 to 20 of about 16,570 (263)
Automatic Modulation Classification (AMC) is a technique used in wireless communication systems to identify the modulation type of received signals at the receiver.
Kuchul Jung +2 more
doaj +1 more source
CUDA Memory Optimizations for Large Data-Structures in the Gravit Simulator
Modern GPUs open a completely new field to optimize embarrassingly parallel algorithms. Implementing an algorithm on a GPU confronts the programmer with a new set of challenges for program optimization.
Jakob Siegel +2 more
doaj +1 more source
Metrics, Noise Propagation Models, and Design Framework for Floating-Point Approximate Computing
Approximate computing has emerged as an efficient solution for energy saving at the expense of calculation accuracy, especially for floating-point operation intensive applications, which have urgent demands for some uniform design frameworks for floating-
Yiyao Xiang +4 more
doaj +1 more source
Digital filter optimization for C language
A method for transforming C code with floating-point values into C code with integer variables is developed. The objective is to avoid any operations with floating-point data types, thereby increasing the execution speed of the program on a ...
BARLEANU, A., BAITOIU, V., STAN, A.
doaj +1 more source
Algorithm for output of floating-point numbers in fixed-point form
Input/output of data items and information are very common operations performed on computer systems. Various data types require special ways to perform these operations on them. Floating-point numbers are no exceptions.
CB Ndeekor
doaj +1 more source
Optimizing floating point operations in Scheme [PDF]
Summary: It is well known that dynamic typing in languages like Lisp is costly in terms of performance. Besides the cost of tag checking, the other major source of inefficiency comes from the need to place and retrieve data from dynamically allocated objects, i.e. boxing and unboxing.
openaire +2 more sources
Numerical behavior of NVIDIA tensor cores [PDF]
We explore the floating-point arithmetic implemented in the NVIDIA tensor cores, which are hardware accelerators for mixed-precision matrix multiplication available on the Volta, Turing, and Ampere microarchitectures.
Massimiliano Fasi +3 more
doaj +2 more sources
BitFlow-Net: Toward Fully Binarized Convolutional Neural Networks
Binarization can greatly compress and accelerate deep convolutional neural networks (CNNs) for real-time industrial applications. However, existing binarized CNNs (BCNNs) rely on scaling factor (SF) and batch normalization (BatchNorm) that still involve ...
Lijun Wu +6 more
doaj +1 more source
An Efficient Implementation of the Sign LMS Algorithm Using Block Floating Point Format
An efficient scheme is presented for implementing the sign LMS algorithm in block floating point format, which permits processing of data over a wide dynamic range at a processor complexity and cost as low as that of a fixed point processor.
Lee Moon Ho +2 more
doaj +2 more sources
The rapid advancement in AI requires efficient accelerators for training on edge devices, which often face challenges related to the high hardware costs of floating-point arithmetic operations.
Muhammad Junaid +5 more
doaj +1 more source

