CUDA is an abbreviation for Compute Unified Device Architecture, developed by Nvidia company. It has two main features, parallel computing and Application Programming Interface (API). CUDA core has the capability of breaking down or splitting an extensive calculation into many small parts and then performing all these small calculations simultaneously (also known as parallel processing). Many of such processing rules are bunched together as commands to enable communication between different software components of a device (also known as API), which can be used to build custom software and can provide access to the parallel computational elements.

    In computers, a Graphics Processing Unit (GPU) renders and displays graphics content on a computer screen. A core is a physical processing unit inside the GPU that executes CUDA API and works simultaneously (or in parallel) with other processing units (or cores) to process and display an image or video pixels to the screen that humans can see. GPU cores are optimized to do more calculations as a group. When hundreds and thousands of these cores are combined to form a GPU that supports CUDA developed by Nvidia, it is known as the CUDA core.

    CUDA cores process and render image and video data in parallel to each other, meaning that each core works on a separate task simultaneously. It makes a highly efficient system, where no single core is waiting for any other core to finish its task. Therefore, more cores mean more parallel processing and more multitasking capability, and therefore, more individual tasks can be done in a given time, leading to faster performance and overall efficiency.

    « Back to Definition Index