blob: a035a34c0d29294599e4d60fb892dfcb35fbab45 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.sysds.runtime.matrix.data;
import static jcuda.jcusparse.JCusparse.*;
import static jcuda.runtime.JCuda.cudaMemcpy;
import static jcuda.runtime.cudaMemcpyKind.cudaMemcpyDeviceToHost;
import static jcuda.runtime.cudaMemcpyKind.cudaMemcpyHostToDevice;
import java.nio.ByteBuffer;
import java.nio.ByteOrder;
import java.nio.FloatBuffer;
import java.util.stream.IntStream;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.sysds.api.DMLScript;
import org.apache.sysds.runtime.instructions.gpu.context.GPUContext;
import org.apache.sysds.utils.GPUStatistics;
import jcuda.Pointer;
import jcuda.Sizeof;
import jcuda.jcublas.JCublas2;
import jcuda.jcublas.cublasHandle;
import jcuda.jcusolver.JCusolverDn;
import jcuda.jcusolver.cusolverDnHandle;
import jcuda.jcusparse.JCusparse;
import jcuda.jcusparse.cusparseHandle;
import jcuda.jcusparse.cusparseMatDescr;
import jcuda.jcusparse.cusparseSpMatDescr;
import jcuda.jcusparse.cusparseSpGEMMDescr;
import static jcuda.jcusparse.cusparseIndexType.CUSPARSE_INDEX_32I;
import static jcuda.cudaDataType.CUDA_R_32F;
import static jcuda.jcusparse.cusparseStatus.CUSPARSE_STATUS_SUCCESS;
import static jcuda.jcusparse.cusparseSpMVAlg.CUSPARSE_SPMV_ALG_DEFAULT;
import static jcuda.jcusparse.cusparseOrder.CUSPARSE_ORDER_COL;
import static jcuda.jcusparse.cusparseSpMMAlg.CUSPARSE_SPMM_ALG_DEFAULT;
import static jcuda.jcusparse.cusparseCsr2CscAlg.CUSPARSE_CSR2CSC_ALG1;
import static jcuda.jcusparse.cusparseSparseToDenseAlg.CUSPARSE_SPARSETODENSE_ALG_DEFAULT;
import static jcuda.jcusparse.cusparseDenseToSparseAlg.CUSPARSE_DENSETOSPARSE_ALG_DEFAULT;
import static jcuda.runtime.JCuda.cudaMalloc;
import static jcuda.runtime.JCuda.cudaFree;
import jcuda.jcusparse.cusparseDnVecDescr;
import jcuda.jcusparse.cusparseDnMatDescr;
public class SinglePrecisionCudaSupportFunctions implements CudaSupportFunctions {
private static final Log LOG = LogFactory.getLog(SinglePrecisionCudaSupportFunctions.class.getName());
@Override
public int cusparsecsrgemm(cusparseHandle handle, int transA, int transB, int alg, cusparseSpMatDescr spMatDescrA,
cusparseSpMatDescr spMatDescrB, cusparseSpMatDescr spMatDescrC, cusparseSpGEMMDescr spgemmDescr) {
double[] alpha = {1.0}, beta = {0.0};
Pointer alphaPtr = Pointer.to(alpha), betaPtr = Pointer.to(beta);
return cusparseSpGEMM_copy(handle, transA, transB, alphaPtr, spMatDescrA.asConst(), spMatDescrB.asConst(),
betaPtr, spMatDescrC, CUDA_R_32F, alg, spgemmDescr);
}
@Override
public int cublasgeam(cublasHandle handle, int transa, int transb, int m, int n, Pointer alpha, Pointer A, int lda,
Pointer beta, Pointer B, int ldb, Pointer C, int ldc) {
return JCublas2.cublasSgeam(handle, transa, transb, m, n, alpha, A, lda, beta, B, ldb, C, ldc);
}
@Override
public int cusparsecsrmv(cusparseHandle handle, int transA, int m, int n, int nnz, Pointer alpha,
cusparseSpMatDescr spMatDescrA, cusparseMatDescr descrA, Pointer csrValA, Pointer csrRowPtrA,
Pointer csrColIndA, Pointer x, Pointer beta, Pointer y) {
// Create sparse matrix A in CSR format
int idxBase = cusparseGetMatIndexBase(descrA);
int dataType = CUDA_R_32F;
cusparseCreateCsr(spMatDescrA, m, n, nnz, csrRowPtrA, csrColIndA, csrValA, CUSPARSE_INDEX_32I,
CUSPARSE_INDEX_32I, idxBase, dataType);
// Create dense vectors vecX and vecY
cusparseDnVecDescr vecX = new cusparseDnVecDescr();
cusparseDnVecDescr vecY = new cusparseDnVecDescr();
cusparseCreateDnVec(vecX, n, x, dataType);
cusparseCreateDnVec(vecY, m, y, dataType);
// allocate an external buffer if needed
long[] bufferSize = {0};
int alg = CUSPARSE_SPMV_ALG_DEFAULT;
cusparseSpMV_bufferSize(handle, transA, alpha, spMatDescrA.asConst(), vecX.asConst(), beta, vecY, dataType, alg,
bufferSize);
// execute SpMV
Pointer dBuffer = new Pointer();
if(bufferSize[0] > 0)
cudaMalloc(dBuffer, bufferSize[0]);
try {
return cusparseSpMV(handle, transA, alpha, spMatDescrA.asConst(), vecX.asConst(), beta, vecY, dataType, alg,
dBuffer);
}
finally {
if(bufferSize[0] > 0)
cudaFree(dBuffer);
cusparseDestroyDnVec(vecX.asConst());
cusparseDestroyDnVec(vecY.asConst());
}
}
@Override
public int cusparsecsrmm2(cusparseHandle handle, int transA, int transB, int m, int n, int k, int nnz,
Pointer alpha, cusparseMatDescr descrA, cusparseSpMatDescr spMatDescrA, Pointer csrValA, Pointer csrRowPtrA,
Pointer csrColIndA, Pointer B, int ldb, Pointer beta, Pointer C, int ldc) {
int dataType = CUDA_R_32F;
int idxBase = cusparseGetMatIndexBase(descrA);
// Create sparse matrix A in CSR format
cusparseCreateCsr(spMatDescrA, m, n, nnz, csrRowPtrA, csrColIndA, csrValA, CUSPARSE_INDEX_32I,
CUSPARSE_INDEX_32I, idxBase, dataType);
// Create dense matrix B
cusparseDnMatDescr dnMatB = new cusparseDnMatDescr();
cusparseCreateDnMat(dnMatB, k, n, ldb, B, dataType, CUSPARSE_ORDER_COL);
// Create dense matrix C
cusparseDnMatDescr dnMatC = new cusparseDnMatDescr();
cusparseCreateDnMat(dnMatC, m, n, ldc, C, dataType, CUSPARSE_ORDER_COL);
// allocate an external buffer if needed
long[] bufferSize = {0};
int alg = CUSPARSE_SPMM_ALG_DEFAULT;
cusparseSpMM_bufferSize(handle, transA, transB, alpha, spMatDescrA.asConst(), dnMatB.asConst(), beta, dnMatC,
dataType, alg, bufferSize);
// execute SpMM
Pointer dBuffer = new Pointer();
if(bufferSize[0] > 0)
cudaMalloc(dBuffer, bufferSize[0]);
try {
return cusparseSpMM(handle, transA, transB, alpha, spMatDescrA.asConst(), dnMatB.asConst(), beta, dnMatC,
dataType, alg, dBuffer);
}
finally {
if(bufferSize[0] > 0)
cudaFree(dBuffer);
cusparseDestroySpMat(spMatDescrA.asConst());
cusparseDestroyDnMat(dnMatB.asConst());
cusparseDestroyDnMat(dnMatC.asConst());
}
}
@Override
public int cublasdot(cublasHandle handle, int n, Pointer x, int incx, Pointer y, int incy, Pointer result) {
return JCublas2.cublasSdot(handle, n, x, incx, y, incy, result);
}
@Override
public int cublasgemv(cublasHandle handle, int trans, int m, int n, Pointer alpha, Pointer A, int lda, Pointer x,
int incx, Pointer beta, Pointer y, int incy) {
return JCublas2.cublasSgemv(handle, trans, m, n, alpha, A, lda, x, incx, beta, y, incy);
}
@Override
public int cublasgemm(cublasHandle handle, int transa, int transb, int m, int n, int k, Pointer alpha, Pointer A,
int lda, Pointer B, int ldb, Pointer beta, Pointer C, int ldc) {
return JCublas2.cublasSgemm(handle, transa, transb, m, n, k, alpha, A, lda, B, ldb, beta, C, ldc);
}
@Override
public int cusparsecsr2csc(cusparseHandle handle, int m, int n, int nnz, Pointer csrVal, Pointer csrRowPtr,
Pointer csrColInd, Pointer cscVal, Pointer cscRowInd, Pointer cscColPtr, int copyValues, int idxBase) {
int valType = CUDA_R_32F; // single precision
int alg = CUSPARSE_CSR2CSC_ALG1; // always supported
long[] bufferSize = {0};
cusparseCsr2cscEx2_bufferSize(handle, m, n, nnz, csrVal, csrRowPtr, csrColInd, cscVal, cscColPtr, cscRowInd,
valType, copyValues, idxBase, alg, bufferSize);
Pointer buffer = new Pointer();
if(bufferSize[0] > 0)
cudaMalloc(buffer, bufferSize[0]);
try {
return cusparseCsr2cscEx2(handle, m, n, nnz, csrVal, csrRowPtr, csrColInd, cscVal, cscColPtr, cscRowInd,
valType, copyValues, idxBase, alg, buffer);
}
finally {
if(bufferSize[0] > 0)
cudaFree(buffer);
}
}
@Override
public int cublassyrk(cublasHandle handle, int uplo, int trans, int n, int k, Pointer alpha, Pointer A, int lda,
Pointer beta, Pointer C, int ldc) {
return JCublas2.cublasSsyrk(handle, uplo, trans, n, k, alpha, A, lda, beta, C, ldc);
}
@Override
public int cublasaxpy(cublasHandle handle, int n, Pointer alpha, Pointer x, int incx, Pointer y, int incy) {
return JCublas2.cublasSaxpy(handle, n, alpha, x, incx, y, incy);
}
@Override
public int cublastrsm(cublasHandle handle, int side, int uplo, int trans, int diag, int m, int n, Pointer alpha,
Pointer A, int lda, Pointer B, int ldb) {
return JCublas2.cublasStrsm(handle, side, uplo, trans, diag, m, n, alpha, A, lda, B, ldb);
}
@Override
public int cusolverDngeqrf_bufferSize(cusolverDnHandle handle, int m, int n, Pointer A, int lda, int[] Lwork) {
return JCusolverDn.cusolverDnSgeqrf_bufferSize(handle, m, n, A, lda, Lwork);
}
@Override
public int cusolverDngeqrf(cusolverDnHandle handle, int m, int n, Pointer A, int lda, Pointer TAU,
Pointer Workspace, int Lwork, Pointer devInfo) {
return JCusolverDn.cusolverDnSgeqrf(handle, m, n, A, lda, TAU, Workspace, Lwork, devInfo);
}
@Override
public int cusolverDnormqr(cusolverDnHandle handle, int side, int trans, int m, int n, int k, Pointer A, int lda,
Pointer tau, Pointer C, int ldc, Pointer work, int lwork, Pointer devInfo) {
return JCusolverDn.cusolverDnSormqr(handle, side, trans, m, n, k, A, lda, tau, C, ldc, work, lwork, devInfo);
}
@Override
public int cusparsecsrgeam(cusparseHandle handle, int m, int n, Pointer alpha, cusparseMatDescr descrA, int nnzA,
Pointer csrValA, Pointer csrRowPtrA, Pointer csrColIndA, Pointer beta, cusparseMatDescr descrB, int nnzB,
Pointer csrValB, Pointer csrRowPtrB, Pointer csrColIndB, cusparseMatDescr descrC, Pointer csrValC,
Pointer csrRowPtrC, Pointer csrColIndC) {
long[] pBufferSizeInBytes = {0};
int status = JCusparse.cusparseDcsrgeam2_bufferSizeExt(handle, m, n, alpha, descrA, nnzA, csrValA, csrRowPtrA,
csrColIndA, beta, descrB, nnzB, csrValB, csrRowPtrB, csrColIndB, descrC, csrValC, csrRowPtrC, csrColIndC,
pBufferSizeInBytes);
if(status != CUSPARSE_STATUS_SUCCESS)
return status;
Pointer buffer = new Pointer();
if(pBufferSizeInBytes[0] > 0)
cudaMalloc(buffer, pBufferSizeInBytes[0]);
try {
// C = α*A + β*B
return JCusparse.cusparseDcsrgeam2(handle, m, n, alpha, descrA, nnzA, csrValA, csrRowPtrA, csrColIndA, beta,
descrB, nnzB, csrValB, csrRowPtrB, csrColIndB, descrC, csrValC, csrRowPtrC, csrColIndC, buffer);
}
finally {
if(pBufferSizeInBytes[0] > 0)
cudaFree(buffer);
}
}
@Override
public int cusparsecsr2dense(cusparseHandle handle, int m, int n, cusparseMatDescr descrA, Pointer csrValA,
Pointer csrRowPtrA, Pointer csrColIndA, Pointer A, int lda, long nnz) {
// Get index base from legacy descriptor -> 0 or 1
int idxBase = JCusparse.cusparseGetMatIndexBase(descrA);
// Create generric sparse-matrix descriptor required by CUDA 12
cusparseSpMatDescr spMatA = new cusparseSpMatDescr();
// Build CSR descriptor
cusparseCreateCsr(spMatA, m, n, nnz, csrRowPtrA, csrColIndA, csrValA, CUSPARSE_INDEX_32I,
CUSPARSE_INDEX_32I, idxBase, CUDA_R_32F);
// Build dense descriptor
cusparseDnMatDescr matB = new cusparseDnMatDescr();
cusparseCreateDnMat(matB, m, n, lda, A, CUDA_R_32F, CUSPARSE_ORDER_COL);
// Determine buffer size
int alg = CUSPARSE_SPARSETODENSE_ALG_DEFAULT;
long[] bufSize = {0};
cusparseSparseToDense_bufferSize(handle, spMatA.asConst(), matB, alg,
bufSize); //bufSize[0] now holds the exact byte count
// Allocate scratch space of the requested size
Pointer dBuffer = new Pointer();
if(bufSize[0] > 0) {
cudaMalloc(dBuffer, bufSize[0]);
}
try {
// Write dense matrix
int algSparseToDense = CUSPARSE_SPARSETODENSE_ALG_DEFAULT;
return cusparseSparseToDense(handle, spMatA.asConst(), matB, algSparseToDense, dBuffer);
}
finally {
if(bufSize[0] > 0)
cudaFree(dBuffer);
cusparseDestroyDnMat(matB.asConst());
cusparseDestroySpMat(spMatA.asConst());
}
}
@Override
public int cusparsedense2csr(cusparseHandle handle, int m, int n, cusparseMatDescr descrA, Pointer A, int lda,
Pointer nnzPerRow, Pointer csrValA, Pointer csrRowPtrA, Pointer csrColIndA) {
// setup
int dataType = CUDA_R_32F;
cusparseSpMatDescr csrDesc = new cusparseSpMatDescr();
cusparseDnMatDescr denseDesc = new cusparseDnMatDescr();
int idxBase = cusparseGetMatIndexBase(descrA);
int alg = CUSPARSE_DENSETOSPARSE_ALG_DEFAULT;
long[] bufferSize = {0};
Pointer dBuffer = new Pointer();
// Create dense matrix A
cusparseCreateDnMat(denseDesc, m, n, lda, A, dataType, CUSPARSE_ORDER_COL);
// Create sparse matrix B in CSR format
cusparseCreateCsr(csrDesc, m, n, 0, csrRowPtrA, csrColIndA, csrValA, CUSPARSE_INDEX_32I, CUSPARSE_INDEX_32I,
idxBase, dataType);
// allocate an external buffer if needed
cusparseDenseToSparse_bufferSize(handle, denseDesc.asConst(), csrDesc, alg, bufferSize);
if(bufferSize[0] > 0)
cudaMalloc(dBuffer, bufferSize[0]);
// prepare Sparse to Dense conversion
cusparseDenseToSparse_analysis(handle, denseDesc.asConst(), csrDesc, alg, dBuffer);
/** Keep this in case needed later for debugging*/
/*long[] rowsTmp = {0}, colsTmp = {0}, nnz = {0};
JCusparse.cusparseSpMatGetSize(csrDesc.asConst(), rowsTmp, colsTmp, nnz);
// only allocate if caller passed null pointers
if (csrColIndA == null)
cudaMalloc(csrColIndA, nnz[0] * Sizeof.INT);
if (csrValA == null)
cudaMalloc(csrValA, nnz[0] * Sizeof.DOUBLE);*/
// re-attach column & value pointers
JCusparse.cusparseCsrSetPointers(csrDesc, csrRowPtrA, csrColIndA, csrValA);
try {
// execute Sparse to Dense conversion
return cusparseDenseToSparse_convert(handle, denseDesc.asConst(), csrDesc, alg, dBuffer);
}
finally {
cusparseDestroyDnMat(denseDesc.asConst());
cusparseDestroySpMat(csrDesc.asConst());
if(bufferSize[0] > 0)
cudaFree(dBuffer);
}
}
@Override
public int cusparsennz(cusparseHandle handle, int dirA, int m, int n, cusparseMatDescr descrA, Pointer A, int lda,
Pointer nnzPerRowCol, Pointer nnzTotalDevHostPtr) {
return JCusparse.cusparseSnnz(handle, dirA, m, n, descrA, A, lda, nnzPerRowCol, nnzTotalDevHostPtr);
}
@Override
public void deviceToHost(GPUContext gCtx, Pointer src, double[] dest, String instName, boolean isEviction) {
long t0 = DMLScript.STATISTICS ? System.nanoTime() : 0;
// We invoke transfer matrix from device to host in two cases:
// 1. During eviction of unlocked matrices
// 2. During acquireHostRead
//
// If the single-precision support is enabled, then float-to-double conversion is required as CP expects the data to be in double format.
// This conversion can be done on host or on device. We typically prefer to do this conversion on device due to GPU's high-memory bandwidth.
// However, the conversion requires an additional space to be allocated for the conversion, which can lead to infinite recursion
// during eviction: `evict -> devictToHost -> float2double -> allocate -> ensureFreeSpace -> evict`.
// To avoid this recursion, it is necessary to perform this conversion in host.
if(PERFORM_CONVERSION_ON_DEVICE && !isEviction) {
Pointer deviceDoubleData = gCtx.allocate(instName, ((long)dest.length)*Sizeof.DOUBLE, false);
LibMatrixCUDA.float2double(gCtx, src, deviceDoubleData, dest.length);
cudaMemcpy(Pointer.to(dest), deviceDoubleData, ((long)dest.length)*Sizeof.DOUBLE, cudaMemcpyDeviceToHost);
gCtx.cudaFreeHelper(instName, deviceDoubleData, DMLScript.EAGER_CUDA_FREE);
}
else {
LOG.debug("Potential OOM: Allocated additional space on host in deviceToHost");
FloatBuffer floatData = ByteBuffer.allocateDirect(Sizeof.FLOAT*dest.length).order(ByteOrder.nativeOrder()).asFloatBuffer();
cudaMemcpy(Pointer.to(floatData), src, ((long)dest.length)*Sizeof.FLOAT, cudaMemcpyDeviceToHost);
LibMatrixNative.fromFloatBuffer(floatData, dest);
}
if(DMLScript.STATISTICS) {
long totalTime = System.nanoTime() - t0;
GPUStatistics.cudaFloat2DoubleTime.add(totalTime);
GPUStatistics.cudaFloat2DoubleCount.add(1);
}
}
@Override
public void hostToDevice(GPUContext gCtx, double[] src, Pointer dest, String instName) {
LOG.debug("Potential OOM: Allocated additional space in hostToDevice");
// TODO: Perform conversion on GPU using double2float and float2double kernels
long t0 = DMLScript.STATISTICS ? System.nanoTime() : 0;
if(PERFORM_CONVERSION_ON_DEVICE) {
Pointer deviceDoubleData = gCtx.allocate(instName, ((long)src.length)*Sizeof.DOUBLE, false);
cudaMemcpy(deviceDoubleData, Pointer.to(src), ((long)src.length)*Sizeof.DOUBLE, cudaMemcpyHostToDevice);
LibMatrixCUDA.double2float(gCtx, deviceDoubleData, dest, src.length);
gCtx.cudaFreeHelper(instName, deviceDoubleData, DMLScript.EAGER_CUDA_FREE);
}
else {
FloatBuffer floatData = ByteBuffer.allocateDirect(Sizeof.FLOAT*src.length).order(ByteOrder.nativeOrder()).asFloatBuffer();
IntStream.range(0, src.length).parallel().forEach(i -> floatData.put(i, (float)src[i]));
cudaMemcpy(dest, Pointer.to(floatData), ((long)src.length)*Sizeof.FLOAT, cudaMemcpyHostToDevice);
}
if(DMLScript.STATISTICS) {
long totalTime = System.nanoTime() - t0;
GPUStatistics.cudaDouble2FloatTime.add(totalTime);
GPUStatistics.cudaDouble2FloatCount.add(1);
}
}
}