@@ -471,7 +471,9 @@ func (sfa *snowflakeFileTransferAgent) processFileCompressionType() error {
471
471
if err != nil {
472
472
return err
473
473
}
474
- io .ReadAll (r ) // flush out tee buffer
474
+ if _ , err = io .ReadAll (r ); err != nil { // flush out tee buffer
475
+ return err
476
+ }
475
477
} else {
476
478
mtype , err = mimetype .DetectFile (fileName )
477
479
}
@@ -880,7 +882,7 @@ func (sfa *snowflakeFileTransferAgent) uploadOneFile(meta *fileMetadata) (*fileM
880
882
881
883
if meta .srcStream != nil {
882
884
if meta .realSrcStream != nil {
883
- // the whole file has been read in compressFileWithGzipFromStream
885
+ // the file has been fully read in compressFileWithGzipFromStream
884
886
meta .sha256Digest , meta .uploadSize , err = fileUtil .getDigestAndSizeForStream (& meta .realSrcStream )
885
887
} else {
886
888
r := getReaderFromContext (sfa .ctx )
@@ -889,7 +891,9 @@ func (sfa *snowflakeFileTransferAgent) uploadOneFile(meta *fileMetadata) (*fileM
889
891
}
890
892
891
893
var fullSrcStream bytes.Buffer
892
- io .Copy (& fullSrcStream , meta .srcStream )
894
+ if _ , err = io .Copy (& fullSrcStream , meta .srcStream ); err != nil {
895
+ return nil , err
896
+ }
893
897
894
898
// continue reading the rest of the data in chunks
895
899
chunk := make ([]byte , fileChunkSize )
@@ -902,7 +906,9 @@ func (sfa *snowflakeFileTransferAgent) uploadOneFile(meta *fileMetadata) (*fileM
902
906
}
903
907
fullSrcStream .Write (chunk [:n ])
904
908
}
905
- io .Copy (meta .srcStream , & fullSrcStream )
909
+ if _ , err = io .Copy (meta .srcStream , & fullSrcStream ); err != nil {
910
+ return nil , err
911
+ }
906
912
meta .sha256Digest , meta .uploadSize , err = fileUtil .getDigestAndSizeForStream (& meta .srcStream )
907
913
}
908
914
} else {
0 commit comments