|
@@ -20,16 +20,21 @@ import botan "../bindings"
|
|
|
High level API
|
|
|
*/
|
|
|
|
|
|
+DIGEST_SIZE_224 :: 28
|
|
|
+DIGEST_SIZE_256 :: 32
|
|
|
+DIGEST_SIZE_384 :: 48
|
|
|
+DIGEST_SIZE_512 :: 64
|
|
|
+
|
|
|
// hash_string_224 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_string_224 :: proc(data: string) -> [28]byte {
|
|
|
+hash_string_224 :: proc(data: string) -> [DIGEST_SIZE_224]byte {
|
|
|
return hash_bytes_224(transmute([]byte)(data))
|
|
|
}
|
|
|
|
|
|
// hash_bytes_224 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_bytes_224 :: proc(data: []byte) -> [28]byte {
|
|
|
- hash: [28]byte
|
|
|
+hash_bytes_224 :: proc(data: []byte) -> [DIGEST_SIZE_224]byte {
|
|
|
+ hash: [DIGEST_SIZE_224]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_224, 0)
|
|
|
botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
@@ -38,10 +43,29 @@ hash_bytes_224 :: proc(data: []byte) -> [28]byte {
|
|
|
return hash
|
|
|
}
|
|
|
|
|
|
+// hash_string_to_buffer_224 will hash the given input and assign the
|
|
|
+// computed hash to the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_string_to_buffer_224 :: proc(data: string, hash: []byte) {
|
|
|
+ hash_bytes_to_buffer_224(transmute([]byte)(data), hash);
|
|
|
+}
|
|
|
+
|
|
|
+// hash_bytes_to_buffer_224 will hash the given input and write the
|
|
|
+// computed hash into the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_bytes_to_buffer_224 :: proc(data, hash: []byte) {
|
|
|
+ assert(len(hash) >= DIGEST_SIZE_224, "Size of destination buffer is smaller than the digest size")
|
|
|
+ ctx: botan.hash_t
|
|
|
+ botan.hash_init(&ctx, botan.HASH_SHA3_224, 0)
|
|
|
+ botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
|
+ botan.hash_final(ctx, &hash[0])
|
|
|
+ botan.hash_destroy(ctx)
|
|
|
+}
|
|
|
+
|
|
|
// hash_stream_224 will read the stream in chunks and compute a
|
|
|
// hash from its contents
|
|
|
-hash_stream_224 :: proc(s: io.Stream) -> ([28]byte, bool) {
|
|
|
- hash: [28]byte
|
|
|
+hash_stream_224 :: proc(s: io.Stream) -> ([DIGEST_SIZE_224]byte, bool) {
|
|
|
+ hash: [DIGEST_SIZE_224]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_224, 0)
|
|
|
buf := make([]byte, 512)
|
|
@@ -60,7 +84,7 @@ hash_stream_224 :: proc(s: io.Stream) -> ([28]byte, bool) {
|
|
|
|
|
|
// hash_file_224 will read the file provided by the given handle
|
|
|
// and compute a hash
|
|
|
-hash_file_224 :: proc(hd: os.Handle, load_at_once := false) -> ([28]byte, bool) {
|
|
|
+hash_file_224 :: proc(hd: os.Handle, load_at_once := false) -> ([DIGEST_SIZE_224]byte, bool) {
|
|
|
if !load_at_once {
|
|
|
return hash_stream_224(os.stream_from_handle(hd))
|
|
|
} else {
|
|
@@ -68,7 +92,7 @@ hash_file_224 :: proc(hd: os.Handle, load_at_once := false) -> ([28]byte, bool)
|
|
|
return hash_bytes_224(buf[:]), ok
|
|
|
}
|
|
|
}
|
|
|
- return [28]byte{}, false
|
|
|
+ return [DIGEST_SIZE_224]byte{}, false
|
|
|
}
|
|
|
|
|
|
hash_224 :: proc {
|
|
@@ -76,18 +100,20 @@ hash_224 :: proc {
|
|
|
hash_file_224,
|
|
|
hash_bytes_224,
|
|
|
hash_string_224,
|
|
|
+ hash_bytes_to_buffer_224,
|
|
|
+ hash_string_to_buffer_224,
|
|
|
}
|
|
|
|
|
|
// hash_string_256 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_string_256 :: proc(data: string) -> [32]byte {
|
|
|
+hash_string_256 :: proc(data: string) -> [DIGEST_SIZE_256]byte {
|
|
|
return hash_bytes_256(transmute([]byte)(data))
|
|
|
}
|
|
|
|
|
|
// hash_bytes_256 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_bytes_256 :: proc(data: []byte) -> [32]byte {
|
|
|
- hash: [32]byte
|
|
|
+hash_bytes_256 :: proc(data: []byte) -> [DIGEST_SIZE_256]byte {
|
|
|
+ hash: [DIGEST_SIZE_256]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_256, 0)
|
|
|
botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
@@ -96,10 +122,29 @@ hash_bytes_256 :: proc(data: []byte) -> [32]byte {
|
|
|
return hash
|
|
|
}
|
|
|
|
|
|
+// hash_string_to_buffer_256 will hash the given input and assign the
|
|
|
+// computed hash to the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_string_to_buffer_256 :: proc(data: string, hash: []byte) {
|
|
|
+ hash_bytes_to_buffer_256(transmute([]byte)(data), hash);
|
|
|
+}
|
|
|
+
|
|
|
+// hash_bytes_to_buffer_256 will hash the given input and write the
|
|
|
+// computed hash into the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_bytes_to_buffer_256 :: proc(data, hash: []byte) {
|
|
|
+ assert(len(hash) >= DIGEST_SIZE_256, "Size of destination buffer is smaller than the digest size")
|
|
|
+ ctx: botan.hash_t
|
|
|
+ botan.hash_init(&ctx, botan.HASH_SHA3_256, 0)
|
|
|
+ botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
|
+ botan.hash_final(ctx, &hash[0])
|
|
|
+ botan.hash_destroy(ctx)
|
|
|
+}
|
|
|
+
|
|
|
// hash_stream_256 will read the stream in chunks and compute a
|
|
|
// hash from its contents
|
|
|
-hash_stream_256 :: proc(s: io.Stream) -> ([32]byte, bool) {
|
|
|
- hash: [32]byte
|
|
|
+hash_stream_256 :: proc(s: io.Stream) -> ([DIGEST_SIZE_256]byte, bool) {
|
|
|
+ hash: [DIGEST_SIZE_256]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_256, 0)
|
|
|
buf := make([]byte, 512)
|
|
@@ -118,7 +163,7 @@ hash_stream_256 :: proc(s: io.Stream) -> ([32]byte, bool) {
|
|
|
|
|
|
// hash_file_256 will read the file provided by the given handle
|
|
|
// and compute a hash
|
|
|
-hash_file_256 :: proc(hd: os.Handle, load_at_once := false) -> ([32]byte, bool) {
|
|
|
+hash_file_256 :: proc(hd: os.Handle, load_at_once := false) -> ([DIGEST_SIZE_256]byte, bool) {
|
|
|
if !load_at_once {
|
|
|
return hash_stream_256(os.stream_from_handle(hd))
|
|
|
} else {
|
|
@@ -126,7 +171,7 @@ hash_file_256 :: proc(hd: os.Handle, load_at_once := false) -> ([32]byte, bool)
|
|
|
return hash_bytes_256(buf[:]), ok
|
|
|
}
|
|
|
}
|
|
|
- return [32]byte{}, false
|
|
|
+ return [DIGEST_SIZE_256]byte{}, false
|
|
|
}
|
|
|
|
|
|
hash_256 :: proc {
|
|
@@ -134,18 +179,20 @@ hash_256 :: proc {
|
|
|
hash_file_256,
|
|
|
hash_bytes_256,
|
|
|
hash_string_256,
|
|
|
+ hash_bytes_to_buffer_256,
|
|
|
+ hash_string_to_buffer_256,
|
|
|
}
|
|
|
|
|
|
// hash_string_384 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_string_384 :: proc(data: string) -> [48]byte {
|
|
|
+hash_string_384 :: proc(data: string) -> [DIGEST_SIZE_384]byte {
|
|
|
return hash_bytes_384(transmute([]byte)(data))
|
|
|
}
|
|
|
|
|
|
// hash_bytes_384 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_bytes_384 :: proc(data: []byte) -> [48]byte {
|
|
|
- hash: [48]byte
|
|
|
+hash_bytes_384 :: proc(data: []byte) -> [DIGEST_SIZE_384]byte {
|
|
|
+ hash: [DIGEST_SIZE_384]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_384, 0)
|
|
|
botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
@@ -154,10 +201,29 @@ hash_bytes_384 :: proc(data: []byte) -> [48]byte {
|
|
|
return hash
|
|
|
}
|
|
|
|
|
|
+// hash_string_to_buffer_384 will hash the given input and assign the
|
|
|
+// computed hash to the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_string_to_buffer_384 :: proc(data: string, hash: []byte) {
|
|
|
+ hash_bytes_to_buffer_384(transmute([]byte)(data), hash);
|
|
|
+}
|
|
|
+
|
|
|
+// hash_bytes_to_buffer_384 will hash the given input and write the
|
|
|
+// computed hash into the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_bytes_to_buffer_384 :: proc(data, hash: []byte) {
|
|
|
+ assert(len(hash) >= DIGEST_SIZE_384, "Size of destination buffer is smaller than the digest size")
|
|
|
+ ctx: botan.hash_t
|
|
|
+ botan.hash_init(&ctx, botan.HASH_SHA3_384, 0)
|
|
|
+ botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
|
+ botan.hash_final(ctx, &hash[0])
|
|
|
+ botan.hash_destroy(ctx)
|
|
|
+}
|
|
|
+
|
|
|
// hash_stream_384 will read the stream in chunks and compute a
|
|
|
// hash from its contents
|
|
|
-hash_stream_384 :: proc(s: io.Stream) -> ([48]byte, bool) {
|
|
|
- hash: [48]byte
|
|
|
+hash_stream_384 :: proc(s: io.Stream) -> ([DIGEST_SIZE_384]byte, bool) {
|
|
|
+ hash: [DIGEST_SIZE_384]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_384, 0)
|
|
|
buf := make([]byte, 512)
|
|
@@ -176,7 +242,7 @@ hash_stream_384 :: proc(s: io.Stream) -> ([48]byte, bool) {
|
|
|
|
|
|
// hash_file_384 will read the file provided by the given handle
|
|
|
// and compute a hash
|
|
|
-hash_file_384 :: proc(hd: os.Handle, load_at_once := false) -> ([48]byte, bool) {
|
|
|
+hash_file_384 :: proc(hd: os.Handle, load_at_once := false) -> ([DIGEST_SIZE_384]byte, bool) {
|
|
|
if !load_at_once {
|
|
|
return hash_stream_384(os.stream_from_handle(hd))
|
|
|
} else {
|
|
@@ -184,7 +250,7 @@ hash_file_384 :: proc(hd: os.Handle, load_at_once := false) -> ([48]byte, bool)
|
|
|
return hash_bytes_384(buf[:]), ok
|
|
|
}
|
|
|
}
|
|
|
- return [48]byte{}, false
|
|
|
+ return [DIGEST_SIZE_384]byte{}, false
|
|
|
}
|
|
|
|
|
|
hash_384 :: proc {
|
|
@@ -192,18 +258,20 @@ hash_384 :: proc {
|
|
|
hash_file_384,
|
|
|
hash_bytes_384,
|
|
|
hash_string_384,
|
|
|
+ hash_bytes_to_buffer_384,
|
|
|
+ hash_string_to_buffer_384,
|
|
|
}
|
|
|
|
|
|
// hash_string_512 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_string_512 :: proc(data: string) -> [64]byte {
|
|
|
+hash_string_512 :: proc(data: string) -> [DIGEST_SIZE_512]byte {
|
|
|
return hash_bytes_512(transmute([]byte)(data))
|
|
|
}
|
|
|
|
|
|
// hash_bytes_512 will hash the given input and return the
|
|
|
// computed hash
|
|
|
-hash_bytes_512 :: proc(data: []byte) -> [64]byte {
|
|
|
- hash: [64]byte
|
|
|
+hash_bytes_512 :: proc(data: []byte) -> [DIGEST_SIZE_512]byte {
|
|
|
+ hash: [DIGEST_SIZE_512]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_512, 0)
|
|
|
botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
@@ -212,10 +280,29 @@ hash_bytes_512 :: proc(data: []byte) -> [64]byte {
|
|
|
return hash
|
|
|
}
|
|
|
|
|
|
+// hash_string_to_buffer_512 will hash the given input and assign the
|
|
|
+// computed hash to the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_string_to_buffer_512 :: proc(data: string, hash: []byte) {
|
|
|
+ hash_bytes_to_buffer_512(transmute([]byte)(data), hash);
|
|
|
+}
|
|
|
+
|
|
|
+// hash_bytes_to_buffer_512 will hash the given input and write the
|
|
|
+// computed hash into the second parameter.
|
|
|
+// It requires that the destination buffer is at least as big as the digest size
|
|
|
+hash_bytes_to_buffer_512 :: proc(data, hash: []byte) {
|
|
|
+ assert(len(hash) >= DIGEST_SIZE_512, "Size of destination buffer is smaller than the digest size")
|
|
|
+ ctx: botan.hash_t
|
|
|
+ botan.hash_init(&ctx, botan.HASH_SHA3_512, 0)
|
|
|
+ botan.hash_update(ctx, len(data) == 0 ? nil : &data[0], uint(len(data)))
|
|
|
+ botan.hash_final(ctx, &hash[0])
|
|
|
+ botan.hash_destroy(ctx)
|
|
|
+}
|
|
|
+
|
|
|
// hash_stream_512 will read the stream in chunks and compute a
|
|
|
// hash from its contents
|
|
|
-hash_stream_512 :: proc(s: io.Stream) -> ([64]byte, bool) {
|
|
|
- hash: [64]byte
|
|
|
+hash_stream_512 :: proc(s: io.Stream) -> ([DIGEST_SIZE_512]byte, bool) {
|
|
|
+ hash: [DIGEST_SIZE_512]byte
|
|
|
ctx: botan.hash_t
|
|
|
botan.hash_init(&ctx, botan.HASH_SHA3_512, 0)
|
|
|
buf := make([]byte, 512)
|
|
@@ -234,7 +321,7 @@ hash_stream_512 :: proc(s: io.Stream) -> ([64]byte, bool) {
|
|
|
|
|
|
// hash_file_512 will read the file provided by the given handle
|
|
|
// and compute a hash
|
|
|
-hash_file_512 :: proc(hd: os.Handle, load_at_once := false) -> ([64]byte, bool) {
|
|
|
+hash_file_512 :: proc(hd: os.Handle, load_at_once := false) -> ([DIGEST_SIZE_512]byte, bool) {
|
|
|
if !load_at_once {
|
|
|
return hash_stream_512(os.stream_from_handle(hd))
|
|
|
} else {
|
|
@@ -242,7 +329,7 @@ hash_file_512 :: proc(hd: os.Handle, load_at_once := false) -> ([64]byte, bool)
|
|
|
return hash_bytes_512(buf[:]), ok
|
|
|
}
|
|
|
}
|
|
|
- return [64]byte{}, false
|
|
|
+ return [DIGEST_SIZE_512]byte{}, false
|
|
|
}
|
|
|
|
|
|
hash_512 :: proc {
|
|
@@ -250,6 +337,8 @@ hash_512 :: proc {
|
|
|
hash_file_512,
|
|
|
hash_bytes_512,
|
|
|
hash_string_512,
|
|
|
+ hash_bytes_to_buffer_512,
|
|
|
+ hash_string_to_buffer_512,
|
|
|
}
|
|
|
|
|
|
/*
|