summaryrefslogtreecommitdiff
path: root/lib/crypto/s390/aes.h
blob: 5466f6ecbce783d28e870661a250250e47b1bc31 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * AES optimized using the CP Assist for Cryptographic Functions (CPACF)
 *
 * Copyright 2026 Google LLC
 */
#include <asm/cpacf.h>
#include <linux/cpufeature.h>

static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_aes128);
static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_aes192);
static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_aes256);

/*
 * When the CPU supports CPACF AES for the requested key length, we need only
 * save a copy of the raw AES key, as that's what the CPACF instructions need.
 *
 * When unsupported, fall back to the generic key expansion and en/decryption.
 */
static void aes_preparekey_arch(union aes_enckey_arch *k,
				union aes_invkey_arch *inv_k,
				const u8 *in_key, int key_len, int nrounds)
{
	if (key_len == AES_KEYSIZE_128) {
		if (static_branch_likely(&have_cpacf_aes128)) {
			memcpy(k->raw_key, in_key, AES_KEYSIZE_128);
			return;
		}
	} else if (key_len == AES_KEYSIZE_192) {
		if (static_branch_likely(&have_cpacf_aes192)) {
			memcpy(k->raw_key, in_key, AES_KEYSIZE_192);
			return;
		}
	} else {
		if (static_branch_likely(&have_cpacf_aes256)) {
			memcpy(k->raw_key, in_key, AES_KEYSIZE_256);
			return;
		}
	}
	aes_expandkey_generic(k->rndkeys, inv_k ? inv_k->inv_rndkeys : NULL,
			      in_key, key_len);
}

static inline bool aes_crypt_s390(const struct aes_enckey *key,
				  u8 out[AES_BLOCK_SIZE],
				  const u8 in[AES_BLOCK_SIZE], int decrypt)
{
	if (key->len == AES_KEYSIZE_128) {
		if (static_branch_likely(&have_cpacf_aes128)) {
			cpacf_km(CPACF_KM_AES_128 | decrypt,
				 (void *)key->k.raw_key, out, in,
				 AES_BLOCK_SIZE);
			return true;
		}
	} else if (key->len == AES_KEYSIZE_192) {
		if (static_branch_likely(&have_cpacf_aes192)) {
			cpacf_km(CPACF_KM_AES_192 | decrypt,
				 (void *)key->k.raw_key, out, in,
				 AES_BLOCK_SIZE);
			return true;
		}
	} else {
		if (static_branch_likely(&have_cpacf_aes256)) {
			cpacf_km(CPACF_KM_AES_256 | decrypt,
				 (void *)key->k.raw_key, out, in,
				 AES_BLOCK_SIZE);
			return true;
		}
	}
	return false;
}

static void aes_encrypt_arch(const struct aes_enckey *key,
			     u8 out[AES_BLOCK_SIZE],
			     const u8 in[AES_BLOCK_SIZE])
{
	if (likely(aes_crypt_s390(key, out, in, 0)))
		return;
	aes_encrypt_generic(key->k.rndkeys, key->nrounds, out, in);
}

static void aes_decrypt_arch(const struct aes_key *key,
			     u8 out[AES_BLOCK_SIZE],
			     const u8 in[AES_BLOCK_SIZE])
{
	if (likely(aes_crypt_s390((const struct aes_enckey *)key, out, in,
				  CPACF_DECRYPT)))
		return;
	aes_decrypt_generic(key->inv_k.inv_rndkeys, key->nrounds, out, in);
}

#define aes_mod_init_arch aes_mod_init_arch
static void aes_mod_init_arch(void)
{
	if (cpu_have_feature(S390_CPU_FEATURE_MSA)) {
		cpacf_mask_t km_functions;

		cpacf_query(CPACF_KM, &km_functions);
		if (cpacf_test_func(&km_functions, CPACF_KM_AES_128))
			static_branch_enable(&have_cpacf_aes128);
		if (cpacf_test_func(&km_functions, CPACF_KM_AES_192))
			static_branch_enable(&have_cpacf_aes192);
		if (cpacf_test_func(&km_functions, CPACF_KM_AES_256))
			static_branch_enable(&have_cpacf_aes256);
	}
}