1*0090dad1Sjhigh /*
2*0090dad1Sjhigh * Argon2 reference source code package - reference C implementations
3*0090dad1Sjhigh *
4*0090dad1Sjhigh * Copyright 2015
5*0090dad1Sjhigh * Daniel Dinu, Dmitry Khovratovich, Jean-Philippe Aumasson, and Samuel Neves
6*0090dad1Sjhigh *
7*0090dad1Sjhigh * You may use this work under the terms of a Creative Commons CC0 1.0
8*0090dad1Sjhigh * License/Waiver or the Apache Public License 2.0, at your option. The terms of
9*0090dad1Sjhigh * these licenses can be found at:
10*0090dad1Sjhigh *
11*0090dad1Sjhigh * - CC0 1.0 Universal : http://creativecommons.org/publicdomain/zero/1.0
12*0090dad1Sjhigh * - Apache 2.0 : http://www.apache.org/licenses/LICENSE-2.0
13*0090dad1Sjhigh *
14*0090dad1Sjhigh * You should have received a copy of both of these licenses along with this
15*0090dad1Sjhigh * software. If not, they may be obtained at the above URLs.
16*0090dad1Sjhigh */
17*0090dad1Sjhigh
18*0090dad1Sjhigh #include <stdint.h>
19*0090dad1Sjhigh #include <string.h>
20*0090dad1Sjhigh #include <stdlib.h>
21*0090dad1Sjhigh
22*0090dad1Sjhigh #include "argon2.h"
23*0090dad1Sjhigh #include "core.h"
24*0090dad1Sjhigh
25*0090dad1Sjhigh #include "blake2/blamka-round-ref.h"
26*0090dad1Sjhigh #include "blake2/blake2-impl.h"
27*0090dad1Sjhigh #include "blake2/blake2.h"
28*0090dad1Sjhigh
29*0090dad1Sjhigh
30*0090dad1Sjhigh /*
31*0090dad1Sjhigh * Function fills a new memory block and optionally XORs the old block over the new one.
32*0090dad1Sjhigh * @next_block must be initialized.
33*0090dad1Sjhigh * @param prev_block Pointer to the previous block
34*0090dad1Sjhigh * @param ref_block Pointer to the reference block
35*0090dad1Sjhigh * @param next_block Pointer to the block to be constructed
36*0090dad1Sjhigh * @param with_xor Whether to XOR into the new block (1) or just overwrite (0)
37*0090dad1Sjhigh * @pre all block pointers must be valid
38*0090dad1Sjhigh */
fill_block(const block * prev_block,const block * ref_block,block * next_block,int with_xor)39*0090dad1Sjhigh static void fill_block(const block *prev_block, const block *ref_block,
40*0090dad1Sjhigh block *next_block, int with_xor) {
41*0090dad1Sjhigh block blockR, block_tmp;
42*0090dad1Sjhigh unsigned i;
43*0090dad1Sjhigh
44*0090dad1Sjhigh copy_block(&blockR, ref_block);
45*0090dad1Sjhigh xor_block(&blockR, prev_block);
46*0090dad1Sjhigh copy_block(&block_tmp, &blockR);
47*0090dad1Sjhigh /* Now blockR = ref_block + prev_block and block_tmp = ref_block + prev_block */
48*0090dad1Sjhigh if (with_xor) {
49*0090dad1Sjhigh /* Saving the next block contents for XOR over: */
50*0090dad1Sjhigh xor_block(&block_tmp, next_block);
51*0090dad1Sjhigh /* Now blockR = ref_block + prev_block and
52*0090dad1Sjhigh block_tmp = ref_block + prev_block + next_block */
53*0090dad1Sjhigh }
54*0090dad1Sjhigh
55*0090dad1Sjhigh /* Apply Blake2 on columns of 64-bit words: (0,1,...,15) , then
56*0090dad1Sjhigh (16,17,..31)... finally (112,113,...127) */
57*0090dad1Sjhigh for (i = 0; i < 8; ++i) {
58*0090dad1Sjhigh BLAKE2_ROUND_NOMSG(
59*0090dad1Sjhigh blockR.v[16 * i], blockR.v[16 * i + 1], blockR.v[16 * i + 2],
60*0090dad1Sjhigh blockR.v[16 * i + 3], blockR.v[16 * i + 4], blockR.v[16 * i + 5],
61*0090dad1Sjhigh blockR.v[16 * i + 6], blockR.v[16 * i + 7], blockR.v[16 * i + 8],
62*0090dad1Sjhigh blockR.v[16 * i + 9], blockR.v[16 * i + 10], blockR.v[16 * i + 11],
63*0090dad1Sjhigh blockR.v[16 * i + 12], blockR.v[16 * i + 13], blockR.v[16 * i + 14],
64*0090dad1Sjhigh blockR.v[16 * i + 15]);
65*0090dad1Sjhigh }
66*0090dad1Sjhigh
67*0090dad1Sjhigh /* Apply Blake2 on rows of 64-bit words: (0,1,16,17,...112,113), then
68*0090dad1Sjhigh (2,3,18,19,...,114,115).. finally (14,15,30,31,...,126,127) */
69*0090dad1Sjhigh for (i = 0; i < 8; i++) {
70*0090dad1Sjhigh BLAKE2_ROUND_NOMSG(
71*0090dad1Sjhigh blockR.v[2 * i], blockR.v[2 * i + 1], blockR.v[2 * i + 16],
72*0090dad1Sjhigh blockR.v[2 * i + 17], blockR.v[2 * i + 32], blockR.v[2 * i + 33],
73*0090dad1Sjhigh blockR.v[2 * i + 48], blockR.v[2 * i + 49], blockR.v[2 * i + 64],
74*0090dad1Sjhigh blockR.v[2 * i + 65], blockR.v[2 * i + 80], blockR.v[2 * i + 81],
75*0090dad1Sjhigh blockR.v[2 * i + 96], blockR.v[2 * i + 97], blockR.v[2 * i + 112],
76*0090dad1Sjhigh blockR.v[2 * i + 113]);
77*0090dad1Sjhigh }
78*0090dad1Sjhigh
79*0090dad1Sjhigh copy_block(next_block, &block_tmp);
80*0090dad1Sjhigh xor_block(next_block, &blockR);
81*0090dad1Sjhigh }
82*0090dad1Sjhigh
next_addresses(block * address_block,block * input_block,const block * zero_block)83*0090dad1Sjhigh static void next_addresses(block *address_block, block *input_block,
84*0090dad1Sjhigh const block *zero_block) {
85*0090dad1Sjhigh input_block->v[6]++;
86*0090dad1Sjhigh fill_block(zero_block, input_block, address_block, 0);
87*0090dad1Sjhigh fill_block(zero_block, address_block, address_block, 0);
88*0090dad1Sjhigh }
89*0090dad1Sjhigh
fill_segment(const argon2_instance_t * instance,argon2_position_t position)90*0090dad1Sjhigh void fill_segment(const argon2_instance_t *instance,
91*0090dad1Sjhigh argon2_position_t position) {
92*0090dad1Sjhigh block *ref_block = NULL, *curr_block = NULL;
93*0090dad1Sjhigh block address_block, input_block, zero_block;
94*0090dad1Sjhigh uint64_t pseudo_rand, ref_index, ref_lane;
95*0090dad1Sjhigh uint32_t prev_offset, curr_offset;
96*0090dad1Sjhigh uint32_t starting_index;
97*0090dad1Sjhigh uint32_t i;
98*0090dad1Sjhigh int data_independent_addressing;
99*0090dad1Sjhigh
100*0090dad1Sjhigh if (instance == NULL) {
101*0090dad1Sjhigh return;
102*0090dad1Sjhigh }
103*0090dad1Sjhigh
104*0090dad1Sjhigh data_independent_addressing =
105*0090dad1Sjhigh (instance->type == Argon2_i) ||
106*0090dad1Sjhigh (instance->type == Argon2_id && (position.pass == 0) &&
107*0090dad1Sjhigh (position.slice < ARGON2_SYNC_POINTS / 2));
108*0090dad1Sjhigh
109*0090dad1Sjhigh if (data_independent_addressing) {
110*0090dad1Sjhigh init_block_value(&zero_block, 0);
111*0090dad1Sjhigh init_block_value(&input_block, 0);
112*0090dad1Sjhigh
113*0090dad1Sjhigh input_block.v[0] = position.pass;
114*0090dad1Sjhigh input_block.v[1] = position.lane;
115*0090dad1Sjhigh input_block.v[2] = position.slice;
116*0090dad1Sjhigh input_block.v[3] = instance->memory_blocks;
117*0090dad1Sjhigh input_block.v[4] = instance->passes;
118*0090dad1Sjhigh input_block.v[5] = instance->type;
119*0090dad1Sjhigh }
120*0090dad1Sjhigh
121*0090dad1Sjhigh starting_index = 0;
122*0090dad1Sjhigh
123*0090dad1Sjhigh if ((0 == position.pass) && (0 == position.slice)) {
124*0090dad1Sjhigh starting_index = 2; /* we have already generated the first two blocks */
125*0090dad1Sjhigh
126*0090dad1Sjhigh /* Don't forget to generate the first block of addresses: */
127*0090dad1Sjhigh if (data_independent_addressing) {
128*0090dad1Sjhigh next_addresses(&address_block, &input_block, &zero_block);
129*0090dad1Sjhigh }
130*0090dad1Sjhigh }
131*0090dad1Sjhigh
132*0090dad1Sjhigh /* Offset of the current block */
133*0090dad1Sjhigh curr_offset = position.lane * instance->lane_length +
134*0090dad1Sjhigh position.slice * instance->segment_length + starting_index;
135*0090dad1Sjhigh
136*0090dad1Sjhigh if (0 == curr_offset % instance->lane_length) {
137*0090dad1Sjhigh /* Last block in this lane */
138*0090dad1Sjhigh prev_offset = curr_offset + instance->lane_length - 1;
139*0090dad1Sjhigh } else {
140*0090dad1Sjhigh /* Previous block */
141*0090dad1Sjhigh prev_offset = curr_offset - 1;
142*0090dad1Sjhigh }
143*0090dad1Sjhigh
144*0090dad1Sjhigh for (i = starting_index; i < instance->segment_length;
145*0090dad1Sjhigh ++i, ++curr_offset, ++prev_offset) {
146*0090dad1Sjhigh /*1.1 Rotating prev_offset if needed */
147*0090dad1Sjhigh if (curr_offset % instance->lane_length == 1) {
148*0090dad1Sjhigh prev_offset = curr_offset - 1;
149*0090dad1Sjhigh }
150*0090dad1Sjhigh
151*0090dad1Sjhigh /* 1.2 Computing the index of the reference block */
152*0090dad1Sjhigh /* 1.2.1 Taking pseudo-random value from the previous block */
153*0090dad1Sjhigh if (data_independent_addressing) {
154*0090dad1Sjhigh if (i % ARGON2_ADDRESSES_IN_BLOCK == 0) {
155*0090dad1Sjhigh next_addresses(&address_block, &input_block, &zero_block);
156*0090dad1Sjhigh }
157*0090dad1Sjhigh pseudo_rand = address_block.v[i % ARGON2_ADDRESSES_IN_BLOCK];
158*0090dad1Sjhigh } else {
159*0090dad1Sjhigh pseudo_rand = instance->memory[prev_offset].v[0];
160*0090dad1Sjhigh }
161*0090dad1Sjhigh
162*0090dad1Sjhigh /* 1.2.2 Computing the lane of the reference block */
163*0090dad1Sjhigh ref_lane = ((pseudo_rand >> 32)) % instance->lanes;
164*0090dad1Sjhigh
165*0090dad1Sjhigh if ((position.pass == 0) && (position.slice == 0)) {
166*0090dad1Sjhigh /* Can not reference other lanes yet */
167*0090dad1Sjhigh ref_lane = position.lane;
168*0090dad1Sjhigh }
169*0090dad1Sjhigh
170*0090dad1Sjhigh /* 1.2.3 Computing the number of possible reference block within the
171*0090dad1Sjhigh * lane.
172*0090dad1Sjhigh */
173*0090dad1Sjhigh position.index = i;
174*0090dad1Sjhigh ref_index = index_alpha(instance, &position, pseudo_rand & 0xFFFFFFFF,
175*0090dad1Sjhigh ref_lane == position.lane);
176*0090dad1Sjhigh
177*0090dad1Sjhigh /* 2 Creating a new block */
178*0090dad1Sjhigh ref_block =
179*0090dad1Sjhigh instance->memory + instance->lane_length * ref_lane + ref_index;
180*0090dad1Sjhigh curr_block = instance->memory + curr_offset;
181*0090dad1Sjhigh if (ARGON2_VERSION_10 == instance->version) {
182*0090dad1Sjhigh /* version 1.2.1 and earlier: overwrite, not XOR */
183*0090dad1Sjhigh fill_block(instance->memory + prev_offset, ref_block, curr_block, 0);
184*0090dad1Sjhigh } else {
185*0090dad1Sjhigh if(0 == position.pass) {
186*0090dad1Sjhigh fill_block(instance->memory + prev_offset, ref_block,
187*0090dad1Sjhigh curr_block, 0);
188*0090dad1Sjhigh } else {
189*0090dad1Sjhigh fill_block(instance->memory + prev_offset, ref_block,
190*0090dad1Sjhigh curr_block, 1);
191*0090dad1Sjhigh }
192*0090dad1Sjhigh }
193*0090dad1Sjhigh }
194*0090dad1Sjhigh }
195