/*
 * arch/score/lib/csum_partial.S
 *
 * Score Processor version.
 *
 * Copyright (C) 2009 Sunplus Core Technology Co., Ltd.
 *  Lennox Wu <lennox.wu@sunplusct.com>
 *  Chen Liqin <liqin.chen@sunplusct.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation; either version 2 of the License, or
 * (at your option) any later version.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, see the file COPYING, or write
 * to the Free Software Foundation, Inc.,
 * 51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA
 */
#include <linux/linkage.h>

#define ADDC(sum,reg)			\
	add	sum, sum, reg;		\
	cmp.c	reg, sum;		\
	bleu	9f;			\
	addi	sum, 0x1;		\
9:

#define CSUM_BIGCHUNK(src, offset, sum)		\
	lw	r8, [src, offset + 0x00];	\
	lw	r9, [src, offset + 0x04];	\
	lw	r10, [src, offset + 0x08];	\
	lw	r11, [src, offset + 0x0c];	\
	ADDC(sum, r8);				\
	ADDC(sum, r9);				\
	ADDC(sum, r10);				\
	ADDC(sum, r11);				\
	lw	r8, [src, offset + 0x10];	\
	lw	r9, [src, offset + 0x14];	\
	lw	r10, [src, offset + 0x18]; 	\
	lw	r11, [src, offset + 0x1c]; 	\
	ADDC(sum, r8);				\
	ADDC(sum, r9);				\
	ADDC(sum, r10);				\
	ADDC(sum, r11);				\

#define src r4
#define dest r5
#define sum r27

	.text
/* unknown src alignment and < 8 bytes to go */
small_csumcpy:
	mv	r5, r10
	ldi	r9, 0x0
	cmpi.c	r25, 0x1
	beq pass_small_set_t7	/*already set, jump to pass_small_set_t7*/
	andri.c	r25,r4 , 0x1	/*Is src 2 bytes aligned?*/

pass_small_set_t7:
	beq	aligned
	cmpi.c	r5, 0x0
	beq	fold
	lbu	r9, [src]
	slli	r9,r9, 0x8	/*Little endian*/
	ADDC(sum, r9)
	addi	src, 0x1
	subi.c	r5, 0x1

	/*len still a full word */
aligned:
	andri.c r8, r5, 0x4	/*Len >= 4?*/
	beq	len_less_4bytes

	/* Still a full word (4byte) to go,and the src is word aligned.*/
	andri.c	r8, src, 0x3	/*src is 4bytes aligned, so use LW!!*/
	beq	four_byte_aligned
	lhu 	r9, [src]
	addi	src, 2
	ADDC(sum, r9)
	lhu 	r9, [src]
	addi	src, 2
	ADDC(sum, r9)
	b len_less_4bytes

four_byte_aligned:		/* Len >=4 and four byte aligned */
	lw	r9, [src]
	addi	src, 4
	ADDC(sum, r9)

len_less_4bytes:		/* 2 byte aligned aligned and length<4B */
	andri.c r8, r5, 0x2
	beq	len_less_2bytes
	lhu	r9, [src]
	addi	src, 0x2	/* src+=2 */
	ADDC(sum, r9)

len_less_2bytes:		/* len = 1 */
	andri.c r8, r5, 0x1
	beq 	fold		/* less than 2 and not equal 1--> len=0 -> fold */
	lbu	r9, [src]

fold_ADDC:
	ADDC(sum, r9)
fold:
	/* fold checksum */
	slli	r26, sum, 16
	add	sum, sum, r26
	cmp.c	r26, sum
	srli	sum, sum, 16
	bleu 	1f 		/* if r26<=sum */
	addi	sum, 0x1 	/* r26>sum */
1:
	/* odd buffer alignment? r25 was set in csum_partial */
	cmpi.c	r25, 0x0
	beq	1f
	slli	r26, sum, 8
	srli	sum, sum, 8
	or	sum, sum, r26
	andi	sum, 0xffff
1:
	.set	optimize
	/* Add the passed partial csum. */
	ADDC(sum, r6)
	mv	r4, sum
	br	r3
	.set	volatile

	.align	5
ENTRY(csum_partial)
	ldi sum, 0
	ldi r25, 0
	mv r10, r5
	cmpi.c	r5, 0x8
	blt	small_csumcpy		/* < 8(signed) bytes to copy */
	cmpi.c	r5, 0x0
	beq	out
	andri.c	r25, src, 0x1		/* odd buffer? */

	beq	word_align
hword_align:				/* 1 byte */
	lbu	r8, [src]
	subi	r5, 0x1
	slli	r8, r8, 8
	ADDC(sum, r8)
	addi	src, 0x1

word_align:				/* 2 bytes */
	andri.c r8, src, 0x2		/* 4bytes(dword)_aligned? */
	beq	dword_align		/* not, maybe dword_align */
	lhu	r8, [src]
	subi	r5, 0x2
	ADDC(sum, r8)
	addi	src, 0x2

dword_align:				/* 4bytes */
	mv 	r26, r5			/* maybe useless when len >=56 */
	ldi 	r8, 56
	cmp.c	r8, r5
	bgtu	do_end_words		/* if a1(len)<t0(56) ,unsigned */
	andri.c	r26, src, 0x4
	beq	qword_align
	lw	r8, [src]
	subi	r5, 0x4
	ADDC(sum, r8)
	addi	src, 0x4

qword_align:				/* 8 bytes */
	andri.c r26, src, 0x8
	beq	oword_align
	lw	r8, [src, 0x0]
	lw	r9, [src, 0x4]
	subi	r5, 0x8			/* len-=0x8 */
	ADDC(sum, r8)
	ADDC(sum, r9)
	addi	src, 0x8

oword_align:				/* 16bytes */
	andri.c	r26, src, 0x10
	beq	begin_movement
	lw	r10, [src, 0x08]
	lw	r11, [src, 0x0c]
	lw	r8, [src, 0x00]
	lw	r9, [src, 0x04]
	ADDC(sum, r10)
	ADDC(sum, r11)
	ADDC(sum, r8)
	ADDC(sum, r9)
	subi	r5, 0x10
	addi	src, 0x10

begin_movement:
	srli.c	r26, r5, 0x7		/* len>=128? */
	beq	1f			/* len<128 */

/* r26 is the result that computed in oword_align */
move_128bytes:
	CSUM_BIGCHUNK(src, 0x00, sum)
	CSUM_BIGCHUNK(src, 0x20, sum)
	CSUM_BIGCHUNK(src, 0x40, sum)
	CSUM_BIGCHUNK(src, 0x60, sum)
	subi.c	r26, 0x01		/* r26 equals len/128 */
	addi	src, 0x80
	bne	move_128bytes

1:	/* len<128,we process 64byte here */
	andri.c	r10, r5, 0x40
	beq	1f

move_64bytes:
	CSUM_BIGCHUNK(src, 0x00, sum)
	CSUM_BIGCHUNK(src, 0x20, sum)
	addi	src, 0x40

1:					/* len<64 */
	andri	r26, r5, 0x1c		/* 0x1c=28 */
	andri.c	r10, r5, 0x20
	beq	do_end_words		/* decided by andri */

move_32bytes:
	CSUM_BIGCHUNK(src, 0x00, sum)
	andri	r26, r5, 0x1c
	addri	src, src, 0x20

do_end_words:				/* len<32 */
	/* r26 was set already in dword_align */
	cmpi.c	r26, 0x0
	beq	maybe_end_cruft		/* len<28 or len<56 */
	srli	r26, r26, 0x2

end_words:
	lw	r8, [src]
	subi.c	r26, 0x1		/* unit is 4 byte */
	ADDC(sum, r8)
	addi	src, 0x4
	cmpi.c	r26, 0x0
	bne	end_words		/* r26!=0 */

maybe_end_cruft:			/* len<4 */
	andri	r10, r5, 0x3

small_memcpy:
	mv	r5, r10
	j	small_csumcpy

out:
	mv	r4, sum
	br	r3

END(csum_partial)