Commit message (Collapse) | Author | Age | Files | Lines | |
---|---|---|---|---|---|
* | x86, mem: memmove_64.S: Optimize memmove by enhanced REP MOVSB/STOSB | Fenghua Yu | 2011-05-17 | 1 | -1/+28 |
| | | | | | | | | | | | | | Support memmove() by enhanced rep movsb. On processors supporting enhanced REP MOVSB/STOSB, the alternative memmove() function using enhanced rep movsb overrides the original function. The patch doesn't change the backward memmove case to use enhanced rep movsb. Signed-off-by: Fenghua Yu <fenghua.yu@intel.com> Link: http://lkml.kernel.org/r/1305671358-14478-9-git-send-email-fenghua.yu@intel.com Signed-off-by: H. Peter Anvin <hpa@linux.intel.com> | ||||
* | x86-64, mem: Convert memmove() to assembly file and fix return value bug | Fenghua Yu | 2011-01-25 | 1 | -0/+197 |
memmove_64.c only implements memmove() function which is completely written in inline assembly code. Therefore it doesn't make sense to keep the assembly code in .c file. Currently memmove() doesn't store return value to rax. This may cause issue if caller uses the return value. The patch fixes this issue. Signed-off-by: Fenghua Yu <fenghua.yu@intel.com> LKML-Reference: <1295314755-6625-1-git-send-email-fenghua.yu@intel.com> Signed-off-by: H. Peter Anvin <hpa@linux.intel.com> |