summaryrefslogtreecommitdiffstats
path: root/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/armVCM4P10_InterpolateLuma_HalfHor4x4_unsafe_s.S
blob: 1e2d16b7a4e1261f36c7b555875d6943a4fd2848 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
/*
 * Copyright (C) 2007-2008 ARM Limited
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *      http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 *
 */
/*
 *
 */

    .eabi_attribute 24, 1
    .eabi_attribute 25, 1

    .arm
    .fpu neon
    .text

    .global armVCM4P10_InterpolateLuma_HalfHor4x4_unsafe
    .func   armVCM4P10_InterpolateLuma_HalfHor4x4_unsafe
armVCM4P10_InterpolateLuma_HalfHor4x4_unsafe:
    PUSH     {r4-r12,lr}
    VLD1.8   {d22,d23},[r0],r1
    VEXT.8   d10,d22,d23,#5
    VEXT.8   d12,d22,d23,#1
    VEXT.8   d14,d22,d23,#2
    VEXT.8   d15,d22,d23,#3
    VEXT.8   d13,d22,d23,#4
    VADDL.U8 q11,d22,d10
    VADDL.U8 q4,d14,d15
    VADDL.U8 q6,d12,d13
    VLD1.8   {d24,d25},[r0],r1
    VMLA.I16 d22,d8,d31
    VMUL.I16 d8,d12,d30
    VEXT.8   d10,d24,d25,#5
    VEXT.8   d12,d24,d25,#1
    VEXT.8   d16,d24,d25,#2
    VEXT.8   d17,d24,d25,#3
    VEXT.8   d13,d24,d25,#4
    VADDL.U8 q12,d24,d10
    VSUB.I16 d22,d22,d8
    VADDL.U8 q4,d16,d17
    VADDL.U8 q6,d12,d13
    VLD1.8   {d26,d27},[r0],r1
    VMLA.I16 d24,d8,d31
    VMUL.I16 d8,d12,d30
    VEXT.8   d10,d26,d27,#5
    VEXT.8   d12,d26,d27,#1
    VEXT.8   d18,d26,d27,#2
    VEXT.8   d19,d26,d27,#3
    VEXT.8   d13,d26,d27,#4
    VADDL.U8 q13,d26,d10
    VSUB.I16 d24,d24,d8
    VADDL.U8 q4,d18,d19
    VADDL.U8 q6,d12,d13
    VLD1.8   {d28,d29},[r0],r1
    VMLA.I16 d26,d8,d31
    VMUL.I16 d8,d12,d30
    VEXT.8   d10,d28,d29,#5
    VEXT.8   d12,d28,d29,#1
    VEXT.8   d20,d28,d29,#2
    VEXT.8   d21,d28,d29,#3
    VEXT.8   d13,d28,d29,#4
    VADDL.U8 q14,d28,d10
    VSUB.I16 d26,d26,d8
    VADDL.U8 q4,d20,d21
    VADDL.U8 q6,d12,d13
    VMLA.I16 d28,d8,d31
    VMLS.I16 d28,d12,d30
    VQRSHRUN.S16 d22,q11,#5
    VQRSHRUN.S16 d24,q12,#5
    VQRSHRUN.S16 d26,q13,#5
    VQRSHRUN.S16 d28,q14,#5
    POP      {r4-r12,pc}
    .endfunc

    .end