mirror of
https://github.com/Gericom/teak-llvm.git
synced 2025-06-29 00:08:59 -04:00

In addition to enabling ELFv2 homogeneous aggregate handling, LLVM support to pass array types directly also enables a performance enhancement. We can now pass (non-homogeneous) aggregates that fit fully in registers as direct integer arrays, using an element type to encode the alignment requirement (that would otherwise go to the "byval align" field). This is preferable since "byval" forces the back-end to write the aggregate out to the stack, even if it could be passed fully in registers. This is particularly annoying on ELFv2, if there is no parameter save area available, since we then need to allocate space on the callee's stack just to hold those aggregates. Note that to implement this optimization, this patch does not attempt to fully anticipate register allocation rules as (defined in the ABI and) implemented in the back-end. Instead, the patch is simply passing *any* aggregate passed by value using the array mechanism if its size is up to 64 bytes. This means that some of those will end up being passed in stack slots anyway, but the generated code shouldn't be any worse either. (*Large* aggregates remain passed using "byval" to enable optimized copying via memcpy etc.) llvm-svn: 213495
155 lines
4.7 KiB
C
155 lines
4.7 KiB
C
// RUN: %clang_cc1 -faltivec -triple powerpc64-unknown-linux-gnu -emit-llvm -o - %s | FileCheck %s
|
|
|
|
#include <stdarg.h>
|
|
|
|
struct test1 { int x; int y; };
|
|
struct test2 { int x; int y; } __attribute__((aligned (16)));
|
|
struct test3 { int x; int y; } __attribute__((aligned (32)));
|
|
struct test4 { int x; int y; int z; };
|
|
struct test5 { int x[17]; };
|
|
struct test6 { int x[17]; } __attribute__((aligned (16)));
|
|
struct test7 { int x[17]; } __attribute__((aligned (32)));
|
|
|
|
// CHECK: define void @test1(i32 signext %x, i64 %y.coerce)
|
|
void test1 (int x, struct test1 y)
|
|
{
|
|
}
|
|
|
|
// CHECK: define void @test2(i32 signext %x, [1 x i128] %y.coerce)
|
|
void test2 (int x, struct test2 y)
|
|
{
|
|
}
|
|
|
|
// CHECK: define void @test3(i32 signext %x, [2 x i128] %y.coerce)
|
|
void test3 (int x, struct test3 y)
|
|
{
|
|
}
|
|
|
|
// CHECK: define void @test4(i32 signext %x, [2 x i64] %y.coerce)
|
|
void test4 (int x, struct test4 y)
|
|
{
|
|
}
|
|
|
|
// CHECK: define void @test5(i32 signext %x, %struct.test5* byval align 8 %y)
|
|
void test5 (int x, struct test5 y)
|
|
{
|
|
}
|
|
|
|
// CHECK: define void @test6(i32 signext %x, %struct.test6* byval align 16 %y)
|
|
void test6 (int x, struct test6 y)
|
|
{
|
|
}
|
|
|
|
// This case requires run-time realignment of the incoming struct
|
|
// CHECK: define void @test7(i32 signext %x, %struct.test7* byval align 16)
|
|
// CHECK: %y = alloca %struct.test7, align 32
|
|
// CHECK: call void @llvm.memcpy.p0i8.p0i8.i64
|
|
void test7 (int x, struct test7 y)
|
|
{
|
|
}
|
|
|
|
// CHECK: define void @test1va(%struct.test1* noalias sret %agg.result, i32 signext %x, ...)
|
|
// CHECK: %[[CUR:[^ ]+]] = load i8** %ap
|
|
// CHECK: %[[NEXT:[^ ]+]] = getelementptr i8* %[[CUR]], i64 8
|
|
// CHECK: store i8* %[[NEXT]], i8** %ap
|
|
// CHECK: bitcast i8* %[[CUR]] to %struct.test1*
|
|
struct test1 test1va (int x, ...)
|
|
{
|
|
struct test1 y;
|
|
va_list ap;
|
|
va_start(ap, x);
|
|
y = va_arg (ap, struct test1);
|
|
va_end(ap);
|
|
return y;
|
|
}
|
|
|
|
// CHECK: define void @test2va(%struct.test2* noalias sret %agg.result, i32 signext %x, ...)
|
|
// CHECK: %[[CUR:[^ ]+]] = load i8** %ap
|
|
// CHECK: %[[TMP0:[^ ]+]] = ptrtoint i8* %[[CUR]] to i64
|
|
// CHECK: %[[TMP1:[^ ]+]] = add i64 %[[TMP0]], 15
|
|
// CHECK: %[[TMP2:[^ ]+]] = and i64 %[[TMP1]], -16
|
|
// CHECK: %[[ALIGN:[^ ]+]] = inttoptr i64 %[[TMP2]] to i8*
|
|
// CHECK: %[[NEXT:[^ ]+]] = getelementptr i8* %[[ALIGN]], i64 16
|
|
// CHECK: store i8* %[[NEXT]], i8** %ap
|
|
// CHECK: bitcast i8* %[[ALIGN]] to %struct.test2*
|
|
struct test2 test2va (int x, ...)
|
|
{
|
|
struct test2 y;
|
|
va_list ap;
|
|
va_start(ap, x);
|
|
y = va_arg (ap, struct test2);
|
|
va_end(ap);
|
|
return y;
|
|
}
|
|
|
|
// CHECK: define void @test3va(%struct.test3* noalias sret %agg.result, i32 signext %x, ...)
|
|
// CHECK: %[[CUR:[^ ]+]] = load i8** %ap
|
|
// CHECK: %[[TMP0:[^ ]+]] = ptrtoint i8* %[[CUR]] to i64
|
|
// CHECK: %[[TMP1:[^ ]+]] = add i64 %[[TMP0]], 15
|
|
// CHECK: %[[TMP2:[^ ]+]] = and i64 %[[TMP1]], -16
|
|
// CHECK: %[[ALIGN:[^ ]+]] = inttoptr i64 %[[TMP2]] to i8*
|
|
// CHECK: %[[NEXT:[^ ]+]] = getelementptr i8* %[[ALIGN]], i64 32
|
|
// CHECK: store i8* %[[NEXT]], i8** %ap
|
|
// CHECK: bitcast i8* %[[ALIGN]] to %struct.test3*
|
|
struct test3 test3va (int x, ...)
|
|
{
|
|
struct test3 y;
|
|
va_list ap;
|
|
va_start(ap, x);
|
|
y = va_arg (ap, struct test3);
|
|
va_end(ap);
|
|
return y;
|
|
}
|
|
|
|
// CHECK: define void @test4va(%struct.test4* noalias sret %agg.result, i32 signext %x, ...)
|
|
// CHECK: %[[CUR:[^ ]+]] = load i8** %ap
|
|
// CHECK: %[[NEXT:[^ ]+]] = getelementptr i8* %[[CUR]], i64 16
|
|
// CHECK: store i8* %[[NEXT]], i8** %ap
|
|
// CHECK: bitcast i8* %[[CUR]] to %struct.test4*
|
|
struct test4 test4va (int x, ...)
|
|
{
|
|
struct test4 y;
|
|
va_list ap;
|
|
va_start(ap, x);
|
|
y = va_arg (ap, struct test4);
|
|
va_end(ap);
|
|
return y;
|
|
}
|
|
|
|
// CHECK: define void @testva_longdouble(%struct.test_longdouble* noalias sret %agg.result, i32 signext %x, ...)
|
|
// CHECK: %[[CUR:[^ ]+]] = load i8** %ap
|
|
// CHECK: %[[NEXT:[^ ]+]] = getelementptr i8* %[[CUR]], i64 16
|
|
// CHECK: store i8* %[[NEXT]], i8** %ap
|
|
// CHECK: bitcast i8* %[[CUR]] to %struct.test_longdouble*
|
|
struct test_longdouble { long double x; };
|
|
struct test_longdouble testva_longdouble (int x, ...)
|
|
{
|
|
struct test_longdouble y;
|
|
va_list ap;
|
|
va_start(ap, x);
|
|
y = va_arg (ap, struct test_longdouble);
|
|
va_end(ap);
|
|
return y;
|
|
}
|
|
|
|
// CHECK: define void @testva_vector(%struct.test_vector* noalias sret %agg.result, i32 signext %x, ...)
|
|
// CHECK: %[[CUR:[^ ]+]] = load i8** %ap
|
|
// CHECK: %[[TMP0:[^ ]+]] = ptrtoint i8* %[[CUR]] to i64
|
|
// CHECK: %[[TMP1:[^ ]+]] = add i64 %[[TMP0]], 15
|
|
// CHECK: %[[TMP2:[^ ]+]] = and i64 %[[TMP1]], -16
|
|
// CHECK: %[[ALIGN:[^ ]+]] = inttoptr i64 %[[TMP2]] to i8*
|
|
// CHECK: %[[NEXT:[^ ]+]] = getelementptr i8* %[[ALIGN]], i64 16
|
|
// CHECK: store i8* %[[NEXT]], i8** %ap
|
|
// CHECK: bitcast i8* %[[ALIGN]] to %struct.test_vector*
|
|
struct test_vector { vector int x; };
|
|
struct test_vector testva_vector (int x, ...)
|
|
{
|
|
struct test_vector y;
|
|
va_list ap;
|
|
va_start(ap, x);
|
|
y = va_arg (ap, struct test_vector);
|
|
va_end(ap);
|
|
return y;
|
|
}
|
|
|