121
|
1 //===- SafeStack.cpp - Safe Stack Insertion -------------------------------===//
|
120
|
2 //
|
|
3 // The LLVM Compiler Infrastructure
|
|
4 //
|
|
5 // This file is distributed under the University of Illinois Open Source
|
|
6 // License. See LICENSE.TXT for details.
|
|
7 //
|
|
8 //===----------------------------------------------------------------------===//
|
|
9 //
|
|
10 // This pass splits the stack into the safe stack (kept as-is for LLVM backend)
|
|
11 // and the unsafe stack (explicitly allocated and managed through the runtime
|
|
12 // support library).
|
|
13 //
|
|
14 // http://clang.llvm.org/docs/SafeStack.html
|
|
15 //
|
|
16 //===----------------------------------------------------------------------===//
|
|
17
|
|
18 #include "SafeStackColoring.h"
|
|
19 #include "SafeStackLayout.h"
|
121
|
20 #include "llvm/ADT/APInt.h"
|
|
21 #include "llvm/ADT/ArrayRef.h"
|
|
22 #include "llvm/ADT/SmallPtrSet.h"
|
|
23 #include "llvm/ADT/SmallVector.h"
|
120
|
24 #include "llvm/ADT/Statistic.h"
|
121
|
25 #include "llvm/Analysis/AssumptionCache.h"
|
120
|
26 #include "llvm/Analysis/BranchProbabilityInfo.h"
|
134
|
27 #include "llvm/Analysis/InlineCost.h"
|
121
|
28 #include "llvm/Analysis/LoopInfo.h"
|
120
|
29 #include "llvm/Analysis/ScalarEvolution.h"
|
|
30 #include "llvm/Analysis/ScalarEvolutionExpressions.h"
|
121
|
31 #include "llvm/Analysis/TargetLibraryInfo.h"
|
134
|
32 #include "llvm/CodeGen/TargetLowering.h"
|
121
|
33 #include "llvm/CodeGen/TargetPassConfig.h"
|
134
|
34 #include "llvm/CodeGen/TargetSubtargetInfo.h"
|
121
|
35 #include "llvm/IR/Argument.h"
|
|
36 #include "llvm/IR/Attributes.h"
|
|
37 #include "llvm/IR/CallSite.h"
|
|
38 #include "llvm/IR/ConstantRange.h"
|
120
|
39 #include "llvm/IR/Constants.h"
|
|
40 #include "llvm/IR/DIBuilder.h"
|
|
41 #include "llvm/IR/DataLayout.h"
|
|
42 #include "llvm/IR/DerivedTypes.h"
|
121
|
43 #include "llvm/IR/Dominators.h"
|
120
|
44 #include "llvm/IR/Function.h"
|
|
45 #include "llvm/IR/IRBuilder.h"
|
|
46 #include "llvm/IR/InstIterator.h"
|
121
|
47 #include "llvm/IR/Instruction.h"
|
120
|
48 #include "llvm/IR/Instructions.h"
|
|
49 #include "llvm/IR/IntrinsicInst.h"
|
|
50 #include "llvm/IR/Intrinsics.h"
|
|
51 #include "llvm/IR/MDBuilder.h"
|
|
52 #include "llvm/IR/Module.h"
|
121
|
53 #include "llvm/IR/Type.h"
|
|
54 #include "llvm/IR/Use.h"
|
|
55 #include "llvm/IR/User.h"
|
|
56 #include "llvm/IR/Value.h"
|
120
|
57 #include "llvm/Pass.h"
|
121
|
58 #include "llvm/Support/Casting.h"
|
120
|
59 #include "llvm/Support/Debug.h"
|
121
|
60 #include "llvm/Support/ErrorHandling.h"
|
120
|
61 #include "llvm/Support/MathExtras.h"
|
121
|
62 #include "llvm/Support/raw_ostream.h"
|
|
63 #include "llvm/Target/TargetMachine.h"
|
120
|
64 #include "llvm/Transforms/Utils/BasicBlockUtils.h"
|
134
|
65 #include "llvm/Transforms/Utils/Cloning.h"
|
120
|
66 #include "llvm/Transforms/Utils/Local.h"
|
121
|
67 #include <algorithm>
|
|
68 #include <cassert>
|
|
69 #include <cstdint>
|
|
70 #include <string>
|
|
71 #include <utility>
|
120
|
72
|
|
73 using namespace llvm;
|
|
74 using namespace llvm::safestack;
|
|
75
|
121
|
76 #define DEBUG_TYPE "safe-stack"
|
120
|
77
|
|
78 namespace llvm {
|
|
79
|
|
80 STATISTIC(NumFunctions, "Total number of functions");
|
|
81 STATISTIC(NumUnsafeStackFunctions, "Number of functions with unsafe stack");
|
|
82 STATISTIC(NumUnsafeStackRestorePointsFunctions,
|
|
83 "Number of functions that use setjmp or exceptions");
|
|
84
|
|
85 STATISTIC(NumAllocas, "Total number of allocas");
|
|
86 STATISTIC(NumUnsafeStaticAllocas, "Number of unsafe static allocas");
|
|
87 STATISTIC(NumUnsafeDynamicAllocas, "Number of unsafe dynamic allocas");
|
|
88 STATISTIC(NumUnsafeByValArguments, "Number of unsafe byval arguments");
|
|
89 STATISTIC(NumUnsafeStackRestorePoints, "Number of setjmps and landingpads");
|
|
90
|
|
91 } // namespace llvm
|
|
92
|
134
|
93 /// Use __safestack_pointer_address even if the platform has a faster way of
|
|
94 /// access safe stack pointer.
|
|
95 static cl::opt<bool>
|
|
96 SafeStackUsePointerAddress("safestack-use-pointer-address",
|
|
97 cl::init(false), cl::Hidden);
|
|
98
|
|
99
|
120
|
100 namespace {
|
|
101
|
|
102 /// Rewrite an SCEV expression for a memory access address to an expression that
|
|
103 /// represents offset from the given alloca.
|
|
104 ///
|
|
105 /// The implementation simply replaces all mentions of the alloca with zero.
|
|
106 class AllocaOffsetRewriter : public SCEVRewriteVisitor<AllocaOffsetRewriter> {
|
|
107 const Value *AllocaPtr;
|
|
108
|
|
109 public:
|
|
110 AllocaOffsetRewriter(ScalarEvolution &SE, const Value *AllocaPtr)
|
|
111 : SCEVRewriteVisitor(SE), AllocaPtr(AllocaPtr) {}
|
|
112
|
|
113 const SCEV *visitUnknown(const SCEVUnknown *Expr) {
|
|
114 if (Expr->getValue() == AllocaPtr)
|
|
115 return SE.getZero(Expr->getType());
|
|
116 return Expr;
|
|
117 }
|
|
118 };
|
|
119
|
|
120 /// The SafeStack pass splits the stack of each function into the safe
|
|
121 /// stack, which is only accessed through memory safe dereferences (as
|
|
122 /// determined statically), and the unsafe stack, which contains all
|
|
123 /// local variables that are accessed in ways that we can't prove to
|
|
124 /// be safe.
|
121
|
125 class SafeStack {
|
|
126 Function &F;
|
|
127 const TargetLoweringBase &TL;
|
|
128 const DataLayout &DL;
|
|
129 ScalarEvolution &SE;
|
120
|
130
|
|
131 Type *StackPtrTy;
|
|
132 Type *IntPtrTy;
|
|
133 Type *Int32Ty;
|
|
134 Type *Int8Ty;
|
|
135
|
|
136 Value *UnsafeStackPtr = nullptr;
|
|
137
|
|
138 /// Unsafe stack alignment. Each stack frame must ensure that the stack is
|
|
139 /// aligned to this value. We need to re-align the unsafe stack if the
|
|
140 /// alignment of any object on the stack exceeds this value.
|
|
141 ///
|
|
142 /// 16 seems like a reasonable upper bound on the alignment of objects that we
|
|
143 /// might expect to appear on the stack on most common targets.
|
|
144 enum { StackAlignment = 16 };
|
|
145
|
|
146 /// \brief Return the value of the stack canary.
|
|
147 Value *getStackGuard(IRBuilder<> &IRB, Function &F);
|
|
148
|
|
149 /// \brief Load stack guard from the frame and check if it has changed.
|
|
150 void checkStackGuard(IRBuilder<> &IRB, Function &F, ReturnInst &RI,
|
|
151 AllocaInst *StackGuardSlot, Value *StackGuard);
|
|
152
|
|
153 /// \brief Find all static allocas, dynamic allocas, return instructions and
|
|
154 /// stack restore points (exception unwind blocks and setjmp calls) in the
|
|
155 /// given function and append them to the respective vectors.
|
|
156 void findInsts(Function &F, SmallVectorImpl<AllocaInst *> &StaticAllocas,
|
|
157 SmallVectorImpl<AllocaInst *> &DynamicAllocas,
|
|
158 SmallVectorImpl<Argument *> &ByValArguments,
|
|
159 SmallVectorImpl<ReturnInst *> &Returns,
|
|
160 SmallVectorImpl<Instruction *> &StackRestorePoints);
|
|
161
|
|
162 /// \brief Calculate the allocation size of a given alloca. Returns 0 if the
|
|
163 /// size can not be statically determined.
|
|
164 uint64_t getStaticAllocaAllocationSize(const AllocaInst* AI);
|
|
165
|
|
166 /// \brief Allocate space for all static allocas in \p StaticAllocas,
|
|
167 /// replace allocas with pointers into the unsafe stack and generate code to
|
|
168 /// restore the stack pointer before all return instructions in \p Returns.
|
|
169 ///
|
|
170 /// \returns A pointer to the top of the unsafe stack after all unsafe static
|
|
171 /// allocas are allocated.
|
|
172 Value *moveStaticAllocasToUnsafeStack(IRBuilder<> &IRB, Function &F,
|
|
173 ArrayRef<AllocaInst *> StaticAllocas,
|
|
174 ArrayRef<Argument *> ByValArguments,
|
|
175 ArrayRef<ReturnInst *> Returns,
|
|
176 Instruction *BasePointer,
|
|
177 AllocaInst *StackGuardSlot);
|
|
178
|
|
179 /// \brief Generate code to restore the stack after all stack restore points
|
|
180 /// in \p StackRestorePoints.
|
|
181 ///
|
|
182 /// \returns A local variable in which to maintain the dynamic top of the
|
|
183 /// unsafe stack if needed.
|
|
184 AllocaInst *
|
|
185 createStackRestorePoints(IRBuilder<> &IRB, Function &F,
|
|
186 ArrayRef<Instruction *> StackRestorePoints,
|
|
187 Value *StaticTop, bool NeedDynamicTop);
|
|
188
|
|
189 /// \brief Replace all allocas in \p DynamicAllocas with code to allocate
|
|
190 /// space dynamically on the unsafe stack and store the dynamic unsafe stack
|
|
191 /// top to \p DynamicTop if non-null.
|
|
192 void moveDynamicAllocasToUnsafeStack(Function &F, Value *UnsafeStackPtr,
|
|
193 AllocaInst *DynamicTop,
|
|
194 ArrayRef<AllocaInst *> DynamicAllocas);
|
|
195
|
|
196 bool IsSafeStackAlloca(const Value *AllocaPtr, uint64_t AllocaSize);
|
|
197
|
|
198 bool IsMemIntrinsicSafe(const MemIntrinsic *MI, const Use &U,
|
|
199 const Value *AllocaPtr, uint64_t AllocaSize);
|
|
200 bool IsAccessSafe(Value *Addr, uint64_t Size, const Value *AllocaPtr,
|
|
201 uint64_t AllocaSize);
|
|
202
|
134
|
203 bool ShouldInlinePointerAddress(CallSite &CS);
|
|
204 void TryInlinePointerAddress();
|
|
205
|
120
|
206 public:
|
121
|
207 SafeStack(Function &F, const TargetLoweringBase &TL, const DataLayout &DL,
|
|
208 ScalarEvolution &SE)
|
|
209 : F(F), TL(TL), DL(DL), SE(SE),
|
|
210 StackPtrTy(Type::getInt8PtrTy(F.getContext())),
|
|
211 IntPtrTy(DL.getIntPtrType(F.getContext())),
|
|
212 Int32Ty(Type::getInt32Ty(F.getContext())),
|
|
213 Int8Ty(Type::getInt8Ty(F.getContext())) {}
|
120
|
214
|
121
|
215 // Run the transformation on the associated function.
|
|
216 // Returns whether the function was changed.
|
|
217 bool run();
|
|
218 };
|
120
|
219
|
|
220 uint64_t SafeStack::getStaticAllocaAllocationSize(const AllocaInst* AI) {
|
121
|
221 uint64_t Size = DL.getTypeAllocSize(AI->getAllocatedType());
|
120
|
222 if (AI->isArrayAllocation()) {
|
|
223 auto C = dyn_cast<ConstantInt>(AI->getArraySize());
|
|
224 if (!C)
|
|
225 return 0;
|
|
226 Size *= C->getZExtValue();
|
|
227 }
|
|
228 return Size;
|
|
229 }
|
|
230
|
|
231 bool SafeStack::IsAccessSafe(Value *Addr, uint64_t AccessSize,
|
|
232 const Value *AllocaPtr, uint64_t AllocaSize) {
|
121
|
233 AllocaOffsetRewriter Rewriter(SE, AllocaPtr);
|
|
234 const SCEV *Expr = Rewriter.visit(SE.getSCEV(Addr));
|
120
|
235
|
121
|
236 uint64_t BitWidth = SE.getTypeSizeInBits(Expr->getType());
|
|
237 ConstantRange AccessStartRange = SE.getUnsignedRange(Expr);
|
120
|
238 ConstantRange SizeRange =
|
|
239 ConstantRange(APInt(BitWidth, 0), APInt(BitWidth, AccessSize));
|
|
240 ConstantRange AccessRange = AccessStartRange.add(SizeRange);
|
|
241 ConstantRange AllocaRange =
|
|
242 ConstantRange(APInt(BitWidth, 0), APInt(BitWidth, AllocaSize));
|
|
243 bool Safe = AllocaRange.contains(AccessRange);
|
|
244
|
|
245 DEBUG(dbgs() << "[SafeStack] "
|
|
246 << (isa<AllocaInst>(AllocaPtr) ? "Alloca " : "ByValArgument ")
|
|
247 << *AllocaPtr << "\n"
|
|
248 << " Access " << *Addr << "\n"
|
|
249 << " SCEV " << *Expr
|
121
|
250 << " U: " << SE.getUnsignedRange(Expr)
|
|
251 << ", S: " << SE.getSignedRange(Expr) << "\n"
|
120
|
252 << " Range " << AccessRange << "\n"
|
|
253 << " AllocaRange " << AllocaRange << "\n"
|
|
254 << " " << (Safe ? "safe" : "unsafe") << "\n");
|
|
255
|
|
256 return Safe;
|
|
257 }
|
|
258
|
|
259 bool SafeStack::IsMemIntrinsicSafe(const MemIntrinsic *MI, const Use &U,
|
|
260 const Value *AllocaPtr,
|
|
261 uint64_t AllocaSize) {
|
|
262 // All MemIntrinsics have destination address in Arg0 and size in Arg2.
|
|
263 if (MI->getRawDest() != U) return true;
|
|
264 const auto *Len = dyn_cast<ConstantInt>(MI->getLength());
|
|
265 // Non-constant size => unsafe. FIXME: try SCEV getRange.
|
|
266 if (!Len) return false;
|
|
267 return IsAccessSafe(U, Len->getZExtValue(), AllocaPtr, AllocaSize);
|
|
268 }
|
|
269
|
|
270 /// Check whether a given allocation must be put on the safe
|
|
271 /// stack or not. The function analyzes all uses of AI and checks whether it is
|
|
272 /// only accessed in a memory safe way (as decided statically).
|
|
273 bool SafeStack::IsSafeStackAlloca(const Value *AllocaPtr, uint64_t AllocaSize) {
|
|
274 // Go through all uses of this alloca and check whether all accesses to the
|
|
275 // allocated object are statically known to be memory safe and, hence, the
|
|
276 // object can be placed on the safe stack.
|
|
277 SmallPtrSet<const Value *, 16> Visited;
|
|
278 SmallVector<const Value *, 8> WorkList;
|
|
279 WorkList.push_back(AllocaPtr);
|
|
280
|
|
281 // A DFS search through all uses of the alloca in bitcasts/PHI/GEPs/etc.
|
|
282 while (!WorkList.empty()) {
|
|
283 const Value *V = WorkList.pop_back_val();
|
|
284 for (const Use &UI : V->uses()) {
|
|
285 auto I = cast<const Instruction>(UI.getUser());
|
|
286 assert(V == UI.get());
|
|
287
|
|
288 switch (I->getOpcode()) {
|
121
|
289 case Instruction::Load:
|
|
290 if (!IsAccessSafe(UI, DL.getTypeStoreSize(I->getType()), AllocaPtr,
|
120
|
291 AllocaSize))
|
|
292 return false;
|
|
293 break;
|
121
|
294
|
120
|
295 case Instruction::VAArg:
|
|
296 // "va-arg" from a pointer is safe.
|
|
297 break;
|
121
|
298 case Instruction::Store:
|
120
|
299 if (V == I->getOperand(0)) {
|
|
300 // Stored the pointer - conservatively assume it may be unsafe.
|
|
301 DEBUG(dbgs() << "[SafeStack] Unsafe alloca: " << *AllocaPtr
|
|
302 << "\n store of address: " << *I << "\n");
|
|
303 return false;
|
|
304 }
|
|
305
|
121
|
306 if (!IsAccessSafe(UI, DL.getTypeStoreSize(I->getOperand(0)->getType()),
|
120
|
307 AllocaPtr, AllocaSize))
|
|
308 return false;
|
|
309 break;
|
121
|
310
|
|
311 case Instruction::Ret:
|
120
|
312 // Information leak.
|
|
313 return false;
|
|
314
|
|
315 case Instruction::Call:
|
|
316 case Instruction::Invoke: {
|
|
317 ImmutableCallSite CS(I);
|
|
318
|
|
319 if (const IntrinsicInst *II = dyn_cast<IntrinsicInst>(I)) {
|
|
320 if (II->getIntrinsicID() == Intrinsic::lifetime_start ||
|
|
321 II->getIntrinsicID() == Intrinsic::lifetime_end)
|
|
322 continue;
|
|
323 }
|
|
324
|
|
325 if (const MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) {
|
|
326 if (!IsMemIntrinsicSafe(MI, UI, AllocaPtr, AllocaSize)) {
|
|
327 DEBUG(dbgs() << "[SafeStack] Unsafe alloca: " << *AllocaPtr
|
|
328 << "\n unsafe memintrinsic: " << *I
|
|
329 << "\n");
|
|
330 return false;
|
|
331 }
|
|
332 continue;
|
|
333 }
|
|
334
|
|
335 // LLVM 'nocapture' attribute is only set for arguments whose address
|
|
336 // is not stored, passed around, or used in any other non-trivial way.
|
|
337 // We assume that passing a pointer to an object as a 'nocapture
|
|
338 // readnone' argument is safe.
|
|
339 // FIXME: a more precise solution would require an interprocedural
|
|
340 // analysis here, which would look at all uses of an argument inside
|
|
341 // the function being called.
|
|
342 ImmutableCallSite::arg_iterator B = CS.arg_begin(), E = CS.arg_end();
|
|
343 for (ImmutableCallSite::arg_iterator A = B; A != E; ++A)
|
|
344 if (A->get() == V)
|
|
345 if (!(CS.doesNotCapture(A - B) && (CS.doesNotAccessMemory(A - B) ||
|
|
346 CS.doesNotAccessMemory()))) {
|
|
347 DEBUG(dbgs() << "[SafeStack] Unsafe alloca: " << *AllocaPtr
|
|
348 << "\n unsafe call: " << *I << "\n");
|
|
349 return false;
|
|
350 }
|
|
351 continue;
|
|
352 }
|
|
353
|
|
354 default:
|
|
355 if (Visited.insert(I).second)
|
|
356 WorkList.push_back(cast<const Instruction>(I));
|
|
357 }
|
|
358 }
|
|
359 }
|
|
360
|
|
361 // All uses of the alloca are safe, we can place it on the safe stack.
|
|
362 return true;
|
|
363 }
|
|
364
|
|
365 Value *SafeStack::getStackGuard(IRBuilder<> &IRB, Function &F) {
|
121
|
366 Value *StackGuardVar = TL.getIRStackGuard(IRB);
|
120
|
367 if (!StackGuardVar)
|
|
368 StackGuardVar =
|
|
369 F.getParent()->getOrInsertGlobal("__stack_chk_guard", StackPtrTy);
|
|
370 return IRB.CreateLoad(StackGuardVar, "StackGuard");
|
|
371 }
|
|
372
|
|
373 void SafeStack::findInsts(Function &F,
|
|
374 SmallVectorImpl<AllocaInst *> &StaticAllocas,
|
|
375 SmallVectorImpl<AllocaInst *> &DynamicAllocas,
|
|
376 SmallVectorImpl<Argument *> &ByValArguments,
|
|
377 SmallVectorImpl<ReturnInst *> &Returns,
|
|
378 SmallVectorImpl<Instruction *> &StackRestorePoints) {
|
|
379 for (Instruction &I : instructions(&F)) {
|
|
380 if (auto AI = dyn_cast<AllocaInst>(&I)) {
|
|
381 ++NumAllocas;
|
|
382
|
|
383 uint64_t Size = getStaticAllocaAllocationSize(AI);
|
|
384 if (IsSafeStackAlloca(AI, Size))
|
|
385 continue;
|
|
386
|
|
387 if (AI->isStaticAlloca()) {
|
|
388 ++NumUnsafeStaticAllocas;
|
|
389 StaticAllocas.push_back(AI);
|
|
390 } else {
|
|
391 ++NumUnsafeDynamicAllocas;
|
|
392 DynamicAllocas.push_back(AI);
|
|
393 }
|
|
394 } else if (auto RI = dyn_cast<ReturnInst>(&I)) {
|
|
395 Returns.push_back(RI);
|
|
396 } else if (auto CI = dyn_cast<CallInst>(&I)) {
|
|
397 // setjmps require stack restore.
|
|
398 if (CI->getCalledFunction() && CI->canReturnTwice())
|
|
399 StackRestorePoints.push_back(CI);
|
|
400 } else if (auto LP = dyn_cast<LandingPadInst>(&I)) {
|
|
401 // Exception landing pads require stack restore.
|
|
402 StackRestorePoints.push_back(LP);
|
|
403 } else if (auto II = dyn_cast<IntrinsicInst>(&I)) {
|
|
404 if (II->getIntrinsicID() == Intrinsic::gcroot)
|
121
|
405 report_fatal_error(
|
120
|
406 "gcroot intrinsic not compatible with safestack attribute");
|
|
407 }
|
|
408 }
|
|
409 for (Argument &Arg : F.args()) {
|
|
410 if (!Arg.hasByValAttr())
|
|
411 continue;
|
|
412 uint64_t Size =
|
121
|
413 DL.getTypeStoreSize(Arg.getType()->getPointerElementType());
|
120
|
414 if (IsSafeStackAlloca(&Arg, Size))
|
|
415 continue;
|
|
416
|
|
417 ++NumUnsafeByValArguments;
|
|
418 ByValArguments.push_back(&Arg);
|
|
419 }
|
|
420 }
|
|
421
|
|
422 AllocaInst *
|
|
423 SafeStack::createStackRestorePoints(IRBuilder<> &IRB, Function &F,
|
|
424 ArrayRef<Instruction *> StackRestorePoints,
|
|
425 Value *StaticTop, bool NeedDynamicTop) {
|
|
426 assert(StaticTop && "The stack top isn't set.");
|
|
427
|
|
428 if (StackRestorePoints.empty())
|
|
429 return nullptr;
|
|
430
|
|
431 // We need the current value of the shadow stack pointer to restore
|
|
432 // after longjmp or exception catching.
|
|
433
|
|
434 // FIXME: On some platforms this could be handled by the longjmp/exception
|
|
435 // runtime itself.
|
|
436
|
|
437 AllocaInst *DynamicTop = nullptr;
|
|
438 if (NeedDynamicTop) {
|
|
439 // If we also have dynamic alloca's, the stack pointer value changes
|
|
440 // throughout the function. For now we store it in an alloca.
|
|
441 DynamicTop = IRB.CreateAlloca(StackPtrTy, /*ArraySize=*/nullptr,
|
|
442 "unsafe_stack_dynamic_ptr");
|
|
443 IRB.CreateStore(StaticTop, DynamicTop);
|
|
444 }
|
|
445
|
|
446 // Restore current stack pointer after longjmp/exception catch.
|
|
447 for (Instruction *I : StackRestorePoints) {
|
|
448 ++NumUnsafeStackRestorePoints;
|
|
449
|
|
450 IRB.SetInsertPoint(I->getNextNode());
|
|
451 Value *CurrentTop = DynamicTop ? IRB.CreateLoad(DynamicTop) : StaticTop;
|
|
452 IRB.CreateStore(CurrentTop, UnsafeStackPtr);
|
|
453 }
|
|
454
|
|
455 return DynamicTop;
|
|
456 }
|
|
457
|
|
458 void SafeStack::checkStackGuard(IRBuilder<> &IRB, Function &F, ReturnInst &RI,
|
|
459 AllocaInst *StackGuardSlot, Value *StackGuard) {
|
|
460 Value *V = IRB.CreateLoad(StackGuardSlot);
|
|
461 Value *Cmp = IRB.CreateICmpNE(StackGuard, V);
|
|
462
|
|
463 auto SuccessProb = BranchProbabilityInfo::getBranchProbStackProtector(true);
|
|
464 auto FailureProb = BranchProbabilityInfo::getBranchProbStackProtector(false);
|
|
465 MDNode *Weights = MDBuilder(F.getContext())
|
|
466 .createBranchWeights(SuccessProb.getNumerator(),
|
|
467 FailureProb.getNumerator());
|
|
468 Instruction *CheckTerm =
|
|
469 SplitBlockAndInsertIfThen(Cmp, &RI,
|
|
470 /* Unreachable */ true, Weights);
|
|
471 IRBuilder<> IRBFail(CheckTerm);
|
|
472 // FIXME: respect -fsanitize-trap / -ftrap-function here?
|
|
473 Constant *StackChkFail = F.getParent()->getOrInsertFunction(
|
121
|
474 "__stack_chk_fail", IRB.getVoidTy());
|
120
|
475 IRBFail.CreateCall(StackChkFail, {});
|
|
476 }
|
|
477
|
|
478 /// We explicitly compute and set the unsafe stack layout for all unsafe
|
|
479 /// static alloca instructions. We save the unsafe "base pointer" in the
|
|
480 /// prologue into a local variable and restore it in the epilogue.
|
|
481 Value *SafeStack::moveStaticAllocasToUnsafeStack(
|
|
482 IRBuilder<> &IRB, Function &F, ArrayRef<AllocaInst *> StaticAllocas,
|
|
483 ArrayRef<Argument *> ByValArguments, ArrayRef<ReturnInst *> Returns,
|
|
484 Instruction *BasePointer, AllocaInst *StackGuardSlot) {
|
|
485 if (StaticAllocas.empty() && ByValArguments.empty())
|
|
486 return BasePointer;
|
|
487
|
|
488 DIBuilder DIB(*F.getParent());
|
|
489
|
|
490 StackColoring SSC(F, StaticAllocas);
|
|
491 SSC.run();
|
|
492 SSC.removeAllMarkers();
|
|
493
|
|
494 // Unsafe stack always grows down.
|
|
495 StackLayout SSL(StackAlignment);
|
|
496 if (StackGuardSlot) {
|
|
497 Type *Ty = StackGuardSlot->getAllocatedType();
|
|
498 unsigned Align =
|
121
|
499 std::max(DL.getPrefTypeAlignment(Ty), StackGuardSlot->getAlignment());
|
120
|
500 SSL.addObject(StackGuardSlot, getStaticAllocaAllocationSize(StackGuardSlot),
|
|
501 Align, SSC.getFullLiveRange());
|
|
502 }
|
|
503
|
|
504 for (Argument *Arg : ByValArguments) {
|
|
505 Type *Ty = Arg->getType()->getPointerElementType();
|
121
|
506 uint64_t Size = DL.getTypeStoreSize(Ty);
|
120
|
507 if (Size == 0)
|
|
508 Size = 1; // Don't create zero-sized stack objects.
|
|
509
|
|
510 // Ensure the object is properly aligned.
|
121
|
511 unsigned Align = std::max((unsigned)DL.getPrefTypeAlignment(Ty),
|
120
|
512 Arg->getParamAlignment());
|
|
513 SSL.addObject(Arg, Size, Align, SSC.getFullLiveRange());
|
|
514 }
|
|
515
|
|
516 for (AllocaInst *AI : StaticAllocas) {
|
|
517 Type *Ty = AI->getAllocatedType();
|
|
518 uint64_t Size = getStaticAllocaAllocationSize(AI);
|
|
519 if (Size == 0)
|
|
520 Size = 1; // Don't create zero-sized stack objects.
|
|
521
|
|
522 // Ensure the object is properly aligned.
|
|
523 unsigned Align =
|
121
|
524 std::max((unsigned)DL.getPrefTypeAlignment(Ty), AI->getAlignment());
|
120
|
525
|
|
526 SSL.addObject(AI, Size, Align, SSC.getLiveRange(AI));
|
|
527 }
|
|
528
|
|
529 SSL.computeLayout();
|
|
530 unsigned FrameAlignment = SSL.getFrameAlignment();
|
|
531
|
|
532 // FIXME: tell SSL that we start at a less-then-MaxAlignment aligned location
|
|
533 // (AlignmentSkew).
|
|
534 if (FrameAlignment > StackAlignment) {
|
|
535 // Re-align the base pointer according to the max requested alignment.
|
|
536 assert(isPowerOf2_32(FrameAlignment));
|
|
537 IRB.SetInsertPoint(BasePointer->getNextNode());
|
|
538 BasePointer = cast<Instruction>(IRB.CreateIntToPtr(
|
|
539 IRB.CreateAnd(IRB.CreatePtrToInt(BasePointer, IntPtrTy),
|
|
540 ConstantInt::get(IntPtrTy, ~uint64_t(FrameAlignment - 1))),
|
|
541 StackPtrTy));
|
|
542 }
|
|
543
|
|
544 IRB.SetInsertPoint(BasePointer->getNextNode());
|
|
545
|
|
546 if (StackGuardSlot) {
|
|
547 unsigned Offset = SSL.getObjectOffset(StackGuardSlot);
|
|
548 Value *Off = IRB.CreateGEP(BasePointer, // BasePointer is i8*
|
|
549 ConstantInt::get(Int32Ty, -Offset));
|
|
550 Value *NewAI =
|
|
551 IRB.CreateBitCast(Off, StackGuardSlot->getType(), "StackGuardSlot");
|
|
552
|
|
553 // Replace alloc with the new location.
|
|
554 StackGuardSlot->replaceAllUsesWith(NewAI);
|
|
555 StackGuardSlot->eraseFromParent();
|
|
556 }
|
|
557
|
|
558 for (Argument *Arg : ByValArguments) {
|
|
559 unsigned Offset = SSL.getObjectOffset(Arg);
|
134
|
560 unsigned Align = SSL.getObjectAlignment(Arg);
|
120
|
561 Type *Ty = Arg->getType()->getPointerElementType();
|
|
562
|
121
|
563 uint64_t Size = DL.getTypeStoreSize(Ty);
|
120
|
564 if (Size == 0)
|
|
565 Size = 1; // Don't create zero-sized stack objects.
|
|
566
|
|
567 Value *Off = IRB.CreateGEP(BasePointer, // BasePointer is i8*
|
|
568 ConstantInt::get(Int32Ty, -Offset));
|
|
569 Value *NewArg = IRB.CreateBitCast(Off, Arg->getType(),
|
|
570 Arg->getName() + ".unsafe-byval");
|
|
571
|
|
572 // Replace alloc with the new location.
|
|
573 replaceDbgDeclare(Arg, BasePointer, BasePointer->getNextNode(), DIB,
|
134
|
574 DIExpression::NoDeref, -Offset, DIExpression::NoDeref);
|
120
|
575 Arg->replaceAllUsesWith(NewArg);
|
|
576 IRB.SetInsertPoint(cast<Instruction>(NewArg)->getNextNode());
|
134
|
577 IRB.CreateMemCpy(Off, Align, Arg, Arg->getParamAlignment(), Size);
|
120
|
578 }
|
|
579
|
|
580 // Allocate space for every unsafe static AllocaInst on the unsafe stack.
|
|
581 for (AllocaInst *AI : StaticAllocas) {
|
|
582 IRB.SetInsertPoint(AI);
|
|
583 unsigned Offset = SSL.getObjectOffset(AI);
|
|
584
|
|
585 uint64_t Size = getStaticAllocaAllocationSize(AI);
|
|
586 if (Size == 0)
|
|
587 Size = 1; // Don't create zero-sized stack objects.
|
|
588
|
134
|
589 replaceDbgDeclareForAlloca(AI, BasePointer, DIB, DIExpression::NoDeref,
|
|
590 -Offset, DIExpression::NoDeref);
|
120
|
591 replaceDbgValueForAlloca(AI, BasePointer, DIB, -Offset);
|
|
592
|
|
593 // Replace uses of the alloca with the new location.
|
|
594 // Insert address calculation close to each use to work around PR27844.
|
|
595 std::string Name = std::string(AI->getName()) + ".unsafe";
|
|
596 while (!AI->use_empty()) {
|
|
597 Use &U = *AI->use_begin();
|
|
598 Instruction *User = cast<Instruction>(U.getUser());
|
|
599
|
|
600 Instruction *InsertBefore;
|
|
601 if (auto *PHI = dyn_cast<PHINode>(User))
|
|
602 InsertBefore = PHI->getIncomingBlock(U)->getTerminator();
|
|
603 else
|
|
604 InsertBefore = User;
|
|
605
|
|
606 IRBuilder<> IRBUser(InsertBefore);
|
|
607 Value *Off = IRBUser.CreateGEP(BasePointer, // BasePointer is i8*
|
|
608 ConstantInt::get(Int32Ty, -Offset));
|
|
609 Value *Replacement = IRBUser.CreateBitCast(Off, AI->getType(), Name);
|
|
610
|
|
611 if (auto *PHI = dyn_cast<PHINode>(User)) {
|
|
612 // PHI nodes may have multiple incoming edges from the same BB (why??),
|
|
613 // all must be updated at once with the same incoming value.
|
|
614 auto *BB = PHI->getIncomingBlock(U);
|
|
615 for (unsigned I = 0; I < PHI->getNumIncomingValues(); ++I)
|
|
616 if (PHI->getIncomingBlock(I) == BB)
|
|
617 PHI->setIncomingValue(I, Replacement);
|
|
618 } else {
|
|
619 U.set(Replacement);
|
|
620 }
|
|
621 }
|
|
622
|
|
623 AI->eraseFromParent();
|
|
624 }
|
|
625
|
|
626 // Re-align BasePointer so that our callees would see it aligned as
|
|
627 // expected.
|
|
628 // FIXME: no need to update BasePointer in leaf functions.
|
|
629 unsigned FrameSize = alignTo(SSL.getFrameSize(), StackAlignment);
|
|
630
|
|
631 // Update shadow stack pointer in the function epilogue.
|
|
632 IRB.SetInsertPoint(BasePointer->getNextNode());
|
|
633
|
|
634 Value *StaticTop =
|
|
635 IRB.CreateGEP(BasePointer, ConstantInt::get(Int32Ty, -FrameSize),
|
|
636 "unsafe_stack_static_top");
|
|
637 IRB.CreateStore(StaticTop, UnsafeStackPtr);
|
|
638 return StaticTop;
|
|
639 }
|
|
640
|
|
641 void SafeStack::moveDynamicAllocasToUnsafeStack(
|
|
642 Function &F, Value *UnsafeStackPtr, AllocaInst *DynamicTop,
|
|
643 ArrayRef<AllocaInst *> DynamicAllocas) {
|
|
644 DIBuilder DIB(*F.getParent());
|
|
645
|
|
646 for (AllocaInst *AI : DynamicAllocas) {
|
|
647 IRBuilder<> IRB(AI);
|
|
648
|
|
649 // Compute the new SP value (after AI).
|
|
650 Value *ArraySize = AI->getArraySize();
|
|
651 if (ArraySize->getType() != IntPtrTy)
|
|
652 ArraySize = IRB.CreateIntCast(ArraySize, IntPtrTy, false);
|
|
653
|
|
654 Type *Ty = AI->getAllocatedType();
|
121
|
655 uint64_t TySize = DL.getTypeAllocSize(Ty);
|
120
|
656 Value *Size = IRB.CreateMul(ArraySize, ConstantInt::get(IntPtrTy, TySize));
|
|
657
|
|
658 Value *SP = IRB.CreatePtrToInt(IRB.CreateLoad(UnsafeStackPtr), IntPtrTy);
|
|
659 SP = IRB.CreateSub(SP, Size);
|
|
660
|
|
661 // Align the SP value to satisfy the AllocaInst, type and stack alignments.
|
|
662 unsigned Align = std::max(
|
121
|
663 std::max((unsigned)DL.getPrefTypeAlignment(Ty), AI->getAlignment()),
|
120
|
664 (unsigned)StackAlignment);
|
|
665
|
|
666 assert(isPowerOf2_32(Align));
|
|
667 Value *NewTop = IRB.CreateIntToPtr(
|
|
668 IRB.CreateAnd(SP, ConstantInt::get(IntPtrTy, ~uint64_t(Align - 1))),
|
|
669 StackPtrTy);
|
|
670
|
|
671 // Save the stack pointer.
|
|
672 IRB.CreateStore(NewTop, UnsafeStackPtr);
|
|
673 if (DynamicTop)
|
|
674 IRB.CreateStore(NewTop, DynamicTop);
|
|
675
|
|
676 Value *NewAI = IRB.CreatePointerCast(NewTop, AI->getType());
|
|
677 if (AI->hasName() && isa<Instruction>(NewAI))
|
|
678 NewAI->takeName(AI);
|
|
679
|
134
|
680 replaceDbgDeclareForAlloca(AI, NewAI, DIB, DIExpression::NoDeref, 0,
|
|
681 DIExpression::NoDeref);
|
120
|
682 AI->replaceAllUsesWith(NewAI);
|
|
683 AI->eraseFromParent();
|
|
684 }
|
|
685
|
|
686 if (!DynamicAllocas.empty()) {
|
|
687 // Now go through the instructions again, replacing stacksave/stackrestore.
|
|
688 for (inst_iterator It = inst_begin(&F), Ie = inst_end(&F); It != Ie;) {
|
|
689 Instruction *I = &*(It++);
|
|
690 auto II = dyn_cast<IntrinsicInst>(I);
|
|
691 if (!II)
|
|
692 continue;
|
|
693
|
|
694 if (II->getIntrinsicID() == Intrinsic::stacksave) {
|
|
695 IRBuilder<> IRB(II);
|
|
696 Instruction *LI = IRB.CreateLoad(UnsafeStackPtr);
|
|
697 LI->takeName(II);
|
|
698 II->replaceAllUsesWith(LI);
|
|
699 II->eraseFromParent();
|
|
700 } else if (II->getIntrinsicID() == Intrinsic::stackrestore) {
|
|
701 IRBuilder<> IRB(II);
|
|
702 Instruction *SI = IRB.CreateStore(II->getArgOperand(0), UnsafeStackPtr);
|
|
703 SI->takeName(II);
|
|
704 assert(II->use_empty());
|
|
705 II->eraseFromParent();
|
|
706 }
|
|
707 }
|
|
708 }
|
|
709 }
|
|
710
|
134
|
711 bool SafeStack::ShouldInlinePointerAddress(CallSite &CS) {
|
|
712 Function *Callee = CS.getCalledFunction();
|
|
713 if (CS.hasFnAttr(Attribute::AlwaysInline) && isInlineViable(*Callee))
|
|
714 return true;
|
|
715 if (Callee->isInterposable() || Callee->hasFnAttribute(Attribute::NoInline) ||
|
|
716 CS.isNoInline())
|
|
717 return false;
|
|
718 return true;
|
|
719 }
|
|
720
|
|
721 void SafeStack::TryInlinePointerAddress() {
|
|
722 if (!isa<CallInst>(UnsafeStackPtr))
|
|
723 return;
|
|
724
|
|
725 if(F.hasFnAttribute(Attribute::OptimizeNone))
|
|
726 return;
|
|
727
|
|
728 CallSite CS(UnsafeStackPtr);
|
|
729 Function *Callee = CS.getCalledFunction();
|
|
730 if (!Callee || Callee->isDeclaration())
|
|
731 return;
|
|
732
|
|
733 if (!ShouldInlinePointerAddress(CS))
|
|
734 return;
|
|
735
|
|
736 InlineFunctionInfo IFI;
|
|
737 InlineFunction(CS, IFI);
|
|
738 }
|
|
739
|
121
|
740 bool SafeStack::run() {
|
|
741 assert(F.hasFnAttribute(Attribute::SafeStack) &&
|
|
742 "Can't run SafeStack on a function without the attribute");
|
|
743 assert(!F.isDeclaration() && "Can't run SafeStack on a function declaration");
|
120
|
744
|
|
745 ++NumFunctions;
|
|
746
|
|
747 SmallVector<AllocaInst *, 16> StaticAllocas;
|
|
748 SmallVector<AllocaInst *, 4> DynamicAllocas;
|
|
749 SmallVector<Argument *, 4> ByValArguments;
|
|
750 SmallVector<ReturnInst *, 4> Returns;
|
|
751
|
|
752 // Collect all points where stack gets unwound and needs to be restored
|
|
753 // This is only necessary because the runtime (setjmp and unwind code) is
|
|
754 // not aware of the unsafe stack and won't unwind/restore it properly.
|
|
755 // To work around this problem without changing the runtime, we insert
|
|
756 // instrumentation to restore the unsafe stack pointer when necessary.
|
|
757 SmallVector<Instruction *, 4> StackRestorePoints;
|
|
758
|
|
759 // Find all static and dynamic alloca instructions that must be moved to the
|
|
760 // unsafe stack, all return instructions and stack restore points.
|
|
761 findInsts(F, StaticAllocas, DynamicAllocas, ByValArguments, Returns,
|
|
762 StackRestorePoints);
|
|
763
|
|
764 if (StaticAllocas.empty() && DynamicAllocas.empty() &&
|
|
765 ByValArguments.empty() && StackRestorePoints.empty())
|
|
766 return false; // Nothing to do in this function.
|
|
767
|
|
768 if (!StaticAllocas.empty() || !DynamicAllocas.empty() ||
|
|
769 !ByValArguments.empty())
|
|
770 ++NumUnsafeStackFunctions; // This function has the unsafe stack.
|
|
771
|
|
772 if (!StackRestorePoints.empty())
|
|
773 ++NumUnsafeStackRestorePointsFunctions;
|
|
774
|
|
775 IRBuilder<> IRB(&F.front(), F.begin()->getFirstInsertionPt());
|
134
|
776 if (SafeStackUsePointerAddress) {
|
|
777 Value *Fn = F.getParent()->getOrInsertFunction(
|
|
778 "__safestack_pointer_address", StackPtrTy->getPointerTo(0));
|
|
779 UnsafeStackPtr = IRB.CreateCall(Fn);
|
|
780 } else {
|
|
781 UnsafeStackPtr = TL.getSafeStackPointerLocation(IRB);
|
|
782 }
|
120
|
783
|
|
784 // Load the current stack pointer (we'll also use it as a base pointer).
|
|
785 // FIXME: use a dedicated register for it ?
|
|
786 Instruction *BasePointer =
|
|
787 IRB.CreateLoad(UnsafeStackPtr, false, "unsafe_stack_ptr");
|
|
788 assert(BasePointer->getType() == StackPtrTy);
|
|
789
|
|
790 AllocaInst *StackGuardSlot = nullptr;
|
|
791 // FIXME: implement weaker forms of stack protector.
|
|
792 if (F.hasFnAttribute(Attribute::StackProtect) ||
|
|
793 F.hasFnAttribute(Attribute::StackProtectStrong) ||
|
|
794 F.hasFnAttribute(Attribute::StackProtectReq)) {
|
|
795 Value *StackGuard = getStackGuard(IRB, F);
|
|
796 StackGuardSlot = IRB.CreateAlloca(StackPtrTy, nullptr);
|
|
797 IRB.CreateStore(StackGuard, StackGuardSlot);
|
|
798
|
|
799 for (ReturnInst *RI : Returns) {
|
|
800 IRBuilder<> IRBRet(RI);
|
|
801 checkStackGuard(IRBRet, F, *RI, StackGuardSlot, StackGuard);
|
|
802 }
|
|
803 }
|
|
804
|
|
805 // The top of the unsafe stack after all unsafe static allocas are
|
|
806 // allocated.
|
|
807 Value *StaticTop =
|
|
808 moveStaticAllocasToUnsafeStack(IRB, F, StaticAllocas, ByValArguments,
|
|
809 Returns, BasePointer, StackGuardSlot);
|
|
810
|
|
811 // Safe stack object that stores the current unsafe stack top. It is updated
|
|
812 // as unsafe dynamic (non-constant-sized) allocas are allocated and freed.
|
|
813 // This is only needed if we need to restore stack pointer after longjmp
|
|
814 // or exceptions, and we have dynamic allocations.
|
|
815 // FIXME: a better alternative might be to store the unsafe stack pointer
|
|
816 // before setjmp / invoke instructions.
|
|
817 AllocaInst *DynamicTop = createStackRestorePoints(
|
|
818 IRB, F, StackRestorePoints, StaticTop, !DynamicAllocas.empty());
|
|
819
|
|
820 // Handle dynamic allocas.
|
|
821 moveDynamicAllocasToUnsafeStack(F, UnsafeStackPtr, DynamicTop,
|
|
822 DynamicAllocas);
|
|
823
|
|
824 // Restore the unsafe stack pointer before each return.
|
|
825 for (ReturnInst *RI : Returns) {
|
|
826 IRB.SetInsertPoint(RI);
|
|
827 IRB.CreateStore(BasePointer, UnsafeStackPtr);
|
|
828 }
|
|
829
|
134
|
830 TryInlinePointerAddress();
|
|
831
|
120
|
832 DEBUG(dbgs() << "[SafeStack] safestack applied\n");
|
|
833 return true;
|
|
834 }
|
|
835
|
121
|
836 class SafeStackLegacyPass : public FunctionPass {
|
|
837 const TargetMachine *TM = nullptr;
|
|
838
|
|
839 public:
|
|
840 static char ID; // Pass identification, replacement for typeid..
|
|
841
|
|
842 SafeStackLegacyPass() : FunctionPass(ID) {
|
|
843 initializeSafeStackLegacyPassPass(*PassRegistry::getPassRegistry());
|
|
844 }
|
|
845
|
|
846 void getAnalysisUsage(AnalysisUsage &AU) const override {
|
|
847 AU.addRequired<TargetPassConfig>();
|
|
848 AU.addRequired<TargetLibraryInfoWrapperPass>();
|
|
849 AU.addRequired<AssumptionCacheTracker>();
|
|
850 }
|
|
851
|
|
852 bool runOnFunction(Function &F) override {
|
|
853 DEBUG(dbgs() << "[SafeStack] Function: " << F.getName() << "\n");
|
|
854
|
|
855 if (!F.hasFnAttribute(Attribute::SafeStack)) {
|
|
856 DEBUG(dbgs() << "[SafeStack] safestack is not requested"
|
|
857 " for this function\n");
|
|
858 return false;
|
|
859 }
|
|
860
|
|
861 if (F.isDeclaration()) {
|
|
862 DEBUG(dbgs() << "[SafeStack] function definition"
|
|
863 " is not available\n");
|
|
864 return false;
|
|
865 }
|
120
|
866
|
121
|
867 TM = &getAnalysis<TargetPassConfig>().getTM<TargetMachine>();
|
|
868 auto *TL = TM->getSubtargetImpl(F)->getTargetLowering();
|
|
869 if (!TL)
|
|
870 report_fatal_error("TargetLowering instance is required");
|
|
871
|
|
872 auto *DL = &F.getParent()->getDataLayout();
|
|
873 auto &TLI = getAnalysis<TargetLibraryInfoWrapperPass>().getTLI();
|
|
874 auto &ACT = getAnalysis<AssumptionCacheTracker>().getAssumptionCache(F);
|
|
875
|
|
876 // Compute DT and LI only for functions that have the attribute.
|
|
877 // This is only useful because the legacy pass manager doesn't let us
|
|
878 // compute analyzes lazily.
|
|
879 // In the backend pipeline, nothing preserves DT before SafeStack, so we
|
|
880 // would otherwise always compute it wastefully, even if there is no
|
|
881 // function with the safestack attribute.
|
|
882 DominatorTree DT(F);
|
|
883 LoopInfo LI(DT);
|
120
|
884
|
121
|
885 ScalarEvolution SE(F, TLI, ACT, DT, LI);
|
|
886
|
|
887 return SafeStack(F, *TL, *DL, SE).run();
|
|
888 }
|
|
889 };
|
|
890
|
|
891 } // end anonymous namespace
|
|
892
|
|
893 char SafeStackLegacyPass::ID = 0;
|
|
894
|
|
895 INITIALIZE_PASS_BEGIN(SafeStackLegacyPass, DEBUG_TYPE,
|
|
896 "Safe Stack instrumentation pass", false, false)
|
|
897 INITIALIZE_PASS_DEPENDENCY(TargetPassConfig)
|
|
898 INITIALIZE_PASS_END(SafeStackLegacyPass, DEBUG_TYPE,
|
|
899 "Safe Stack instrumentation pass", false, false)
|
|
900
|
|
901 FunctionPass *llvm::createSafeStackPass() { return new SafeStackLegacyPass(); }
|