blob: e22e9fdac95c2612ce4ee8a60ce65567fe2c17f3 [file] [log] [blame]
//===- RewriteAtomics.cpp - Stabilize instructions used for concurrency ---===//
//
// The LLVM Compiler Infrastructure
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
//
//===----------------------------------------------------------------------===//
//
// This pass encodes atomics, volatiles and fences using NaCl intrinsics
// instead of LLVM's regular IR instructions.
//
// All of the above are transformed into one of the
// @llvm.nacl.atomic.* intrinsics.
//
//===----------------------------------------------------------------------===//
#include "llvm/ADT/Twine.h"
#include "llvm/IR/DataLayout.h"
#include "llvm/IR/Function.h"
#include "llvm/IR/InlineAsm.h"
#include "llvm/IR/Instructions.h"
#include "llvm/IR/Intrinsics.h"
#include "llvm/IR/Module.h"
#include "llvm/IR/NaClAtomicIntrinsics.h"
#include "llvm/InstVisitor.h"
#include "llvm/Pass.h"
#include "llvm/Support/Compiler.h"
#include "llvm/Support/raw_ostream.h"
#include "llvm/Transforms/NaCl.h"
#include <climits>
#include <string>
using namespace llvm;
namespace {
class RewriteAtomics : public ModulePass {
public:
static char ID; // Pass identification, replacement for typeid
RewriteAtomics() : ModulePass(ID) {
// This is a module pass because it may have to introduce
// intrinsic declarations into the module and modify a global function.
initializeRewriteAtomicsPass(*PassRegistry::getPassRegistry());
}
virtual bool runOnModule(Module &M);
virtual void getAnalysisUsage(AnalysisUsage &Info) const {
Info.addRequired<DataLayout>();
}
};
template <class T> std::string ToStr(const T &V) {
std::string S;
raw_string_ostream OS(S);
OS << const_cast<T &>(V);
return OS.str();
}
class AtomicVisitor : public InstVisitor<AtomicVisitor> {
public:
AtomicVisitor(Module &M, Pass &P)
: M(M), C(M.getContext()), TD(P.getAnalysis<DataLayout>()), AI(C),
ModifiedModule(false) {}
~AtomicVisitor() {}
bool modifiedModule() const { return ModifiedModule; }
void visitLoadInst(LoadInst &I);
void visitStoreInst(StoreInst &I);
void visitAtomicCmpXchgInst(AtomicCmpXchgInst &I);
void visitAtomicRMWInst(AtomicRMWInst &I);
void visitFenceInst(FenceInst &I);
private:
Module &M;
LLVMContext &C;
const DataLayout TD;
NaCl::AtomicIntrinsics AI;
bool ModifiedModule;
AtomicVisitor() LLVM_DELETED_FUNCTION;
AtomicVisitor(const AtomicVisitor &) LLVM_DELETED_FUNCTION;
AtomicVisitor &operator=(const AtomicVisitor &) LLVM_DELETED_FUNCTION;
/// Create an integer constant holding a NaCl::MemoryOrder that can be
/// passed as an argument to one of the @llvm.nacl.atomic.*
/// intrinsics. This function may strengthen the ordering initially
/// specified by the instruction \p I for stability purpose.
template <class Instruction>
ConstantInt *freezeMemoryOrder(const Instruction &I) const;
/// Sanity-check that instruction \p I which has pointer and value
/// parameters have matching sizes \p BitSize for the type-pointed-to
/// and the value's type \p T.
void checkSizeMatchesType(const Instruction &I, unsigned BitSize,
const Type *T) const;
/// Verify that loads and stores are at least naturally aligned. Use
/// byte alignment because converting to bits could truncate the
/// value.
void checkAlignment(const Instruction &I, unsigned ByteAlignment,
unsigned ByteSize) const;
/// Create a cast before Instruction \p I from \p Src to \p Dst with \p Name.
CastInst *createCast(Instruction &I, Value *Src, Type *Dst, Twine Name) const;
/// Helper function which rewrites a single instruction \p I to a
/// particular intrinsic \p ID with overloaded type \p OverloadedType,
/// and argument list \p Args. Will perform a bitcast to the proper \p
/// DstType, if different from \p OverloadedType.
void replaceInstructionWithIntrinsicCall(Instruction &I, Intrinsic::ID ID,
Type *DstType, Type *OverloadedType,
ArrayRef<Value *> Args);
/// Most atomics instructions deal with at least one pointer, this
/// struct automates some of this and has generic sanity checks.
template <class Instruction> struct PointerHelper {
Value *P;
Type *OriginalPET;
Type *PET;
unsigned BitSize;
PointerHelper(const AtomicVisitor &AV, Instruction &I)
: P(I.getPointerOperand()) {
if (I.getPointerAddressSpace() != 0)
report_fatal_error("unhandled pointer address space " +
Twine(I.getPointerAddressSpace()) + " for atomic: " +
ToStr(I));
assert(P->getType()->isPointerTy() && "expected a pointer");
PET = OriginalPET = P->getType()->getPointerElementType();
BitSize = AV.TD.getTypeSizeInBits(OriginalPET);
if (!OriginalPET->isIntegerTy()) {
// The pointer wasn't to an integer type. We define atomics in
// terms of integers, so bitcast the pointer to an integer of
// the proper width.
Type *IntNPtr = Type::getIntNPtrTy(AV.C, BitSize);
P = AV.createCast(I, P, IntNPtr, P->getName() + ".cast");
PET = P->getType()->getPointerElementType();
}
AV.checkSizeMatchesType(I, BitSize, PET);
}
};
};
}
char RewriteAtomics::ID = 0;
INITIALIZE_PASS(RewriteAtomics, "nacl-rewrite-atomics",
"rewrite atomics, volatiles and fences into stable "
"@llvm.nacl.atomics.* intrinsics",
false, false)
bool RewriteAtomics::runOnModule(Module &M) {
AtomicVisitor AV(M, *this);
AV.visit(M);
return AV.modifiedModule();
}
template <class Instruction>
ConstantInt *AtomicVisitor::freezeMemoryOrder(const Instruction &I) const {
NaCl::MemoryOrder AO = NaCl::MemoryOrderInvalid;
// TODO Volatile load/store are promoted to sequentially consistent
// for now. We could do something weaker.
if (const LoadInst *L = dyn_cast<LoadInst>(&I)) {
if (L->isVolatile())
AO = NaCl::MemoryOrderSequentiallyConsistent;
} else if (const StoreInst *S = dyn_cast<StoreInst>(&I)) {
if (S->isVolatile())
AO = NaCl::MemoryOrderSequentiallyConsistent;
}
if (AO == NaCl::MemoryOrderInvalid) {
switch (I.getOrdering()) {
default:
case NotAtomic: llvm_unreachable("unexpected memory order");
// Monotonic is a strict superset of Unordered. Both can therefore
// map to Relaxed ordering, which is in the C11/C++11 standard.
case Unordered: AO = NaCl::MemoryOrderRelaxed; break;
case Monotonic: AO = NaCl::MemoryOrderRelaxed; break;
// TODO Consume is currently unspecified by LLVM's internal IR.
case Acquire: AO = NaCl::MemoryOrderAcquire; break;
case Release: AO = NaCl::MemoryOrderRelease; break;
case AcquireRelease: AO = NaCl::MemoryOrderAcquireRelease; break;
case SequentiallyConsistent:
AO = NaCl::MemoryOrderSequentiallyConsistent; break;
}
}
// TODO For now only sequential consistency is allowed.
AO = NaCl::MemoryOrderSequentiallyConsistent;
return ConstantInt::get(Type::getInt32Ty(C), AO);
}
void AtomicVisitor::checkSizeMatchesType(const Instruction &I, unsigned BitSize,
const Type *T) const {
Type *IntType = Type::getIntNTy(C, BitSize);
if (IntType && T == IntType)
return;
report_fatal_error("unsupported atomic type " + ToStr(*T) + " of size " +
Twine(BitSize) + " bits in: " + ToStr(I));
}
void AtomicVisitor::checkAlignment(const Instruction &I, unsigned ByteAlignment,
unsigned ByteSize) const {
if (ByteAlignment < ByteSize)
report_fatal_error("atomic load/store must be at least naturally aligned, "
"got " +
Twine(ByteAlignment) + ", bytes expected at least " +
Twine(ByteSize) + " bytes, in: " + ToStr(I));
}
CastInst *AtomicVisitor::createCast(Instruction &I, Value *Src, Type *Dst,
Twine Name) const {
Type *SrcT = Src->getType();
Instruction::CastOps Op = SrcT->isIntegerTy() && Dst->isPointerTy()
? Instruction::IntToPtr
: SrcT->isPointerTy() && Dst->isIntegerTy()
? Instruction::PtrToInt
: Instruction::BitCast;
if (!CastInst::castIsValid(Op, Src, Dst))
report_fatal_error("cannot emit atomic instruction while converting type " +
ToStr(*SrcT) + " to " + ToStr(*Dst) + " for " + Name +
" in " + ToStr(I));
return CastInst::Create(Op, Src, Dst, Name, &I);
}
void AtomicVisitor::replaceInstructionWithIntrinsicCall(
Instruction &I, Intrinsic::ID ID, Type *DstType, Type *OverloadedType,
ArrayRef<Value *> Args) {
std::string Name(I.getName());
Function *F = AI.find(ID, OverloadedType)->getDeclaration(&M);
CallInst *Call = CallInst::Create(F, Args, "", &I);
Instruction *Res = Call;
if (!Call->getType()->isVoidTy() && DstType != OverloadedType) {
// The call returns a value which needs to be cast to a non-integer.
Res = createCast(I, Call, DstType, Name + ".cast");
Res->setDebugLoc(I.getDebugLoc());
}
Call->setDebugLoc(I.getDebugLoc());
I.replaceAllUsesWith(Res);
I.eraseFromParent();
Call->setName(Name);
ModifiedModule = true;
}
/// %res = load {atomic|volatile} T* %ptr memory_order, align sizeof(T)
/// becomes:
/// %res = call T @llvm.nacl.atomic.load.i<size>(%ptr, memory_order)
void AtomicVisitor::visitLoadInst(LoadInst &I) {
if (I.isSimple())
return;
PointerHelper<LoadInst> PH(*this, I);
checkAlignment(I, I.getAlignment(), PH.BitSize / CHAR_BIT);
Value *Args[] = { PH.P, freezeMemoryOrder(I) };
replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_load,
PH.OriginalPET, PH.PET, Args);
}
/// store {atomic|volatile} T %val, T* %ptr memory_order, align sizeof(T)
/// becomes:
/// call void @llvm.nacl.atomic.store.i<size>(%val, %ptr, memory_order)
void AtomicVisitor::visitStoreInst(StoreInst &I) {
if (I.isSimple())
return;
PointerHelper<StoreInst> PH(*this, I);
checkAlignment(I, I.getAlignment(), PH.BitSize / CHAR_BIT);
Value *V = I.getValueOperand();
if (!V->getType()->isIntegerTy()) {
// The store isn't of an integer type. We define atomics in terms of
// integers, so bitcast the value to store to an integer of the
// proper width.
CastInst *Cast = createCast(I, V, Type::getIntNTy(C, PH.BitSize),
V->getName() + ".cast");
Cast->setDebugLoc(I.getDebugLoc());
V = Cast;
}
checkSizeMatchesType(I, PH.BitSize, V->getType());
Value *Args[] = { V, PH.P, freezeMemoryOrder(I) };
replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_store,
PH.OriginalPET, PH.PET, Args);
}
/// %res = atomicrmw OP T* %ptr, T %val memory_order
/// becomes:
/// %res = call T @llvm.nacl.atomic.rmw.i<size>(OP, %ptr, %val, memory_order)
void AtomicVisitor::visitAtomicRMWInst(AtomicRMWInst &I) {
NaCl::AtomicRMWOperation Op;
switch (I.getOperation()) {
default: report_fatal_error("unsupported atomicrmw operation: " + ToStr(I));
case AtomicRMWInst::Add: Op = NaCl::AtomicAdd; break;
case AtomicRMWInst::Sub: Op = NaCl::AtomicSub; break;
case AtomicRMWInst::And: Op = NaCl::AtomicAnd; break;
case AtomicRMWInst::Or: Op = NaCl::AtomicOr; break;
case AtomicRMWInst::Xor: Op = NaCl::AtomicXor; break;
case AtomicRMWInst::Xchg: Op = NaCl::AtomicExchange; break;
}
PointerHelper<AtomicRMWInst> PH(*this, I);
checkSizeMatchesType(I, PH.BitSize, I.getValOperand()->getType());
Value *Args[] = { ConstantInt::get(Type::getInt32Ty(C), Op), PH.P,
I.getValOperand(), freezeMemoryOrder(I) };
replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_rmw,
PH.OriginalPET, PH.PET, Args);
}
/// %res = cmpxchg T* %ptr, T %old, T %new memory_order
/// becomes:
/// %res = call T @llvm.nacl.atomic.cmpxchg.i<size>(
/// %object, %expected, %desired, memory_order_success,
/// memory_order_failure)
void AtomicVisitor::visitAtomicCmpXchgInst(AtomicCmpXchgInst &I) {
PointerHelper<AtomicCmpXchgInst> PH(*this, I);
checkSizeMatchesType(I, PH.BitSize, I.getCompareOperand()->getType());
checkSizeMatchesType(I, PH.BitSize, I.getNewValOperand()->getType());
// TODO LLVM currently doesn't support specifying separate memory
// orders for compare exchange's success and failure cases: LLVM
// IR implicitly drops the Release part of the specified memory
// order on failure.
Value *Args[] = { PH.P, I.getCompareOperand(), I.getNewValOperand(),
freezeMemoryOrder(I), freezeMemoryOrder(I) };
replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_cmpxchg,
PH.OriginalPET, PH.PET, Args);
}
/// fence memory_order
/// becomes:
/// call void @llvm.nacl.atomic.fence(memory_order)
/// and
/// call void asm sideeffect "", "~{memory}"()
/// fence seq_cst
/// call void asm sideeffect "", "~{memory}"()
/// becomes:
/// call void asm sideeffect "", "~{memory}"()
/// call void @llvm.nacl.atomic.fence.all()
/// call void asm sideeffect "", "~{memory}"()
/// Note that the assembly gets eliminated by the -remove-asm-memory pass.
void AtomicVisitor::visitFenceInst(FenceInst &I) {
Type *T = Type::getInt32Ty(C); // Fences aren't overloaded on type.
BasicBlock::InstListType &IL(I.getParent()->getInstList());
bool isFirst = IL.empty() || &*I.getParent()->getInstList().begin() == &I;
bool isLast = IL.empty() || &*I.getParent()->getInstList().rbegin() == &I;
CallInst *PrevC = isFirst ? 0 : dyn_cast<CallInst>(I.getPrevNode());
CallInst *NextC = isLast ? 0 : dyn_cast<CallInst>(I.getNextNode());
if ((PrevC && PrevC->isInlineAsm() &&
cast<InlineAsm>(PrevC->getCalledValue())->isAsmMemory()) &&
(NextC && NextC->isInlineAsm() &&
cast<InlineAsm>(NextC->getCalledValue())->isAsmMemory()) &&
I.getOrdering() == SequentiallyConsistent) {
replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_fence_all, T,
T, ArrayRef<Value *>());
} else {
Value *Args[] = { freezeMemoryOrder(I) };
replaceInstructionWithIntrinsicCall(I, Intrinsic::nacl_atomic_fence, T, T,
Args);
}
}
ModulePass *llvm::createRewriteAtomicsPass() { return new RewriteAtomics(); }