//===--- CGVtable.cpp - Emit LLVM Code for C++ vtables --------------------===// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// // // This contains code dealing with C++ code generation of virtual tables. // //===----------------------------------------------------------------------===// #include "CodeGenModule.h" #include "CodeGenFunction.h" #include "clang/AST/RecordLayout.h" #include using namespace clang; using namespace CodeGen; class VtableBuilder { public: /// Index_t - Vtable index type. typedef uint64_t Index_t; private: std::vector &methods; std::vector submethods; llvm::Type *Ptr8Ty; /// Class - The most derived class that this vtable is being built for. const CXXRecordDecl *Class; /// LayoutClass - The most derived class used for virtual base layout /// information. const CXXRecordDecl *LayoutClass; /// LayoutOffset - The offset for Class in LayoutClass. uint64_t LayoutOffset; /// BLayout - Layout for the most derived class that this vtable is being /// built for. const ASTRecordLayout &BLayout; llvm::SmallSet IndirectPrimary; llvm::SmallSet SeenVBase; llvm::Constant *rtti; llvm::LLVMContext &VMContext; CodeGenModule &CGM; // Per-module state. /// Index - Maps a method decl into a vtable index. Useful for virtual /// dispatch codegen. llvm::DenseMap Index; llvm::DenseMap VCall; llvm::DenseMap VCallOffset; // This is the offset to the nearest virtual base llvm::DenseMap NonVirtualOffset; llvm::DenseMap VBIndex; typedef llvm::DenseMap Pures_t; Pures_t Pures; typedef std::pair CallOffset; typedef llvm::DenseMap Thunks_t; Thunks_t Thunks; typedef llvm::DenseMap, CanQualType> > CovariantThunks_t; CovariantThunks_t CovariantThunks; std::vector VCalls; typedef std::pair CtorVtable_t; // CtorVtable - Used to hold the AddressPoints (offsets) into the built vtable // for use in computing the initializers for the VTT. llvm::DenseMap &AddressPoints; typedef CXXRecordDecl::method_iterator method_iter; const bool Extern; const uint32_t LLVMPointerWidth; Index_t extra; typedef std::vector > Path_t; llvm::Constant *cxa_pure; public: VtableBuilder(std::vector &meth, const CXXRecordDecl *c, const CXXRecordDecl *l, uint64_t lo, CodeGenModule &cgm) : methods(meth), Class(c), LayoutClass(l), LayoutOffset(lo), BLayout(cgm.getContext().getASTRecordLayout(l)), rtti(cgm.GenerateRttiRef(c)), VMContext(cgm.getModule().getContext()), CGM(cgm), AddressPoints(*new llvm::DenseMap), Extern(!l->isInAnonymousNamespace()), LLVMPointerWidth(cgm.getContext().Target.getPointerWidth(0)) { Ptr8Ty = llvm::PointerType::get(llvm::Type::getInt8Ty(VMContext), 0); // Calculate pointer for ___cxa_pure_virtual. const llvm::FunctionType *FTy; std::vector ArgTys; const llvm::Type *ResultType = llvm::Type::getVoidTy(VMContext); FTy = llvm::FunctionType::get(ResultType, ArgTys, false); cxa_pure = wrap(CGM.CreateRuntimeFunction(FTy, "__cxa_pure_virtual")); } llvm::DenseMap &getIndex() { return Index; } llvm::DenseMap &getVBIndex() { return VBIndex; } llvm::DenseMap *getAddressPoints() { return &AddressPoints; } llvm::Constant *wrap(Index_t i) { llvm::Constant *m; m = llvm::ConstantInt::get(llvm::Type::getInt64Ty(VMContext), i); return llvm::ConstantExpr::getIntToPtr(m, Ptr8Ty); } llvm::Constant *wrap(llvm::Constant *m) { return llvm::ConstantExpr::getBitCast(m, Ptr8Ty); } //#define D1(x) #define D1(X) do { if (getenv("DEBUG")) { X; } } while (0) void GenerateVBaseOffsets(const CXXRecordDecl *RD, uint64_t Offset, bool updateVBIndex, Index_t current_vbindex) { for (CXXRecordDecl::base_class_const_iterator i = RD->bases_begin(), e = RD->bases_end(); i != e; ++i) { const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); Index_t next_vbindex = current_vbindex; if (i->isVirtual() && !SeenVBase.count(Base)) { SeenVBase.insert(Base); if (updateVBIndex) { next_vbindex = (ssize_t)(-(VCalls.size()*LLVMPointerWidth/8) - 3*LLVMPointerWidth/8); VBIndex[Base] = next_vbindex; } int64_t BaseOffset = -(Offset/8) + BLayout.getVBaseClassOffset(Base)/8; VCalls.push_back((0?700:0) + BaseOffset); D1(printf(" vbase for %s at %d delta %d most derived %s\n", Base->getNameAsCString(), (int)-VCalls.size()-3, (int)BaseOffset, Class->getNameAsCString())); } // We also record offsets for non-virtual bases to closest enclosing // virtual base. We do this so that we don't have to search // for the nearst virtual base class when generating thunks. if (updateVBIndex && VBIndex.count(Base) == 0) VBIndex[Base] = next_vbindex; GenerateVBaseOffsets(Base, Offset, updateVBIndex, next_vbindex); } } void StartNewTable() { SeenVBase.clear(); } Index_t VBlookup(CXXRecordDecl *D, CXXRecordDecl *B); Index_t getNVOffset_1(const CXXRecordDecl *D, const CXXRecordDecl *B, Index_t Offset = 0) { if (B == D) return Offset; const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(D); for (CXXRecordDecl::base_class_const_iterator i = D->bases_begin(), e = D->bases_end(); i != e; ++i) { const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); int64_t BaseOffset = 0; if (!i->isVirtual()) BaseOffset = Offset + Layout.getBaseClassOffset(Base); int64_t o = getNVOffset_1(Base, B, BaseOffset); if (o >= 0) return o; } return -1; } /// getNVOffset - Returns the non-virtual offset for the given (B) base of the /// derived class D. Index_t getNVOffset(QualType qB, QualType qD) { qD = qD->getPointeeType(); qB = qB->getPointeeType(); CXXRecordDecl *D = cast(qD->getAs()->getDecl()); CXXRecordDecl *B = cast(qB->getAs()->getDecl()); int64_t o = getNVOffset_1(D, B); if (o >= 0) return o; assert(false && "FIXME: non-virtual base not found"); return 0; } /// getVbaseOffset - Returns the index into the vtable for the virtual base /// offset for the given (B) virtual base of the derived class D. Index_t getVbaseOffset(QualType qB, QualType qD) { qD = qD->getPointeeType(); qB = qB->getPointeeType(); CXXRecordDecl *D = cast(qD->getAs()->getDecl()); CXXRecordDecl *B = cast(qB->getAs()->getDecl()); if (D != Class) return VBlookup(D, B); llvm::DenseMap::iterator i; i = VBIndex.find(B); if (i != VBIndex.end()) return i->second; assert(false && "FIXME: Base not found"); return 0; } bool OverrideMethod(GlobalDecl GD, llvm::Constant *m, bool MorallyVirtual, Index_t OverrideOffset, Index_t Offset, int64_t CurrentVBaseOffset) { const CXXMethodDecl *MD = cast(GD.getDecl()); const bool isPure = MD->isPure(); typedef CXXMethodDecl::method_iterator meth_iter; // FIXME: Should OverrideOffset's be Offset? // FIXME: Don't like the nested loops. For very large inheritance // heirarchies we could have a table on the side with the final overridder // and just replace each instance of an overridden method once. Would be // nice to measure the cost/benefit on real code. for (meth_iter mi = MD->begin_overridden_methods(), e = MD->end_overridden_methods(); mi != e; ++mi) { GlobalDecl OGD; const CXXMethodDecl *OMD = *mi; if (const CXXDestructorDecl *DD = dyn_cast(OMD)) OGD = GlobalDecl(DD, GD.getDtorType()); else OGD = OMD; llvm::Constant *om; om = WrapAddrOf(OGD); om = llvm::ConstantExpr::getBitCast(om, Ptr8Ty); for (Index_t i = 0, e = submethods.size(); i != e; ++i) { // FIXME: begin_overridden_methods might be too lax, covariance */ if (submethods[i] != om) continue; QualType nc_oret = OMD->getType()->getAs()->getResultType(); CanQualType oret = CGM.getContext().getCanonicalType(nc_oret); QualType nc_ret = MD->getType()->getAs()->getResultType(); CanQualType ret = CGM.getContext().getCanonicalType(nc_ret); CallOffset ReturnOffset = std::make_pair(0, 0); if (oret != ret) { // FIXME: calculate offsets for covariance if (CovariantThunks.count(OMD)) { oret = CovariantThunks[OMD].second; CovariantThunks.erase(OMD); } // FIXME: Double check oret Index_t nv = getNVOffset(oret, ret)/8; ReturnOffset = std::make_pair(nv, getVbaseOffset(oret, ret)); } Index[GD] = i; submethods[i] = m; if (isPure) Pures[GD] = 1; Pures.erase(OGD); Thunks.erase(OGD); if (MorallyVirtual || VCall.count(OGD)) { Index_t &idx = VCall[OGD]; if (idx == 0) { NonVirtualOffset[GD] = -OverrideOffset/8 + CurrentVBaseOffset/8; VCallOffset[GD] = OverrideOffset/8; idx = VCalls.size()+1; VCalls.push_back(0); D1(printf(" vcall for %s at %d with delta %d most derived %s\n", MD->getNameAsString().c_str(), (int)-idx-3, (int)VCalls[idx-1], Class->getNameAsCString())); } else { NonVirtualOffset[GD] = NonVirtualOffset[OGD]; VCallOffset[GD] = VCallOffset[OGD]; VCalls[idx-1] = -VCallOffset[OGD] + OverrideOffset/8; D1(printf(" vcall patch for %s at %d with delta %d most derived %s\n", MD->getNameAsString().c_str(), (int)-idx-3, (int)VCalls[idx-1], Class->getNameAsCString())); } VCall[GD] = idx; int64_t O = NonVirtualOffset[GD]; int v = -((idx+extra+2)*LLVMPointerWidth/8); // Optimize out virtual adjustments of 0. if (VCalls[idx-1] == 0) v = 0; CallOffset ThisOffset = std::make_pair(O, v); // FIXME: Do we always have to build a covariant thunk to save oret, // which is the containing virtual base class? if (ReturnOffset.first || ReturnOffset.second) CovariantThunks[GD] = std::make_pair(std::make_pair(ThisOffset, ReturnOffset), oret); else if (!isPure && (ThisOffset.first || ThisOffset.second)) Thunks[GD] = ThisOffset; return true; } // FIXME: finish off int64_t O = VCallOffset[OGD] - OverrideOffset/8; if (O || ReturnOffset.first || ReturnOffset.second) { CallOffset ThisOffset = std::make_pair(O, 0); if (ReturnOffset.first || ReturnOffset.second) CovariantThunks[GD] = std::make_pair(std::make_pair(ThisOffset, ReturnOffset), oret); else if (!isPure) Thunks[GD] = ThisOffset; } return true; } } return false; } void InstallThunks() { for (Thunks_t::iterator i = Thunks.begin(), e = Thunks.end(); i != e; ++i) { GlobalDecl GD = i->first; const CXXMethodDecl *MD = cast(GD.getDecl()); assert(!MD->isPure() && "Trying to thunk a pure"); Index_t idx = Index[GD]; Index_t nv_O = i->second.first; Index_t v_O = i->second.second; submethods[idx] = CGM.BuildThunk(MD, Extern, nv_O, v_O); } Thunks.clear(); for (CovariantThunks_t::iterator i = CovariantThunks.begin(), e = CovariantThunks.end(); i != e; ++i) { GlobalDecl GD = i->first; const CXXMethodDecl *MD = cast(GD.getDecl()); if (MD->isPure()) continue; Index_t idx = Index[GD]; Index_t nv_t = i->second.first.first.first; Index_t v_t = i->second.first.first.second; Index_t nv_r = i->second.first.second.first; Index_t v_r = i->second.first.second.second; submethods[idx] = CGM.BuildCovariantThunk(MD, Extern, nv_t, v_t, nv_r, v_r); } CovariantThunks.clear(); for (Pures_t::iterator i = Pures.begin(), e = Pures.end(); i != e; ++i) { GlobalDecl GD = i->first; Index_t idx = Index[GD]; submethods[idx] = cxa_pure; } Pures.clear(); } llvm::Constant *WrapAddrOf(GlobalDecl GD) { const CXXMethodDecl *MD = cast(GD.getDecl()); if (const CXXDestructorDecl *Dtor = dyn_cast(MD)) return wrap(CGM.GetAddrOfCXXDestructor(Dtor, GD.getDtorType())); const FunctionProtoType *FPT = MD->getType()->getAs(); const llvm::Type *Ty = CGM.getTypes().GetFunctionType(CGM.getTypes().getFunctionInfo(MD), FPT->isVariadic()); return wrap(CGM.GetAddrOfFunction(MD, Ty)); } void OverrideMethods(Path_t *Path, bool MorallyVirtual, int64_t Offset, int64_t CurrentVBaseOffset) { for (Path_t::reverse_iterator i = Path->rbegin(), e = Path->rend(); i != e; ++i) { const CXXRecordDecl *RD = i->first; int64_t OverrideOffset = i->second; for (method_iter mi = RD->method_begin(), me = RD->method_end(); mi != me; ++mi) { const CXXMethodDecl *MD = *mi; if (!MD->isVirtual()) continue; if (const CXXDestructorDecl *DD = dyn_cast(MD)) { // Override both the complete and the deleting destructor. GlobalDecl CompDtor(DD, Dtor_Complete); OverrideMethod(CompDtor, WrapAddrOf(CompDtor), MorallyVirtual, OverrideOffset, Offset, CurrentVBaseOffset); GlobalDecl DeletingDtor(DD, Dtor_Deleting); OverrideMethod(DeletingDtor, WrapAddrOf(DeletingDtor), MorallyVirtual, OverrideOffset, Offset, CurrentVBaseOffset); } else { OverrideMethod(MD, WrapAddrOf(MD), MorallyVirtual, OverrideOffset, Offset, CurrentVBaseOffset); } } } } void AddMethod(const GlobalDecl GD, bool MorallyVirtual, Index_t Offset, bool ForVirtualBase, int64_t CurrentVBaseOffset) { llvm::Constant *m = WrapAddrOf(GD); // If we can find a previously allocated slot for this, reuse it. if (OverrideMethod(GD, m, MorallyVirtual, Offset, Offset, CurrentVBaseOffset)) return; const CXXMethodDecl *MD = cast(GD.getDecl()); // else allocate a new slot. Index[GD] = submethods.size(); submethods.push_back(m); D1(printf(" vfn for %s at %d\n", MD->getNameAsString().c_str(), (int)Index[GD])); if (MD->isPure()) Pures[GD] = 1; if (MorallyVirtual) { VCallOffset[GD] = Offset/8; Index_t &idx = VCall[GD]; // Allocate the first one, after that, we reuse the previous one. if (idx == 0) { NonVirtualOffset[GD] = CurrentVBaseOffset/8 - Offset/8; idx = VCalls.size()+1; VCalls.push_back(0); D1(printf(" vcall for %s at %d with delta %d\n", MD->getNameAsString().c_str(), (int)-VCalls.size()-3, 0)); } } } void AddMethods(const CXXRecordDecl *RD, bool MorallyVirtual, Index_t Offset, bool RDisVirtualBase, int64_t CurrentVBaseOffset) { for (method_iter mi = RD->method_begin(), me = RD->method_end(); mi != me; ++mi) { const CXXMethodDecl *MD = *mi; if (!MD->isVirtual()) continue; if (const CXXDestructorDecl *DD = dyn_cast(MD)) { // For destructors, add both the complete and the deleting destructor // to the vtable. AddMethod(GlobalDecl(DD, Dtor_Complete), MorallyVirtual, Offset, RDisVirtualBase, CurrentVBaseOffset); AddMethod(GlobalDecl(DD, Dtor_Deleting), MorallyVirtual, Offset, RDisVirtualBase, CurrentVBaseOffset); } else AddMethod(MD, MorallyVirtual, Offset, RDisVirtualBase, CurrentVBaseOffset); } } void NonVirtualBases(const CXXRecordDecl *RD, const ASTRecordLayout &Layout, const CXXRecordDecl *PrimaryBase, bool PrimaryBaseWasVirtual, bool MorallyVirtual, int64_t Offset, int64_t CurrentVBaseOffset, Path_t *Path) { Path->push_back(std::make_pair(RD, Offset)); for (CXXRecordDecl::base_class_const_iterator i = RD->bases_begin(), e = RD->bases_end(); i != e; ++i) { if (i->isVirtual()) continue; const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); if (Base != PrimaryBase || PrimaryBaseWasVirtual) { uint64_t o = Offset + Layout.getBaseClassOffset(Base); StartNewTable(); GenerateVtableForBase(Base, o, MorallyVirtual, false, CurrentVBaseOffset, Path); } } Path->pop_back(); } // #define D(X) do { X; } while (0) #define D(X) void insertVCalls(int InsertionPoint) { llvm::Constant *e = 0; D1(printf("============= combining vbase/vcall\n")); D(VCalls.insert(VCalls.begin(), 673)); D(VCalls.push_back(672)); methods.insert(methods.begin() + InsertionPoint, VCalls.size(), e); // The vcalls come first... for (std::vector::reverse_iterator i = VCalls.rbegin(), e = VCalls.rend(); i != e; ++i) methods[InsertionPoint++] = wrap((0?600:0) + *i); VCalls.clear(); VCall.clear(); } void AddAddressPoints(const CXXRecordDecl *RD, uint64_t Offset, Index_t AddressPoint) { D1(printf("XXX address point for %s in %s layout %s at offset %d is %d\n", RD->getNameAsCString(), Class->getNameAsCString(), LayoutClass->getNameAsCString(), (int)Offset, (int)AddressPoint)); AddressPoints[std::make_pair(RD, Offset)] = AddressPoint; // Now also add the address point for all our primary bases. while (1) { const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); RD = Layout.getPrimaryBase(); const bool PrimaryBaseWasVirtual = Layout.getPrimaryBaseWasVirtual(); // FIXME: Double check this. if (RD == 0) break; if (PrimaryBaseWasVirtual && BLayout.getVBaseClassOffset(RD) != Offset) break; D1(printf("XXX address point for %s in %s layout %s at offset %d is %d\n", RD->getNameAsCString(), Class->getNameAsCString(), LayoutClass->getNameAsCString(), (int)Offset, (int)AddressPoint)); AddressPoints[std::make_pair(RD, Offset)] = AddressPoint; } } Index_t end(const CXXRecordDecl *RD, const ASTRecordLayout &Layout, const CXXRecordDecl *PrimaryBase, bool PrimaryBaseWasVirtual, bool MorallyVirtual, int64_t Offset, bool ForVirtualBase, int64_t CurrentVBaseOffset, Path_t *Path) { bool alloc = false; if (Path == 0) { alloc = true; Path = new Path_t; } StartNewTable(); extra = 0; bool DeferVCalls = MorallyVirtual || ForVirtualBase; int VCallInsertionPoint = methods.size(); if (!DeferVCalls) { insertVCalls(VCallInsertionPoint); } else // FIXME: just for extra, or for all uses of VCalls.size post this? extra = -VCalls.size(); methods.push_back(wrap(-((Offset-LayoutOffset)/8))); methods.push_back(rtti); Index_t AddressPoint = methods.size(); InstallThunks(); D1(printf("============= combining methods\n")); methods.insert(methods.end(), submethods.begin(), submethods.end()); submethods.clear(); // and then the non-virtual bases. NonVirtualBases(RD, Layout, PrimaryBase, PrimaryBaseWasVirtual, MorallyVirtual, Offset, CurrentVBaseOffset, Path); if (ForVirtualBase) { // FIXME: We're adding to VCalls in callers, we need to do the overrides // in the inner part, so that we know the complete set of vcalls during // the build and don't have to insert into methods. Saving out the // AddressPoint here, would need to be fixed, if we didn't do that. Also // retroactively adding vcalls for overrides later wind up in the wrong // place, the vcall slot has to be alloted during the walk of the base // when the function is first introduces. AddressPoint += VCalls.size(); insertVCalls(VCallInsertionPoint); } AddAddressPoints(RD, Offset, AddressPoint); if (alloc) { delete Path; } return AddressPoint; } void Primaries(const CXXRecordDecl *RD, bool MorallyVirtual, int64_t Offset, bool updateVBIndex, Index_t current_vbindex, bool RDisVirtualBase, int64_t CurrentVBaseOffset) { if (!RD->isDynamicClass()) return; const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); const CXXRecordDecl *PrimaryBase = Layout.getPrimaryBase(); const bool PrimaryBaseWasVirtual = Layout.getPrimaryBaseWasVirtual(); // vtables are composed from the chain of primaries. if (PrimaryBase) { D1(printf(" doing primaries for %s most derived %s\n", RD->getNameAsCString(), Class->getNameAsCString())); int BaseCurrentVBaseOffset = CurrentVBaseOffset; if (PrimaryBaseWasVirtual) BaseCurrentVBaseOffset = BLayout.getVBaseClassOffset(PrimaryBase); if (!PrimaryBaseWasVirtual) Primaries(PrimaryBase, PrimaryBaseWasVirtual|MorallyVirtual, Offset, updateVBIndex, current_vbindex, PrimaryBaseWasVirtual, BaseCurrentVBaseOffset); } D1(printf(" doing vcall entries for %s most derived %s\n", RD->getNameAsCString(), Class->getNameAsCString())); // And add the virtuals for the class to the primary vtable. AddMethods(RD, MorallyVirtual, Offset, RDisVirtualBase, CurrentVBaseOffset); } void VBPrimaries(const CXXRecordDecl *RD, bool MorallyVirtual, int64_t Offset, bool updateVBIndex, Index_t current_vbindex, bool RDisVirtualBase, int64_t CurrentVBaseOffset, bool bottom=false) { if (!RD->isDynamicClass()) return; const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); const CXXRecordDecl *PrimaryBase = Layout.getPrimaryBase(); const bool PrimaryBaseWasVirtual = Layout.getPrimaryBaseWasVirtual(); // vtables are composed from the chain of primaries. if (PrimaryBase) { int BaseCurrentVBaseOffset = CurrentVBaseOffset; if (PrimaryBaseWasVirtual) { IndirectPrimary.insert(PrimaryBase); BaseCurrentVBaseOffset = BLayout.getVBaseClassOffset(PrimaryBase); } D1(printf(" doing primaries for %s most derived %s\n", RD->getNameAsCString(), Class->getNameAsCString())); VBPrimaries(PrimaryBase, PrimaryBaseWasVirtual|MorallyVirtual, Offset, updateVBIndex, current_vbindex, PrimaryBaseWasVirtual, BaseCurrentVBaseOffset); } D1(printf(" doing vbase entries for %s most derived %s\n", RD->getNameAsCString(), Class->getNameAsCString())); GenerateVBaseOffsets(RD, Offset, updateVBIndex, current_vbindex); if (RDisVirtualBase || bottom) { Primaries(RD, MorallyVirtual, Offset, updateVBIndex, current_vbindex, RDisVirtualBase, CurrentVBaseOffset); } } int64_t GenerateVtableForBase(const CXXRecordDecl *RD, int64_t Offset = 0, bool MorallyVirtual = false, bool ForVirtualBase = false, int CurrentVBaseOffset = 0, Path_t *Path = 0) { if (!RD->isDynamicClass()) return 0; // Construction vtable don't need parts that have no virtual bases and // aren't morally virtual. if ((LayoutClass != Class) && RD->getNumVBases() == 0 && !MorallyVirtual) return 0; const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); const CXXRecordDecl *PrimaryBase = Layout.getPrimaryBase(); const bool PrimaryBaseWasVirtual = Layout.getPrimaryBaseWasVirtual(); extra = 0; D1(printf("building entries for base %s most derived %s\n", RD->getNameAsCString(), Class->getNameAsCString())); if (ForVirtualBase) extra = VCalls.size(); VBPrimaries(RD, MorallyVirtual, Offset, !ForVirtualBase, 0, ForVirtualBase, CurrentVBaseOffset, true); if (Path) OverrideMethods(Path, MorallyVirtual, Offset, CurrentVBaseOffset); return end(RD, Layout, PrimaryBase, PrimaryBaseWasVirtual, MorallyVirtual, Offset, ForVirtualBase, CurrentVBaseOffset, Path); } void GenerateVtableForVBases(const CXXRecordDecl *RD, int64_t Offset = 0, Path_t *Path = 0) { bool alloc = false; if (Path == 0) { alloc = true; Path = new Path_t; } // FIXME: We also need to override using all paths to a virtual base, // right now, we just process the first path Path->push_back(std::make_pair(RD, Offset)); for (CXXRecordDecl::base_class_const_iterator i = RD->bases_begin(), e = RD->bases_end(); i != e; ++i) { const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); if (i->isVirtual() && !IndirectPrimary.count(Base)) { // Mark it so we don't output it twice. IndirectPrimary.insert(Base); StartNewTable(); VCall.clear(); int64_t BaseOffset = BLayout.getVBaseClassOffset(Base); int64_t CurrentVBaseOffset = BaseOffset; D1(printf("vtable %s virtual base %s\n", Class->getNameAsCString(), Base->getNameAsCString())); GenerateVtableForBase(Base, BaseOffset, true, true, CurrentVBaseOffset, Path); } int64_t BaseOffset; if (i->isVirtual()) BaseOffset = BLayout.getVBaseClassOffset(Base); else { const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); BaseOffset = Offset + Layout.getBaseClassOffset(Base); } if (Base->getNumVBases()) { GenerateVtableForVBases(Base, BaseOffset, Path); } } Path->pop_back(); if (alloc) delete Path; } }; VtableBuilder::Index_t VtableBuilder::VBlookup(CXXRecordDecl *D, CXXRecordDecl *B) { return CGM.getVtableInfo().getVirtualBaseOffsetIndex(D, B); } int64_t CGVtableInfo::getMethodVtableIndex(GlobalDecl GD) { MethodVtableIndicesTy::iterator I = MethodVtableIndices.find(GD); if (I != MethodVtableIndices.end()) return I->second; const CXXRecordDecl *RD = cast(GD.getDecl())->getParent(); std::vector methods; // FIXME: This seems expensive. Can we do a partial job to get // just this data. VtableBuilder b(methods, RD, RD, 0, CGM); D1(printf("vtable %s\n", RD->getNameAsCString())); b.GenerateVtableForBase(RD); b.GenerateVtableForVBases(RD); MethodVtableIndices.insert(b.getIndex().begin(), b.getIndex().end()); I = MethodVtableIndices.find(GD); assert(I != MethodVtableIndices.end() && "Did not find index!"); return I->second; } int64_t CGVtableInfo::getVirtualBaseOffsetIndex(const CXXRecordDecl *RD, const CXXRecordDecl *VBase) { ClassPairTy ClassPair(RD, VBase); VirtualBaseClassIndiciesTy::iterator I = VirtualBaseClassIndicies.find(ClassPair); if (I != VirtualBaseClassIndicies.end()) return I->second; std::vector methods; // FIXME: This seems expensive. Can we do a partial job to get // just this data. VtableBuilder b(methods, RD, RD, 0, CGM); D1(printf("vtable %s\n", RD->getNameAsCString())); b.GenerateVtableForBase(RD); b.GenerateVtableForVBases(RD); for (llvm::DenseMap::iterator I = b.getVBIndex().begin(), E = b.getVBIndex().end(); I != E; ++I) { // Insert all types. ClassPairTy ClassPair(RD, I->first); VirtualBaseClassIndicies.insert(std::make_pair(ClassPair, I->second)); } I = VirtualBaseClassIndicies.find(ClassPair); assert(I != VirtualBaseClassIndicies.end() && "Did not find index!"); return I->second; } llvm::Constant *CodeGenModule::GenerateVtable(const CXXRecordDecl *LayoutClass, const CXXRecordDecl *RD, uint64_t Offset) { llvm::SmallString<256> OutName; llvm::raw_svector_ostream Out(OutName); if (LayoutClass != RD) mangleCXXCtorVtable(getMangleContext(), LayoutClass, Offset/8, RD, Out); else mangleCXXVtable(getMangleContext(), RD, Out); llvm::StringRef Name = Out.str(); std::vector methods; llvm::Type *Ptr8Ty=llvm::PointerType::get(llvm::Type::getInt8Ty(VMContext),0); int64_t AddressPoint; llvm::GlobalVariable *GV = getModule().getGlobalVariable(Name); if (GV && AddressPoints[LayoutClass] && !GV->isDeclaration()) AddressPoint=(*(*(AddressPoints[LayoutClass]))[RD])[std::make_pair(RD, Offset)]; else { VtableBuilder b(methods, RD, LayoutClass, Offset, *this); D1(printf("vtable %s\n", RD->getNameAsCString())); // First comes the vtables for all the non-virtual bases... AddressPoint = b.GenerateVtableForBase(RD, Offset); // then the vtables for all the virtual bases. b.GenerateVtableForVBases(RD, Offset); CodeGenModule::AddrMap_t *&ref = AddressPoints[LayoutClass]; if (ref == 0) ref = new CodeGenModule::AddrMap_t; (*ref)[RD] = b.getAddressPoints(); bool CreateDefinition = true; if (LayoutClass != RD) CreateDefinition = true; else { // We have to convert it to have a record layout. Types.ConvertTagDeclType(LayoutClass); const CGRecordLayout &CGLayout = Types.getCGRecordLayout(LayoutClass); if (const CXXMethodDecl *KeyFunction = CGLayout.getKeyFunction()) { if (!KeyFunction->getBody()) { // If there is a KeyFunction, and it isn't defined, just build a // reference to the vtable. CreateDefinition = false; } } } llvm::Constant *C = 0; llvm::Type *type = Ptr8Ty; llvm::GlobalVariable::LinkageTypes linktype = llvm::GlobalValue::ExternalLinkage; if (CreateDefinition) { llvm::ArrayType *ntype = llvm::ArrayType::get(Ptr8Ty, methods.size()); C = llvm::ConstantArray::get(ntype, methods); linktype = llvm::GlobalValue::LinkOnceODRLinkage; if (LayoutClass->isInAnonymousNamespace()) linktype = llvm::GlobalValue::InternalLinkage; type = ntype; } llvm::GlobalVariable *OGV = GV; GV = new llvm::GlobalVariable(getModule(), type, true, linktype, C, Name); if (OGV) { GV->takeName(OGV); llvm::Constant *NewPtr = llvm::ConstantExpr::getBitCast(GV, OGV->getType()); OGV->replaceAllUsesWith(NewPtr); OGV->eraseFromParent(); } bool Hidden = getDeclVisibilityMode(RD) == LangOptions::Hidden; if (Hidden) GV->setVisibility(llvm::GlobalVariable::HiddenVisibility); } llvm::Constant *vtable = llvm::ConstantExpr::getBitCast(GV, Ptr8Ty); llvm::Constant *AddressPointC; uint32_t LLVMPointerWidth = getContext().Target.getPointerWidth(0); AddressPointC = llvm::ConstantInt::get(llvm::Type::getInt64Ty(VMContext), AddressPoint*LLVMPointerWidth/8); vtable = llvm::ConstantExpr::getInBoundsGetElementPtr(vtable, &AddressPointC, 1); return vtable; } class VTTBuilder { /// Inits - The list of values built for the VTT. std::vector &Inits; /// Class - The most derived class that this vtable is being built for. const CXXRecordDecl *Class; CodeGenModule &CGM; // Per-module state. llvm::SmallSet SeenVBase; /// BLayout - Layout for the most derived class that this vtable is being /// built for. const ASTRecordLayout &BLayout; CodeGenModule::AddrMap_t &AddressPoints; // vtbl - A pointer to the vtable for Class. llvm::Constant *ClassVtbl; llvm::LLVMContext &VMContext; /// BuildVtablePtr - Build up a referene to the given secondary vtable llvm::Constant *BuildVtablePtr(llvm::Constant *vtbl, const CXXRecordDecl *VtblClass, const CXXRecordDecl *RD, uint64_t Offset) { int64_t AddressPoint; AddressPoint = (*AddressPoints[VtblClass])[std::make_pair(RD, Offset)]; // FIXME: We can never have 0 address point. Do this for now so gepping // retains the same structure. if (AddressPoint == 0) AddressPoint = 1; D1(printf("XXX address point for %s in %s layout %s at offset %d was %d\n", RD->getNameAsCString(), VtblClass->getNameAsCString(), Class->getNameAsCString(), (int)Offset, (int)AddressPoint)); uint32_t LLVMPointerWidth = CGM.getContext().Target.getPointerWidth(0); llvm::Constant *init; init = llvm::ConstantInt::get(llvm::Type::getInt64Ty(VMContext), AddressPoint*LLVMPointerWidth/8); init = llvm::ConstantExpr::getInBoundsGetElementPtr(vtbl, &init, 1); return init; } /// Secondary - Add the secondary vtable pointers to Inits. Offset is the /// current offset in bits to the object we're working on. void Secondary(const CXXRecordDecl *RD, llvm::Constant *vtbl, const CXXRecordDecl *VtblClass, uint64_t Offset=0, bool MorallyVirtual=false) { if (RD->getNumVBases() == 0 && ! MorallyVirtual) return; for (CXXRecordDecl::base_class_const_iterator i = RD->bases_begin(), e = RD->bases_end(); i != e; ++i) { const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); const CXXRecordDecl *PrimaryBase = Layout.getPrimaryBase(); const bool PrimaryBaseWasVirtual = Layout.getPrimaryBaseWasVirtual(); bool NonVirtualPrimaryBase; NonVirtualPrimaryBase = !PrimaryBaseWasVirtual && Base == PrimaryBase; bool BaseMorallyVirtual = MorallyVirtual | i->isVirtual(); uint64_t BaseOffset; if (!i->isVirtual()) { const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); BaseOffset = Offset + Layout.getBaseClassOffset(Base); } else BaseOffset = BLayout.getVBaseClassOffset(Base); llvm::Constant *subvtbl = vtbl; const CXXRecordDecl *subVtblClass = VtblClass; if ((Base->getNumVBases() || BaseMorallyVirtual) && !NonVirtualPrimaryBase) { // FIXME: Slightly too many of these for __ZTT8test8_B2 llvm::Constant *init; if (BaseMorallyVirtual) init = BuildVtablePtr(vtbl, VtblClass, RD, Offset); else { init = CGM.getVtableInfo().getCtorVtable(Class, Base, BaseOffset); subvtbl = dyn_cast(init->getOperand(0)); subVtblClass = Base; } Inits.push_back(init); } Secondary(Base, subvtbl, subVtblClass, BaseOffset, BaseMorallyVirtual); } } /// BuiltVTT - Add the VTT to Inits. Offset is the offset in bits to the /// currnet object we're working on. void BuildVTT(const CXXRecordDecl *RD, uint64_t Offset, bool MorallyVirtual) { if (RD->getNumVBases() == 0 && !MorallyVirtual) return; llvm::Constant *init; const CXXRecordDecl *VtblClass; // First comes the primary virtual table pointer... if (MorallyVirtual) { init = BuildVtablePtr(ClassVtbl, Class, RD, Offset); VtblClass = Class; } else { init = CGM.getVtableInfo().getCtorVtable(Class, RD, Offset); VtblClass = RD; } llvm::Constant *vtbl = dyn_cast(init->getOperand(0)); Inits.push_back(init); // then the secondary VTTs.... SecondaryVTTs(RD, Offset, MorallyVirtual); // and last the secondary vtable pointers. Secondary(RD, vtbl, VtblClass, Offset, MorallyVirtual); } /// SecondaryVTTs - Add the secondary VTTs to Inits. The secondary VTTs are /// built from each direct non-virtual proper base that requires a VTT in /// declaration order. void SecondaryVTTs(const CXXRecordDecl *RD, uint64_t Offset=0, bool MorallyVirtual=false) { for (CXXRecordDecl::base_class_const_iterator i = RD->bases_begin(), e = RD->bases_end(); i != e; ++i) { const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); if (i->isVirtual()) continue; const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD); uint64_t BaseOffset = Offset + Layout.getBaseClassOffset(Base); BuildVTT(Base, BaseOffset, MorallyVirtual); } } /// VirtualVTTs - Add the VTT for each proper virtual base in inheritance /// graph preorder. void VirtualVTTs(const CXXRecordDecl *RD) { for (CXXRecordDecl::base_class_const_iterator i = RD->bases_begin(), e = RD->bases_end(); i != e; ++i) { const CXXRecordDecl *Base = cast(i->getType()->getAs()->getDecl()); if (i->isVirtual() && !SeenVBase.count(Base)) { SeenVBase.insert(Base); uint64_t BaseOffset = BLayout.getVBaseClassOffset(Base); BuildVTT(Base, BaseOffset, true); } VirtualVTTs(Base); } } public: VTTBuilder(std::vector &inits, const CXXRecordDecl *c, CodeGenModule &cgm) : Inits(inits), Class(c), CGM(cgm), BLayout(cgm.getContext().getASTRecordLayout(c)), AddressPoints(*cgm.AddressPoints[c]), VMContext(cgm.getModule().getContext()) { // First comes the primary virtual table pointer for the complete class... ClassVtbl = CGM.getVtableInfo().getVtable(Class); Inits.push_back(ClassVtbl); ClassVtbl = dyn_cast(ClassVtbl->getOperand(0)); // then the secondary VTTs... SecondaryVTTs(Class); // then the secondary vtable pointers... Secondary(Class, ClassVtbl, Class); // and last, the virtual VTTs. VirtualVTTs(Class); } }; llvm::Constant *CodeGenModule::GenerateVTT(const CXXRecordDecl *RD) { // Only classes that have virtual bases need a VTT. if (RD->getNumVBases() == 0) return 0; llvm::SmallString<256> OutName; llvm::raw_svector_ostream Out(OutName); mangleCXXVTT(getMangleContext(), RD, Out); llvm::StringRef Name = Out.str(); llvm::GlobalVariable::LinkageTypes linktype; linktype = llvm::GlobalValue::LinkOnceODRLinkage; if (RD->isInAnonymousNamespace()) linktype = llvm::GlobalValue::InternalLinkage; std::vector inits; llvm::Type *Ptr8Ty=llvm::PointerType::get(llvm::Type::getInt8Ty(VMContext),0); D1(printf("vtt %s\n", RD->getNameAsCString())); VTTBuilder b(inits, RD, *this); llvm::Constant *C; llvm::ArrayType *type = llvm::ArrayType::get(Ptr8Ty, inits.size()); C = llvm::ConstantArray::get(type, inits); llvm::GlobalVariable *vtt = new llvm::GlobalVariable(getModule(), type, true, linktype, C, Name); bool Hidden = getDeclVisibilityMode(RD) == LangOptions::Hidden; if (Hidden) vtt->setVisibility(llvm::GlobalVariable::HiddenVisibility); return llvm::ConstantExpr::getBitCast(vtt, Ptr8Ty); } void CGVtableInfo::GenerateClassData(const CXXRecordDecl *RD) { Vtables[RD] = CGM.GenerateVtable(RD, RD); CGM.GenerateRtti(RD); CGM.GenerateVTT(RD); } llvm::Constant *CGVtableInfo::getVtable(const CXXRecordDecl *RD) { llvm::Constant *&vtbl = Vtables[RD]; if (vtbl) return vtbl; vtbl = CGM.GenerateVtable(RD, RD); bool CreateDefinition = true; // We have to convert it to have a record layout. CGM.getTypes().ConvertTagDeclType(RD); const CGRecordLayout &CGLayout = CGM.getTypes().getCGRecordLayout(RD); if (const CXXMethodDecl *KeyFunction = CGLayout.getKeyFunction()) { if (!KeyFunction->getBody()) { // If there is a KeyFunction, and it isn't defined, just build a // reference to the vtable. CreateDefinition = false; } } if (CreateDefinition) { CGM.GenerateRtti(RD); CGM.GenerateVTT(RD); } return vtbl; } llvm::Constant *CGVtableInfo::getCtorVtable(const CXXRecordDecl *LayoutClass, const CXXRecordDecl *RD, uint64_t Offset) { return CGM.GenerateVtable(LayoutClass, RD, Offset); }