// Copyright 2009-2020 Intel Corporation // SPDX-License-Identifier: Apache-2.0 #include "bvh.h" #include "bvh_statistics.h" namespace embree { template BVHN::BVHN (const PrimitiveType& primTy, Scene* scene) : AccelData((N==4) ? AccelData::TY_BVH4 : (N==8) ? AccelData::TY_BVH8 : AccelData::TY_UNKNOWN), primTy(&primTy), device(scene->device), scene(scene), root(emptyNode), alloc(scene->device,scene->isStaticAccel()), numPrimitives(0), numVertices(0) { } template BVHN::~BVHN () { for (size_t i=0; i void BVHN::clear() { set(BVHN::emptyNode,empty,0); alloc.clear(); } template void BVHN::set (NodeRef root, const LBBox3fa& bounds, size_t numPrimitives) { this->root = root; this->bounds = bounds; this->numPrimitives = numPrimitives; } template void BVHN::clearBarrier(NodeRef& node) { if (node.isBarrier()) node.clearBarrier(); else if (!node.isLeaf()) { BaseNode* n = node.baseNode(); // FIXME: flags should be stored in BVH for (size_t c=0; cchild(c)); } } template void BVHN::layoutLargeNodes(size_t num) { #if defined(__X86_64__) || defined(__aarch64__) // do not use tree rotations on 32 bit platforms, barrier bit in NodeRef will cause issues struct NodeArea { __forceinline NodeArea() {} __forceinline NodeArea(NodeRef& node, const BBox3fa& bounds) : node(&node), A(node.isLeaf() ? float(neg_inf) : area(bounds)) {} __forceinline bool operator< (const NodeArea& other) const { return this->A < other.A; } NodeRef* node; float A; }; std::vector lst; lst.reserve(num); lst.push_back(NodeArea(root,empty)); while (lst.size() < num) { std::pop_heap(lst.begin(), lst.end()); NodeArea n = lst.back(); lst.pop_back(); if (!n.node->isAABBNode()) break; AABBNode* node = n.node->getAABBNode(); for (size_t i=0; ichild(i) == BVHN::emptyNode) continue; lst.push_back(NodeArea(node->child(i),node->bounds(i))); std::push_heap(lst.begin(), lst.end()); } } for (size_t i=0; isetBarrier(); root = layoutLargeNodesRecursion(root,alloc.getCachedAllocator()); #endif } template typename BVHN::NodeRef BVHN::layoutLargeNodesRecursion(NodeRef& node, const FastAllocator::CachedAllocator& allocator) { if (node.isBarrier()) { node.clearBarrier(); return node; } else if (node.isAABBNode()) { AABBNode* oldnode = node.getAABBNode(); AABBNode* newnode = (BVHN::AABBNode*) allocator.malloc0(sizeof(BVHN::AABBNode),byteNodeAlignment); *newnode = *oldnode; for (size_t c=0; cchild(c) = layoutLargeNodesRecursion(oldnode->child(c),allocator); return encodeNode(newnode); } else return node; } template double BVHN::preBuild(const std::string& builderName) { if (builderName == "") return inf; if (device->verbosity(2)) { Lock lock(g_printMutex); std::cout << "building BVH" << N << (builderName.find("MBlur") != std::string::npos ? "MB" : "") << "<" << primTy->name() << "> using " << builderName << " ..." << std::endl << std::flush; } double t0 = 0.0; if (device->benchmark || device->verbosity(2)) t0 = getSeconds(); return t0; } template void BVHN::postBuild(double t0) { if (t0 == double(inf)) return; double dt = 0.0; if (device->benchmark || device->verbosity(2)) dt = getSeconds()-t0; std::unique_ptr> stat; /* print statistics */ if (device->verbosity(2)) { if (!stat) stat.reset(new BVHNStatistics(this)); const size_t usedBytes = alloc.getUsedBytes(); Lock lock(g_printMutex); std::cout << "finished BVH" << N << "<" << primTy->name() << "> : " << 1000.0f*dt << "ms, " << 1E-6*double(numPrimitives)/dt << " Mprim/s, " << 1E-9*double(usedBytes)/dt << " GB/s" << std::endl; if (device->verbosity(2)) std::cout << stat->str(); if (device->verbosity(2)) { FastAllocator::AllStatistics stat(&alloc); for (size_t i=0; ialloc); stat.print(numPrimitives); } if (device->verbosity(3)) { alloc.print_blocks(); for (size_t i=0; ialloc.print_blocks(); } std::cout << std::flush; } /* benchmark mode */ if (device->benchmark) { if (!stat) stat.reset(new BVHNStatistics(this)); Lock lock(g_printMutex); std::cout << "BENCHMARK_BUILD " << dt << " " << double(numPrimitives)/dt << " " << stat->sah() << " " << stat->bytesUsed() << " BVH" << N << "<" << primTy->name() << ">" << std::endl << std::flush; } } #if defined(__AVX__) template class BVHN<8>; #endif #if !defined(__AVX__) || !defined(EMBREE_TARGET_SSE2) && !defined(EMBREE_TARGET_SSE42) || defined(__aarch64__) template class BVHN<4>; #endif }