mirror of
https://gitlab.com/libeigen/eigen.git
synced 2024-12-21 07:19:46 +08:00
Use NULL instead of nullptr to avoid adding a cxx11 requirement.
This commit is contained in:
parent
3810ec228f
commit
26239ee580
@ -187,7 +187,7 @@ struct TensorEvaluator<const TensorAssignOp<LeftArgType, RightArgType>, Device>
|
|||||||
|
|
||||||
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void evalBlock(TensorBlock* block) {
|
EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void evalBlock(TensorBlock* block) {
|
||||||
if (TensorEvaluator<LeftArgType, Device>::RawAccess &&
|
if (TensorEvaluator<LeftArgType, Device>::RawAccess &&
|
||||||
m_leftImpl.data() != nullptr) {
|
m_leftImpl.data() != NULL) {
|
||||||
TensorBlock left_block(block->first_coeff_index(), block->block_sizes(),
|
TensorBlock left_block(block->first_coeff_index(), block->block_sizes(),
|
||||||
block->tensor_strides(), block->tensor_strides(),
|
block->tensor_strides(), block->tensor_strides(),
|
||||||
m_leftImpl.data() + block->first_coeff_index());
|
m_leftImpl.data() + block->first_coeff_index());
|
||||||
|
@ -102,7 +102,7 @@ class Allocator {
|
|||||||
// Build a thread pool device on top the an existing pool of threads.
|
// Build a thread pool device on top the an existing pool of threads.
|
||||||
struct ThreadPoolDevice {
|
struct ThreadPoolDevice {
|
||||||
// The ownership of the thread pool remains with the caller.
|
// The ownership of the thread pool remains with the caller.
|
||||||
ThreadPoolDevice(ThreadPoolInterface* pool, int num_cores, Allocator* allocator = nullptr)
|
ThreadPoolDevice(ThreadPoolInterface* pool, int num_cores, Allocator* allocator = NULL)
|
||||||
: pool_(pool), num_threads_(num_cores), allocator_(allocator) { }
|
: pool_(pool), num_threads_(num_cores), allocator_(allocator) { }
|
||||||
|
|
||||||
EIGEN_STRONG_INLINE void* allocate(size_t num_bytes) const {
|
EIGEN_STRONG_INLINE void* allocate(size_t num_bytes) const {
|
||||||
@ -282,7 +282,7 @@ struct ThreadPoolDevice {
|
|||||||
// Convenience wrapper for parallelFor that does not align blocks.
|
// Convenience wrapper for parallelFor that does not align blocks.
|
||||||
void parallelFor(Index n, const TensorOpCost& cost,
|
void parallelFor(Index n, const TensorOpCost& cost,
|
||||||
std::function<void(Index, Index)> f) const {
|
std::function<void(Index, Index)> f) const {
|
||||||
parallelFor(n, cost, nullptr, std::move(f));
|
parallelFor(n, cost, NULL, std::move(f));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Thread pool accessor.
|
// Thread pool accessor.
|
||||||
|
@ -227,7 +227,7 @@ class TensorExecutor<Expression, ThreadPoolDevice, Vectorizable, Tileable> {
|
|||||||
typedef EvalRange<Evaluator, StorageIndex, Vectorizable> EvalRange;
|
typedef EvalRange<Evaluator, StorageIndex, Vectorizable> EvalRange;
|
||||||
|
|
||||||
Evaluator evaluator(expr, device);
|
Evaluator evaluator(expr, device);
|
||||||
const bool needs_assign = evaluator.evalSubExprsIfNeeded(nullptr);
|
const bool needs_assign = evaluator.evalSubExprsIfNeeded(NULL);
|
||||||
if (needs_assign) {
|
if (needs_assign) {
|
||||||
const StorageIndex PacketSize =
|
const StorageIndex PacketSize =
|
||||||
Vectorizable
|
Vectorizable
|
||||||
@ -271,7 +271,7 @@ class TensorExecutor<Expression, ThreadPoolDevice, Vectorizable, /*Tileable*/ tr
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const bool needs_assign = evaluator.evalSubExprsIfNeeded(nullptr);
|
const bool needs_assign = evaluator.evalSubExprsIfNeeded(NULL);
|
||||||
if (needs_assign) {
|
if (needs_assign) {
|
||||||
TensorBlockShapeType block_shape = TensorBlockShapeType::kSkewedInnerDims;
|
TensorBlockShapeType block_shape = TensorBlockShapeType::kSkewedInnerDims;
|
||||||
Index block_total_size = 0;
|
Index block_total_size = 0;
|
||||||
|
Loading…
Reference in New Issue
Block a user