359 if (consistent_frag_sizes.empty()) {
364 const size_t num_columns =
366 size_t total_group_by_buffer_size{0};
367 for (
size_t i = 0; i < num_columns; ++i) {
368 auto ti = exe_unit.target_exprs[i]->get_type_info();
373 int64_t flatbuffer_size =
query_mem_desc.getFlatBufferSize(slot_idx);
374 total_group_by_buffer_size =
377 const size_t col_width = ti.get_size();
378 const size_t group_buffer_size = num_rows_ * col_width;
379 total_group_by_buffer_size =
386 total_group_by_buffer_size,
nullptr,
thread_idx_, row_set_mem_owner.get());
387 group_by_buffers_.push_back(group_by_buffer);
389 const auto column_frag_offsets =
391 const auto column_frag_sizes =
393 result_sets_.emplace_back(
403 executor->blockSize(),
404 executor->gridSize()));
405 result_sets_.back()->allocateStorage(reinterpret_cast<int8_t*>(group_by_buffer),
std::vector< int64_t > get_consistent_frags_sizes(const std::vector< std::vector< uint64_t >> &frag_offsets)
static QueryMemoryDescriptor fixupQueryMemoryDescriptor(const QueryMemoryDescriptor &)
std::vector< TargetInfo > target_exprs_to_infos(const std::vector< Analyzer::Expr * > &targets, const QueryMemoryDescriptor &query_mem_desc)
int64_t * alloc_group_by_buffer(const size_t numBytes, RenderAllocatorMap *render_allocator_map, const size_t thread_idx, RowSetMemoryOwner *mem_owner)
FORCE_INLINE HOST DEVICE T align_to_int64(T addr)
std::vector< std::vector< int64_t > > get_col_frag_offsets(const std::vector< Analyzer::Expr * > &target_exprs, const std::vector< std::vector< uint64_t >> &table_frag_offsets)