summaryrefslogtreecommitdiff
path: root/thirdparty/basis_universal/encoder/basisu_frontend.cpp
diff options
context:
space:
mode:
authorK. S. Ernest (iFire) Lee <ernest.lee@chibifire.com>2022-03-24 12:39:24 -0700
committerK. S. Ernest (iFire) Lee <ernest.lee@chibifire.com>2022-03-24 22:41:23 -0700
commit3529141b4bf548218f70a713b4328220a566802d (patch)
treea9e0ba416efc19eda40193522f52659feb865256 /thirdparty/basis_universal/encoder/basisu_frontend.cpp
parentd250f12243a74fd36905f46b983e3c142e3b2e67 (diff)
Update basis universal to version 1.16.3.
Enable basis universal uastc internal storage instead of etc1s for better quality.
Diffstat (limited to 'thirdparty/basis_universal/encoder/basisu_frontend.cpp')
-rw-r--r--thirdparty/basis_universal/encoder/basisu_frontend.cpp1789
1 files changed, 1143 insertions, 646 deletions
diff --git a/thirdparty/basis_universal/encoder/basisu_frontend.cpp b/thirdparty/basis_universal/encoder/basisu_frontend.cpp
index 324fc8e447..00210e6679 100644
--- a/thirdparty/basis_universal/encoder/basisu_frontend.cpp
+++ b/thirdparty/basis_universal/encoder/basisu_frontend.cpp
@@ -19,6 +19,7 @@
//
#include "../transcoder/basisu.h"
#include "basisu_frontend.h"
+#include "basisu_opencl.h"
#include <unordered_set>
#include <unordered_map>
@@ -43,61 +44,15 @@ namespace basisu
// TODO - How to handle internal verifies in the basisu lib
static inline void handle_verify_failure(int line)
{
- fprintf(stderr, "ERROR: basisu_frontend: verify check failed at line %i!\n", line);
+ error_printf("basisu_frontend: verify check failed at line %i!\n", line);
abort();
}
bool basisu_frontend::init(const params &p)
{
-#if 0
- // HACK HACK
- FILE* pFile;
- fopen_s(&pFile, "tv.bin", "rb");
- if (pFile)
- {
- debug_printf("Using tv.bin\n");
-
- fseek(pFile, 0, SEEK_END);
- uint32_t size = ftell(pFile);
- fseek(pFile, 0, SEEK_SET);
-
- uint32_t tv = size / sizeof(vec6F_quantizer::training_vec_with_weight);
-
- basisu::vector<vec6F_quantizer::training_vec_with_weight> v(tv);
- fread(&v[0], 1, sizeof(v[0]) * tv, pFile);
-
- for (uint32_t i = 0; i < tv; i++)
- m_endpoint_clusterizer.add_training_vec(v[i].first, v[i].second);
-
- m_endpoint_clusterizer.generate(16128);
- basisu::vector<uint_vec> codebook;
- m_endpoint_clusterizer.retrieve(codebook);
-
- printf("Generated %u entries\n", (uint32_t)codebook.size());
-
- fclose(pFile);
- exit(0);
- }
-#endif
-
- if (p.m_use_hybrid_selector_codebooks)
- {
- if (!p.m_pGlobal_sel_codebook)
- {
- debug_printf("basisu_frontend::init: No global sel codebook!\n");
- assert(0);
- return false;
- }
- }
-
- debug_printf("basisu_frontend::init: Multithreaded: %u, NumEndpointClusters: %u, NumSelectorClusters: %u, Perceptual: %u, CompressionLevel: %u\n",
- p.m_multithreaded, p.m_max_endpoint_clusters, p.m_max_selector_clusters, p.m_perceptual, p.m_compression_level);
-
- debug_printf("Global sel codebook pal bits: %u, Global sel codebook mod bits: %u, Use hybrid selector codebook: %u, Hybrid codebook quality thresh: %f\n",
- p.m_num_global_sel_codebook_pal_bits,
- p.m_num_global_sel_codebook_mod_bits,
- p.m_use_hybrid_selector_codebooks,
- p.m_hybrid_codebook_quality_thresh);
+ debug_printf("basisu_frontend::init: Multithreaded: %u, Job pool total threads: %u, NumEndpointClusters: %u, NumSelectorClusters: %u, Perceptual: %u, CompressionLevel: %u\n",
+ p.m_multithreaded, p.m_pJob_pool ? p.m_pJob_pool->get_total_threads() : 0,
+ p.m_max_endpoint_clusters, p.m_max_selector_clusters, p.m_perceptual, p.m_compression_level);
if ((p.m_max_endpoint_clusters < 1) || (p.m_max_endpoint_clusters > cMaxEndpointClusters))
return false;
@@ -106,8 +61,22 @@ namespace basisu
m_source_blocks.resize(0);
append_vector(m_source_blocks, p.m_pSource_blocks, p.m_num_source_blocks);
-
+
m_params = p;
+
+ if (m_params.m_pOpenCL_context)
+ {
+ BASISU_ASSUME(sizeof(cl_pixel_block) == sizeof(pixel_block));
+
+ // Upload the RGBA pixel blocks a single time.
+ if (!opencl_set_pixel_blocks(m_params.m_pOpenCL_context, m_source_blocks.size(), (cl_pixel_block*)m_source_blocks.data()))
+ {
+ // This is not fatal, we just won't use OpenCL.
+ error_printf("basisu_frontend::init: opencl_set_pixel_blocks() failed\n");
+ m_params.m_pOpenCL_context = nullptr;
+ m_opencl_failed = true;
+ }
+ }
m_encoded_blocks.resize(m_params.m_num_source_blocks);
memset(&m_encoded_blocks[0], 0, m_encoded_blocks.size() * sizeof(m_encoded_blocks[0]));
@@ -194,8 +163,12 @@ namespace basisu
m_total_blocks = m_params.m_num_source_blocks;
m_total_pixels = m_total_blocks * cPixelBlockTotalPixels;
+ // Encode the initial high quality ETC1S texture
+
init_etc1_images();
+ // First quantize the ETC1S endpoints
+
if (m_params.m_pGlobal_codebooks)
{
init_global_codebooks();
@@ -205,16 +178,26 @@ namespace basisu
init_endpoint_training_vectors();
generate_endpoint_clusters();
-
+
for (uint32_t refine_endpoint_step = 0; refine_endpoint_step < m_num_endpoint_codebook_iterations; refine_endpoint_step++)
{
- BASISU_FRONTEND_VERIFY(check_etc1s_constraints());
+ if (m_params.m_validate)
+ {
+ BASISU_FRONTEND_VERIFY(check_etc1s_constraints());
+
+ BASISU_FRONTEND_VERIFY(validate_endpoint_cluster_hierarchy(false));
+ }
if (refine_endpoint_step)
{
introduce_new_endpoint_clusters();
}
+ if (m_params.m_validate)
+ {
+ BASISU_FRONTEND_VERIFY(validate_endpoint_cluster_hierarchy(false));
+ }
+
generate_endpoint_codebook(refine_endpoint_step);
if ((m_params.m_debug_images) && (m_params.m_dump_endpoint_clusterization))
@@ -236,7 +219,7 @@ namespace basisu
if ((m_params.m_tex_type == basist::cBASISTexTypeVideoFrames) && (!refine_endpoint_step) && (m_num_endpoint_codebook_iterations == 1))
{
eliminate_redundant_or_empty_endpoint_clusters();
- generate_endpoint_codebook(refine_endpoint_step);
+ generate_endpoint_codebook(basisu::maximum(1U, refine_endpoint_step));
}
if ((m_params.m_debug_images) && (m_params.m_dump_endpoint_clusterization))
@@ -250,22 +233,37 @@ namespace basisu
dump_endpoint_clusterization_visualization(buf, true);
}
}
+
+ if (m_params.m_validate)
+ {
+ BASISU_FRONTEND_VERIFY(validate_endpoint_cluster_hierarchy(false));
+ }
eliminate_redundant_or_empty_endpoint_clusters();
+ if (m_params.m_validate)
+ {
+ BASISU_FRONTEND_VERIFY(validate_endpoint_cluster_hierarchy(false));
+ }
+
if (m_params.m_debug_stats)
debug_printf("Total endpoint clusters: %u\n", (uint32_t)m_endpoint_clusters.size());
if (early_out)
break;
}
-
- BASISU_FRONTEND_VERIFY(check_etc1s_constraints());
+
+ if (m_params.m_validate)
+ {
+ BASISU_FRONTEND_VERIFY(check_etc1s_constraints());
+ }
generate_block_endpoint_clusters();
create_initial_packed_texture();
+ // Now quantize the ETC1S selectors
+
generate_selector_clusters();
if (m_use_hierarchical_selector_codebooks)
@@ -276,12 +274,12 @@ namespace basisu
create_optimized_selector_codebook(0);
find_optimal_selector_clusters_for_each_block();
-
+
introduce_special_selector_clusters();
}
else
{
- const uint32_t num_refine_selector_steps = m_params.m_pGlobal_sel_codebook ? 1 : m_num_selector_codebook_iterations;
+ const uint32_t num_refine_selector_steps = m_num_selector_codebook_iterations;
for (uint32_t refine_selector_steps = 0; refine_selector_steps < num_refine_selector_steps; refine_selector_steps++)
{
create_optimized_selector_codebook(refine_selector_steps);
@@ -289,7 +287,7 @@ namespace basisu
find_optimal_selector_clusters_for_each_block();
introduce_special_selector_clusters();
-
+
if ((m_params.m_compression_level >= 4) || (m_params.m_tex_type == basist::cBASISTexTypeVideoFrames))
{
if (!refine_block_endpoints_given_selectors())
@@ -297,7 +295,7 @@ namespace basisu
}
}
}
-
+
optimize_selector_codebook();
if (m_params.m_debug_stats)
@@ -574,9 +572,6 @@ namespace basisu
{
debug_printf("introduce_special_selector_clusters\n");
- if (m_params.m_pGlobal_sel_codebook)
- return;
-
uint32_t total_blocks_relocated = 0;
const uint32_t initial_selector_clusters = (uint32_t)m_selector_cluster_block_indices.size();
@@ -722,23 +717,15 @@ namespace basisu
}
basisu::vector<etc_block> new_optimized_cluster_selectors(m_optimized_cluster_selectors.size() ? total_new_entries : 0);
- basist::etc1_global_selector_codebook_entry_id_vec new_optimized_cluster_selector_global_cb_ids(m_optimized_cluster_selector_global_cb_ids.size() ? total_new_entries : 0);
basisu::vector<uint_vec> new_selector_cluster_indices(m_selector_cluster_block_indices.size() ? total_new_entries : 0);
- bool_vec new_selector_cluster_uses_global_cb(m_selector_cluster_uses_global_cb.size() ? total_new_entries : 0);
for (uint32_t i = 0; i < total_new_entries; i++)
{
if (m_optimized_cluster_selectors.size())
new_optimized_cluster_selectors[i] = m_optimized_cluster_selectors[new_to_old[i]];
- if (m_optimized_cluster_selector_global_cb_ids.size())
- new_optimized_cluster_selector_global_cb_ids[i] = m_optimized_cluster_selector_global_cb_ids[new_to_old[i]];
-
//if (m_selector_cluster_block_indices.size())
// new_selector_cluster_indices[i] = m_selector_cluster_block_indices[new_to_old[i]];
-
- if (m_selector_cluster_uses_global_cb.size())
- new_selector_cluster_uses_global_cb[i] = m_selector_cluster_uses_global_cb[new_to_old[i]];
}
for (uint32_t i = 0; i < m_block_selector_cluster_index.size(); i++)
@@ -747,9 +734,7 @@ namespace basisu
}
m_optimized_cluster_selectors.swap(new_optimized_cluster_selectors);
- m_optimized_cluster_selector_global_cb_ids.swap(new_optimized_cluster_selector_global_cb_ids);
m_selector_cluster_block_indices.swap(new_selector_cluster_indices);
- m_selector_cluster_uses_global_cb.swap(new_selector_cluster_uses_global_cb);
// This isn't strictly necessary - doing it for completeness/future sanity.
if (m_selector_clusters_within_each_parent_cluster.size())
@@ -771,66 +756,93 @@ namespace basisu
m_etc1_blocks_etc1s.resize(m_total_blocks);
- const uint32_t N = 4096;
- for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
+ bool use_cpu = true;
+
+ if (m_params.m_pOpenCL_context)
+ {
+ uint32_t total_perms = 64;
+ if (m_params.m_compression_level == 0)
+ total_perms = 4;
+ else if (m_params.m_compression_level == 1)
+ total_perms = 16;
+ else if (m_params.m_compression_level == BASISU_MAX_COMPRESSION_LEVEL)
+ total_perms = OPENCL_ENCODE_ETC1S_MAX_PERMS;
+
+ bool status = opencl_encode_etc1s_blocks(m_params.m_pOpenCL_context, m_etc1_blocks_etc1s.data(), m_params.m_perceptual, total_perms);
+ if (status)
+ use_cpu = false;
+ else
+ {
+ error_printf("basisu_frontend::init_etc1_images: opencl_encode_etc1s_blocks() failed! Using CPU.\n");
+ m_params.m_pOpenCL_context = nullptr;
+ m_opencl_failed = true;
+ }
+ }
+
+ if (use_cpu)
{
- const uint32_t first_index = block_index_iter;
- const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
+ const uint32_t N = 4096;
+ for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
+ {
+ const uint32_t first_index = block_index_iter;
+ const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index] {
+ m_params.m_pJob_pool->add_job([this, first_index, last_index] {
#endif
- for (uint32_t block_index = first_index; block_index < last_index; block_index++)
- {
- const pixel_block &source_blk = get_source_pixel_block(block_index);
+ for (uint32_t block_index = first_index; block_index < last_index; block_index++)
+ {
+ const pixel_block& source_blk = get_source_pixel_block(block_index);
- etc1_optimizer optimizer;
- etc1_optimizer::params optimizer_params;
- etc1_optimizer::results optimizer_results;
-
- if (m_params.m_compression_level == 0)
- optimizer_params.m_quality = cETCQualityFast;
- else if (m_params.m_compression_level == 1)
- optimizer_params.m_quality = cETCQualityMedium;
- else if (m_params.m_compression_level == BASISU_MAX_COMPRESSION_LEVEL)
- optimizer_params.m_quality = cETCQualityUber;
-
- optimizer_params.m_num_src_pixels = 16;
- optimizer_params.m_pSrc_pixels = source_blk.get_ptr();
- optimizer_params.m_perceptual = m_params.m_perceptual;
+ etc1_optimizer optimizer;
+ etc1_optimizer::params optimizer_params;
+ etc1_optimizer::results optimizer_results;
- uint8_t selectors[16];
- optimizer_results.m_pSelectors = selectors;
- optimizer_results.m_n = 16;
+ if (m_params.m_compression_level == 0)
+ optimizer_params.m_quality = cETCQualityFast;
+ else if (m_params.m_compression_level == 1)
+ optimizer_params.m_quality = cETCQualityMedium;
+ else if (m_params.m_compression_level == BASISU_MAX_COMPRESSION_LEVEL)
+ optimizer_params.m_quality = cETCQualityUber;
- optimizer.init(optimizer_params, optimizer_results);
- if (!optimizer.compute())
- BASISU_FRONTEND_VERIFY(false);
+ optimizer_params.m_num_src_pixels = 16;
+ optimizer_params.m_pSrc_pixels = source_blk.get_ptr();
+ optimizer_params.m_perceptual = m_params.m_perceptual;
- etc_block &blk = m_etc1_blocks_etc1s[block_index];
+ uint8_t selectors[16];
+ optimizer_results.m_pSelectors = selectors;
+ optimizer_results.m_n = 16;
- memset(&blk, 0, sizeof(blk));
- blk.set_block_color5_etc1s(optimizer_results.m_block_color_unscaled);
- blk.set_inten_tables_etc1s(optimizer_results.m_block_inten_table);
- blk.set_flip_bit(true);
+ optimizer.init(optimizer_params, optimizer_results);
+ if (!optimizer.compute())
+ BASISU_FRONTEND_VERIFY(false);
- for (uint32_t y = 0; y < 4; y++)
- for (uint32_t x = 0; x < 4; x++)
- blk.set_selector(x, y, selectors[x + y * 4]);
- }
+ etc_block& blk = m_etc1_blocks_etc1s[block_index];
+
+ memset(&blk, 0, sizeof(blk));
+ blk.set_block_color5_etc1s(optimizer_results.m_block_color_unscaled);
+ blk.set_inten_tables_etc1s(optimizer_results.m_block_inten_table);
+ blk.set_flip_bit(true);
+
+ for (uint32_t y = 0; y < 4; y++)
+ for (uint32_t x = 0; x < 4; x++)
+ blk.set_selector(x, y, selectors[x + y * 4]);
+ }
#ifndef __EMSCRIPTEN__
- } );
+ });
#endif
- }
+ }
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->wait_for_all();
+ m_params.m_pJob_pool->wait_for_all();
#endif
- debug_printf("Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
+ } // use_cpu
+
+ debug_printf("init_etc1_images: Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
}
void basisu_frontend::init_endpoint_training_vectors()
@@ -889,13 +901,15 @@ namespace basisu
const uint32_t parent_codebook_size = (m_params.m_max_endpoint_clusters >= 256) ? BASISU_ENDPOINT_PARENT_CODEBOOK_SIZE : 0;
uint32_t max_threads = 0;
max_threads = m_params.m_multithreaded ? minimum<int>(std::thread::hardware_concurrency(), cMaxCodebookCreationThreads) : 0;
+ if (m_params.m_pJob_pool)
+ max_threads = minimum<int>((int)m_params.m_pJob_pool->get_total_threads(), max_threads);
- debug_printf("Using %u threads to create codebook\n", max_threads);
+ debug_printf("max_threads: %u\n", max_threads);
bool status = generate_hierarchical_codebook_threaded(m_endpoint_clusterizer,
m_params.m_max_endpoint_clusters, m_use_hierarchical_endpoint_codebooks ? parent_codebook_size : 0,
m_endpoint_clusters,
m_endpoint_parent_clusters,
- max_threads, m_params.m_pJob_pool);
+ max_threads, m_params.m_pJob_pool, true);
BASISU_FRONTEND_VERIFY(status);
if (m_use_hierarchical_endpoint_codebooks)
@@ -940,6 +954,9 @@ namespace basisu
for (uint32_t j = 0; j < cluster.size(); j++)
{
const uint32_t block_index = cluster[j] >> 1;
+
+ BASISU_FRONTEND_VERIFY(block_index < m_block_parent_endpoint_cluster.size());
+
if (!j)
{
parent_cluster_index = m_block_parent_endpoint_cluster[block_index];
@@ -956,6 +973,7 @@ namespace basisu
debug_printf("Total endpoint clusters: %u, parent clusters: %u\n", (uint32_t)m_endpoint_clusters.size(), (uint32_t)m_endpoint_parent_clusters.size());
}
+ // Iterate through each array of endpoint cluster block indices and set the m_block_endpoint_clusters_indices[][] array to indicaste which cluster index each block uses.
void basisu_frontend::generate_block_endpoint_clusters()
{
m_block_endpoint_clusters_indices.resize(m_total_blocks);
@@ -974,11 +992,14 @@ namespace basisu
} // cluster_indices_iter
}
- for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
+ if (m_params.m_validate)
{
- uint32_t cluster_0 = m_block_endpoint_clusters_indices[block_index][0];
- uint32_t cluster_1 = m_block_endpoint_clusters_indices[block_index][1];
- BASISU_FRONTEND_VERIFY(cluster_0 == cluster_1);
+ for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
+ {
+ uint32_t cluster_0 = m_block_endpoint_clusters_indices[block_index][0];
+ uint32_t cluster_1 = m_block_endpoint_clusters_indices[block_index][1];
+ BASISU_FRONTEND_VERIFY(cluster_0 == cluster_1);
+ }
}
}
@@ -989,6 +1010,7 @@ namespace basisu
m_endpoint_clusters_within_each_parent_cluster.resize(0);
m_endpoint_clusters_within_each_parent_cluster.resize(m_endpoint_parent_clusters.size());
+ // Note: It's possible that some blocks got moved into the same cluster, but live in different parent clusters.
for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
{
const uint32_t cluster_index = m_block_endpoint_clusters_indices[block_index][0];
@@ -1125,6 +1147,8 @@ namespace basisu
std::unordered_set<uint32_t> ignore_cluster;
+ uint32_t total_new_clusters = 0;
+
while (num_new_endpoint_clusters)
{
if (m_subblock_endpoint_quant_err_vec.size() == 0)
@@ -1173,10 +1197,14 @@ namespace basisu
cluster_sizes[subblock_to_move.m_cluster_index] -= 2;
ignore_cluster.insert(subblock_to_move.m_cluster_index);
+
+ total_new_clusters++;
num_new_endpoint_clusters--;
}
+ debug_printf("Introduced %i new endpoint clusters\n", total_new_clusters);
+
for (uint32_t i = 0; i < num_orig_endpoint_clusters; i++)
{
uint_vec &cluster_indices = m_endpoint_clusters[i];
@@ -1200,150 +1228,433 @@ namespace basisu
generate_block_endpoint_clusters();
}
+ struct color_rgba_hasher
+ {
+ inline std::size_t operator()(const color_rgba& k) const
+ {
+ uint32_t v = *(const uint32_t*)&k;
+
+ //return bitmix32(v);
+
+ //v ^= (v << 10);
+ //v ^= (v >> 12);
+
+ return v;
+ }
+ };
+
// Given each endpoint cluster, gather all the block pixels which are in that cluster and compute optimized ETC1S endpoints for them.
// TODO: Don't optimize endpoint clusters which haven't changed.
+ // If step>=1, we check to ensure the new endpoint values actually decrease quantization error.
void basisu_frontend::generate_endpoint_codebook(uint32_t step)
{
debug_printf("generate_endpoint_codebook\n");
+
+ interval_timer tm;
+ tm.start();
m_endpoint_cluster_etc_params.resize(m_endpoint_clusters.size());
- const uint32_t N = 128;
- for (uint32_t cluster_index_iter = 0; cluster_index_iter < m_endpoint_clusters.size(); cluster_index_iter += N)
+ bool use_cpu = true;
+ // TODO: Get this working when step>0
+ if (m_params.m_pOpenCL_context && !step)
{
- const uint32_t first_index = cluster_index_iter;
- const uint32_t last_index = minimum<uint32_t>((uint32_t)m_endpoint_clusters.size(), cluster_index_iter + N);
+ const uint32_t total_clusters = m_endpoint_clusters.size();
-#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index, step ] {
-#endif
+ basisu::vector<cl_pixel_cluster> pixel_clusters(total_clusters);
+
+ std::vector<color_rgba> input_pixels;
+ input_pixels.reserve(m_total_blocks * 16);
- for (uint32_t cluster_index = first_index; cluster_index < last_index; cluster_index++)
+ std::vector<uint32_t> pixel_weights;
+ pixel_weights.reserve(m_total_blocks * 16);
+
+ uint_vec cluster_sizes(total_clusters);
+
+ //typedef basisu::hash_map<color_rgba, uint32_t, color_rgba_hasher> color_hasher_type;
+ //color_hasher_type color_hasher;
+ //color_hasher.reserve(2048);
+
+ interval_timer hash_tm;
+ hash_tm.start();
+
+ basisu::vector<uint32_t> colors, colors2;
+ colors.reserve(65536);
+ colors2.reserve(65536);
+
+ for (uint32_t cluster_index = 0; cluster_index < m_endpoint_clusters.size(); cluster_index++)
+ {
+ const basisu::vector<uint32_t>& cluster_indices = m_endpoint_clusters[cluster_index];
+ assert((cluster_indices.size() & 1) == 0);
+
+#if 0
+ uint64_t first_pixel_index = input_pixels.size();
+ const uint32_t total_pixels = 16 * (cluster_indices.size() / 2);
+
+ input_pixels.resize(input_pixels.size() + total_pixels);
+ pixel_weights.resize(pixel_weights.size() + total_pixels);
+
+ uint64_t dst_ofs = first_pixel_index;
+
+ uint64_t total_r = 0, total_g = 0, total_b = 0;
+ for (uint32_t cluster_indices_iter = 0; cluster_indices_iter < cluster_indices.size(); cluster_indices_iter++)
{
- const basisu::vector<uint32_t>& cluster_indices = m_endpoint_clusters[cluster_index];
+ const uint32_t subblock_index = cluster_indices[cluster_indices_iter] & 1;
+ if (subblock_index)
+ continue;
- BASISU_FRONTEND_VERIFY(cluster_indices.size());
+ const uint32_t block_index = cluster_indices[cluster_indices_iter] >> 1;
+ const color_rgba* pBlock_pixels = get_source_pixel_block(block_index).get_ptr();
- const uint32_t total_pixels = (uint32_t)cluster_indices.size() * 8;
+ for (uint32_t i = 0; i < 16; i++)
+ {
+ input_pixels[dst_ofs] = pBlock_pixels[i];
+ pixel_weights[dst_ofs] = 1;
+ dst_ofs++;
- basisu::vector<color_rgba> cluster_pixels(total_pixels);
+ total_r += pBlock_pixels[i].r;
+ total_g += pBlock_pixels[i].g;
+ total_b += pBlock_pixels[i].b;
+ }
+ }
- for (uint32_t cluster_indices_iter = 0; cluster_indices_iter < cluster_indices.size(); cluster_indices_iter++)
+ //printf("%i %f %f %f\n", cluster_index, total_r / (float)total_pixels, total_g / (float)total_pixels, total_b / (float)total_pixels);
+
+ pixel_clusters[cluster_index].m_first_pixel_index = first_pixel_index;
+ pixel_clusters[cluster_index].m_total_pixels = total_pixels;
+ cluster_sizes[cluster_index] = total_pixels;
+#elif 1
+ colors.resize(cluster_indices.size() * 8);
+ colors2.resize(cluster_indices.size() * 8);
+ uint32_t dst_ofs = 0;
+
+ for (uint32_t cluster_indices_iter = 0; cluster_indices_iter < cluster_indices.size(); cluster_indices_iter++)
+ {
+ const uint32_t subblock_index = cluster_indices[cluster_indices_iter] & 1;
+ if (subblock_index)
+ continue;
+
+ const uint32_t block_index = cluster_indices[cluster_indices_iter] >> 1;
+ const color_rgba* pBlock_pixels = get_source_pixel_block(block_index).get_ptr();
+
+ memcpy(colors.data() + dst_ofs, pBlock_pixels, sizeof(color_rgba) * 16);
+ dst_ofs += 16;
+
+ } // cluster_indices_iter
+
+ uint32_t* pSorted = radix_sort(colors.size(), colors.data(), colors2.data(), 0, 3);
+
+ const uint64_t first_pixel_index = input_pixels.size();
+
+ uint32_t prev_color = 0, cur_weight = 0;
+
+ for (uint32_t i = 0; i < colors.size(); i++)
+ {
+ uint32_t cur_color = pSorted[i];
+ if (cur_color == prev_color)
{
- const uint32_t block_index = cluster_indices[cluster_indices_iter] >> 1;
- const uint32_t subblock_index = cluster_indices[cluster_indices_iter] & 1;
+ if (++cur_weight == 0)
+ cur_weight--;
+ }
+ else
+ {
+ if (cur_weight)
+ {
+ input_pixels.push_back(*(const color_rgba*)&prev_color);
+ pixel_weights.push_back(cur_weight);
+ }
- const bool flipped = true;
+ prev_color = cur_color;
+ cur_weight = 1;
+ }
+ }
- const color_rgba *pBlock_pixels = get_source_pixel_block(block_index).get_ptr();
+ if (cur_weight)
+ {
+ input_pixels.push_back(*(const color_rgba*)&prev_color);
+ pixel_weights.push_back(cur_weight);
+ }
- for (uint32_t pixel_index = 0; pixel_index < 8; pixel_index++)
+ uint32_t total_unique_pixels = (uint32_t)(input_pixels.size() - first_pixel_index);
+
+ pixel_clusters[cluster_index].m_first_pixel_index = first_pixel_index;
+ pixel_clusters[cluster_index].m_total_pixels = total_unique_pixels;
+
+ cluster_sizes[cluster_index] = total_unique_pixels;
+#else
+ color_hasher.reset();
+
+ for (uint32_t cluster_indices_iter = 0; cluster_indices_iter < cluster_indices.size(); cluster_indices_iter++)
+ {
+ const uint32_t subblock_index = cluster_indices[cluster_indices_iter] & 1;
+ if (subblock_index)
+ continue;
+
+ const uint32_t block_index = cluster_indices[cluster_indices_iter] >> 1;
+ const color_rgba* pBlock_pixels = get_source_pixel_block(block_index).get_ptr();
+
+ uint32_t *pPrev_weight = nullptr;
+ color_rgba prev_color;
+
+ {
+ color_rgba cur_color = pBlock_pixels[0];
+ auto res = color_hasher.insert(cur_color, 0);
+
+ uint32_t& weight = (res.first)->second;
+ if (weight != UINT32_MAX)
+ weight++;
+
+ prev_color = cur_color;
+ pPrev_weight = &(res.first)->second;
+ }
+
+ for (uint32_t i = 1; i < 16; i++)
+ {
+ color_rgba cur_color = pBlock_pixels[i];
+
+ if (cur_color == prev_color)
+ {
+ if (*pPrev_weight != UINT32_MAX)
+ *pPrev_weight = *pPrev_weight + 1;
+ }
+ else
{
- const color_rgba &c = pBlock_pixels[g_etc1_pixel_indices[flipped][subblock_index][pixel_index]];
- cluster_pixels[cluster_indices_iter * 8 + pixel_index] = c;
+ auto res = color_hasher.insert(cur_color, 0);
+
+ uint32_t& weight = (res.first)->second;
+ if (weight != UINT32_MAX)
+ weight++;
+
+ prev_color = cur_color;
+ pPrev_weight = &(res.first)->second;
}
}
- endpoint_cluster_etc_params new_subblock_params;
+ } // cluster_indices_iter
+
+ const uint64_t first_pixel_index = input_pixels.size();
+ uint32_t total_unique_pixels = color_hasher.size();
+
+ pixel_clusters[cluster_index].m_first_pixel_index = first_pixel_index;
+ pixel_clusters[cluster_index].m_total_pixels = total_unique_pixels;
+
+ input_pixels.resize(first_pixel_index + total_unique_pixels);
+ pixel_weights.resize(first_pixel_index + total_unique_pixels);
- {
- etc1_optimizer optimizer;
- etc1_solution_coordinates solutions[2];
+ uint32_t j = 0;
+
+ for (auto it = color_hasher.begin(); it != color_hasher.end(); ++it, ++j)
+ {
+ input_pixels[first_pixel_index + j] = it->first;
+ pixel_weights[first_pixel_index + j] = it->second;
+ }
- etc1_optimizer::params cluster_optimizer_params;
- cluster_optimizer_params.m_num_src_pixels = total_pixels;
- cluster_optimizer_params.m_pSrc_pixels = &cluster_pixels[0];
+ cluster_sizes[cluster_index] = total_unique_pixels;
+#endif
- cluster_optimizer_params.m_use_color4 = false;
- cluster_optimizer_params.m_perceptual = m_params.m_perceptual;
+ } // cluster_index
- if (m_params.m_compression_level <= 1)
- cluster_optimizer_params.m_quality = cETCQualityMedium;
- else if (m_params.m_compression_level == BASISU_MAX_COMPRESSION_LEVEL)
- cluster_optimizer_params.m_quality = cETCQualityUber;
+ debug_printf("Total hash time: %3.3f secs\n", hash_tm.get_elapsed_secs());
- etc1_optimizer::results cluster_optimizer_results;
+ debug_printf("Total unique colors: %llu\n", input_pixels.size());
- basisu::vector<uint8_t> cluster_selectors(total_pixels);
- cluster_optimizer_results.m_n = total_pixels;
- cluster_optimizer_results.m_pSelectors = &cluster_selectors[0];
+ uint_vec sorted_cluster_indices_new_to_old(total_clusters);
+ indirect_sort(total_clusters, sorted_cluster_indices_new_to_old.data(), cluster_sizes.data());
+ //for (uint32_t i = 0; i < total_clusters; i++)
+ // sorted_cluster_indices_new_to_old[i] = i;
- optimizer.init(cluster_optimizer_params, cluster_optimizer_results);
+ uint_vec sorted_cluster_indices_old_to_new(total_clusters);
+ for (uint32_t i = 0; i < total_clusters; i++)
+ sorted_cluster_indices_old_to_new[sorted_cluster_indices_new_to_old[i]] = i;
- if (!optimizer.compute())
- BASISU_FRONTEND_VERIFY(false);
+ basisu::vector<cl_pixel_cluster> sorted_pixel_clusters(total_clusters);
+ for (uint32_t i = 0; i < total_clusters; i++)
+ sorted_pixel_clusters[i] = pixel_clusters[sorted_cluster_indices_new_to_old[i]];
- new_subblock_params.m_color_unscaled[0] = cluster_optimizer_results.m_block_color_unscaled;
- new_subblock_params.m_inten_table[0] = cluster_optimizer_results.m_block_inten_table;
- new_subblock_params.m_color_error[0] = cluster_optimizer_results.m_error;
- }
+ uint32_t total_perms = 64;
+ if (m_params.m_compression_level <= 1)
+ total_perms = 16;
+ else if (m_params.m_compression_level == BASISU_MAX_COMPRESSION_LEVEL)
+ total_perms = OPENCL_ENCODE_ETC1S_MAX_PERMS;
- endpoint_cluster_etc_params &prev_etc_params = m_endpoint_cluster_etc_params[cluster_index];
+ basisu::vector<etc_block> output_blocks(total_clusters);
- bool use_new_subblock_params = false;
- if ((!step) || (!prev_etc_params.m_valid))
- use_new_subblock_params = true;
- else
+ if (opencl_encode_etc1s_pixel_clusters(
+ m_params.m_pOpenCL_context,
+ output_blocks.data(),
+ total_clusters,
+ sorted_pixel_clusters.data(),
+ input_pixels.size(),
+ input_pixels.data(),
+ pixel_weights.data(),
+ m_params.m_perceptual, total_perms))
+ {
+ for (uint32_t old_cluster_index = 0; old_cluster_index < m_endpoint_clusters.size(); old_cluster_index++)
+ {
+ const uint32_t new_cluster_index = sorted_cluster_indices_old_to_new[old_cluster_index];
+
+ const etc_block& blk = output_blocks[new_cluster_index];
+
+ endpoint_cluster_etc_params& prev_etc_params = m_endpoint_cluster_etc_params[old_cluster_index];
+
+ prev_etc_params.m_valid = true;
+ etc_block::unpack_color5(prev_etc_params.m_color_unscaled[0], blk.get_base5_color(), false);
+ prev_etc_params.m_inten_table[0] = blk.get_inten_table(0);
+ prev_etc_params.m_color_error[0] = 0; // dummy value - we don't actually use this
+ }
+
+ use_cpu = false;
+ }
+ else
+ {
+ error_printf("basisu_frontend::generate_endpoint_codebook: opencl_encode_etc1s_pixel_clusters() failed! Using CPU.\n");
+ m_params.m_pOpenCL_context = nullptr;
+ m_opencl_failed = true;
+ }
+
+ } // if (opencl_is_available() && m_params.m_use_opencl)
+
+ if (use_cpu)
+ {
+ const uint32_t N = 128;
+ for (uint32_t cluster_index_iter = 0; cluster_index_iter < m_endpoint_clusters.size(); cluster_index_iter += N)
+ {
+ const uint32_t first_index = cluster_index_iter;
+ const uint32_t last_index = minimum<uint32_t>((uint32_t)m_endpoint_clusters.size(), cluster_index_iter + N);
+
+#ifndef __EMSCRIPTEN__
+ m_params.m_pJob_pool->add_job([this, first_index, last_index, step] {
+#endif
+
+ for (uint32_t cluster_index = first_index; cluster_index < last_index; cluster_index++)
{
- assert(prev_etc_params.m_valid);
+ const basisu::vector<uint32_t>& cluster_indices = m_endpoint_clusters[cluster_index];
- uint64_t total_prev_err = 0;
-
+ BASISU_FRONTEND_VERIFY(cluster_indices.size());
+
+ const uint32_t total_pixels = (uint32_t)cluster_indices.size() * 8;
+
+ basisu::vector<color_rgba> cluster_pixels(total_pixels);
+
+ for (uint32_t cluster_indices_iter = 0; cluster_indices_iter < cluster_indices.size(); cluster_indices_iter++)
{
- color_rgba block_colors[4];
+ const uint32_t block_index = cluster_indices[cluster_indices_iter] >> 1;
+ const uint32_t subblock_index = cluster_indices[cluster_indices_iter] & 1;
- etc_block::get_block_colors5(block_colors, prev_etc_params.m_color_unscaled[0], prev_etc_params.m_inten_table[0], false);
+ const bool flipped = true;
- uint64_t total_err = 0;
+ const color_rgba* pBlock_pixels = get_source_pixel_block(block_index).get_ptr();
- for (uint32_t i = 0; i < total_pixels; i++)
+ for (uint32_t pixel_index = 0; pixel_index < 8; pixel_index++)
{
- const color_rgba &c = cluster_pixels[i];
+ const color_rgba& c = pBlock_pixels[g_etc1_pixel_indices[flipped][subblock_index][pixel_index]];
+ cluster_pixels[cluster_indices_iter * 8 + pixel_index] = c;
+ }
+ }
- uint64_t best_err = UINT64_MAX;
- //uint32_t best_index = 0;
+ endpoint_cluster_etc_params new_subblock_params;
+
+ {
+ etc1_optimizer optimizer;
+ etc1_solution_coordinates solutions[2];
+
+ etc1_optimizer::params cluster_optimizer_params;
+ cluster_optimizer_params.m_num_src_pixels = total_pixels;
+ cluster_optimizer_params.m_pSrc_pixels = &cluster_pixels[0];
+
+ cluster_optimizer_params.m_use_color4 = false;
+ cluster_optimizer_params.m_perceptual = m_params.m_perceptual;
+
+ if (m_params.m_compression_level <= 1)
+ cluster_optimizer_params.m_quality = cETCQualityMedium;
+ else if (m_params.m_compression_level == BASISU_MAX_COMPRESSION_LEVEL)
+ cluster_optimizer_params.m_quality = cETCQualityUber;
+
+ etc1_optimizer::results cluster_optimizer_results;
+
+ basisu::vector<uint8_t> cluster_selectors(total_pixels);
+ cluster_optimizer_results.m_n = total_pixels;
+ cluster_optimizer_results.m_pSelectors = &cluster_selectors[0];
+
+ optimizer.init(cluster_optimizer_params, cluster_optimizer_results);
+
+ if (!optimizer.compute())
+ BASISU_FRONTEND_VERIFY(false);
+
+ new_subblock_params.m_color_unscaled[0] = cluster_optimizer_results.m_block_color_unscaled;
+ new_subblock_params.m_inten_table[0] = cluster_optimizer_results.m_block_inten_table;
+ new_subblock_params.m_color_error[0] = cluster_optimizer_results.m_error;
+ }
+
+ endpoint_cluster_etc_params& prev_etc_params = m_endpoint_cluster_etc_params[cluster_index];
+
+ bool use_new_subblock_params = false;
+ if ((!step) || (!prev_etc_params.m_valid))
+ use_new_subblock_params = true;
+ else
+ {
+ assert(prev_etc_params.m_valid);
+
+ uint64_t total_prev_err = 0;
+
+ {
+ color_rgba block_colors[4];
+
+ etc_block::get_block_colors5(block_colors, prev_etc_params.m_color_unscaled[0], prev_etc_params.m_inten_table[0], false);
- for (uint32_t s = 0; s < 4; s++)
+ uint64_t total_err = 0;
+
+ for (uint32_t i = 0; i < total_pixels; i++)
{
- uint64_t err = color_distance(m_params.m_perceptual, c, block_colors[s], false);
- if (err < best_err)
+ const color_rgba& c = cluster_pixels[i];
+
+ uint64_t best_err = UINT64_MAX;
+ //uint32_t best_index = 0;
+
+ for (uint32_t s = 0; s < 4; s++)
{
- best_err = err;
- //best_index = s;
+ uint64_t err = color_distance(m_params.m_perceptual, c, block_colors[s], false);
+ if (err < best_err)
+ {
+ best_err = err;
+ //best_index = s;
+ }
}
+
+ total_err += best_err;
}
- total_err += best_err;
+ total_prev_err += total_err;
}
- total_prev_err += total_err;
+ // See if we should update this cluster's endpoints (if the error has actually fallen)
+ if (total_prev_err > new_subblock_params.m_color_error[0])
+ {
+ use_new_subblock_params = true;
+ }
}
- // See if we should update this cluster's endpoints (if the error has actually fallen)
- if (total_prev_err > new_subblock_params.m_color_error[0])
+ if (use_new_subblock_params)
{
- use_new_subblock_params = true;
- }
- }
+ new_subblock_params.m_valid = true;
- if (use_new_subblock_params)
- {
- new_subblock_params.m_valid = true;
+ prev_etc_params = new_subblock_params;
+ }
- prev_etc_params = new_subblock_params;
- }
-
- } // cluster_index
+ } // cluster_index
#ifndef __EMSCRIPTEN__
- } );
+ });
#endif
- } // cluster_index_iter
+ } // cluster_index_iter
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->wait_for_all();
+ m_params.m_pJob_pool->wait_for_all();
#endif
+ }
+
+ debug_printf("Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
}
bool basisu_frontend::check_etc1s_constraints() const
@@ -1373,6 +1684,8 @@ namespace basisu
return true;
}
+ // For each block, determine which ETC1S endpoint cluster can encode that block with lowest error.
+ // This reassigns blocks to different endpoint clusters.
uint32_t basisu_frontend::refine_endpoint_clusterization()
{
debug_printf("refine_endpoint_clusterization\n");
@@ -1380,6 +1693,8 @@ namespace basisu
if (m_use_hierarchical_endpoint_codebooks)
compute_endpoint_clusters_within_each_parent_cluster();
+ // Note: It's possible that an endpoint cluster may live in more than one parent cluster after the first refinement step.
+
basisu::vector<vec2U> block_clusters(m_total_blocks);
for (int cluster_index = 0; cluster_index < static_cast<int>(m_endpoint_clusters.size()); cluster_index++)
@@ -1400,156 +1715,255 @@ namespace basisu
// Create a new endpoint clusterization
+ interval_timer tm;
+ tm.start();
+
uint_vec best_cluster_indices(m_total_blocks);
- const uint32_t N = 1024;
- for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
+ bool use_cpu = true;
+ // TODO: Support non-hierarchical endpoint codebooks here
+ if (m_params.m_pOpenCL_context && m_use_hierarchical_endpoint_codebooks)
{
- const uint32_t first_index = block_index_iter;
- const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
+ // For the OpenCL kernel, we order the parent endpoint clusters by smallest to largest for efficiency.
+ // We also prepare an array of block info structs that point into this new parent endpoint cluster array.
+ const uint32_t total_parent_clusters = m_endpoint_clusters_within_each_parent_cluster.size();
-#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index, &best_cluster_indices, &block_clusters] {
-#endif
+ basisu::vector<cl_block_info_struct> cl_block_info_structs(m_total_blocks);
+
+ // the size of each parent cluster, in total clusters
+ uint_vec parent_cluster_sizes(total_parent_clusters);
+ for (uint32_t i = 0; i < total_parent_clusters; i++)
+ parent_cluster_sizes[i] = m_endpoint_clusters_within_each_parent_cluster[i].size();
+
+ uint_vec first_parent_cluster_ofs(total_parent_clusters);
+ uint32_t cur_ofs = 0;
+ for (uint32_t i = 0; i < total_parent_clusters; i++)
+ {
+ first_parent_cluster_ofs[i] = cur_ofs;
- for (uint32_t block_index = first_index; block_index < last_index; block_index++)
+ cur_ofs += parent_cluster_sizes[i];
+ }
+
+ // Note: total_actual_endpoint_clusters is not necessarly equal to m_endpoint_clusters.size(), because clusters may live in multiple parent clusters after the first refinement step.
+ BASISU_FRONTEND_VERIFY(cur_ofs >= m_endpoint_clusters.size());
+ const uint32_t total_actual_endpoint_clusters = cur_ofs;
+ basisu::vector<cl_endpoint_cluster_struct> cl_endpoint_cluster_structs(total_actual_endpoint_clusters);
+
+ for (uint32_t i = 0; i < total_parent_clusters; i++)
+ {
+ const uint32_t dst_ofs = first_parent_cluster_ofs[i];
+
+ const uint32_t parent_cluster_size = parent_cluster_sizes[i];
+
+ assert(m_endpoint_clusters_within_each_parent_cluster[i].size() == parent_cluster_size);
+
+ for (uint32_t j = 0; j < parent_cluster_size; j++)
{
- const uint32_t cluster_index = block_clusters[block_index][0];
- BASISU_FRONTEND_VERIFY(cluster_index == block_clusters[block_index][1]);
+ const uint32_t endpoint_cluster_index = m_endpoint_clusters_within_each_parent_cluster[i][j];
- const color_rgba *pSubblock_pixels = get_source_pixel_block(block_index).get_ptr();
- const uint32_t num_subblock_pixels = 16;
+ color_rgba cluster_etc_base_color(m_endpoint_cluster_etc_params[endpoint_cluster_index].m_color_unscaled[0]);
+ uint32_t cluster_etc_inten = m_endpoint_cluster_etc_params[endpoint_cluster_index].m_inten_table[0];
- uint64_t best_cluster_err = INT64_MAX;
- uint32_t best_cluster_index = 0;
+ cl_endpoint_cluster_structs[dst_ofs + j].m_unscaled_color = cluster_etc_base_color;
+ cl_endpoint_cluster_structs[dst_ofs + j].m_etc_inten = (uint8_t)cluster_etc_inten;
+ cl_endpoint_cluster_structs[dst_ofs + j].m_cluster_index = (uint16_t)endpoint_cluster_index;
+ }
+ }
+
+ for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
+ {
+ const uint32_t block_parent_endpoint_cluster_index = m_block_parent_endpoint_cluster[block_index];
+
+ cl_block_info_structs[block_index].m_num_clusters = (uint16_t)(parent_cluster_sizes[block_parent_endpoint_cluster_index]);
+ cl_block_info_structs[block_index].m_first_cluster_ofs = (uint16_t)(first_parent_cluster_ofs[block_parent_endpoint_cluster_index]);
- const uint32_t block_parent_endpoint_cluster_index = m_block_parent_endpoint_cluster.size() ? m_block_parent_endpoint_cluster[block_index] : 0;
- const uint_vec *pCluster_indices = m_endpoint_clusters_within_each_parent_cluster.size() ? &m_endpoint_clusters_within_each_parent_cluster[block_parent_endpoint_cluster_index] : nullptr;
+ const uint32_t block_cluster_index = block_clusters[block_index][0];
+ cl_block_info_structs[block_index].m_cur_cluster_index = (uint16_t)block_cluster_index;
+ cl_block_info_structs[block_index].m_cur_cluster_etc_inten = (uint8_t)m_endpoint_cluster_etc_params[block_cluster_index].m_inten_table[0];
+ }
+
+ uint_vec block_cluster_indices(m_total_blocks);
+ for (uint32_t i = 0; i < m_total_blocks; i++)
+ block_cluster_indices[i] = block_clusters[i][0];
- const uint32_t total_clusters = m_use_hierarchical_endpoint_codebooks ? (uint32_t)pCluster_indices->size() : (uint32_t)m_endpoint_clusters.size();
+ uint_vec sorted_block_indices(m_total_blocks);
+ indirect_sort(m_total_blocks, sorted_block_indices.data(), block_cluster_indices.data());
- for (uint32_t i = 0; i < total_clusters; i++)
+ bool status = opencl_refine_endpoint_clusterization(
+ m_params.m_pOpenCL_context,
+ cl_block_info_structs.data(),
+ total_actual_endpoint_clusters,
+ cl_endpoint_cluster_structs.data(),
+ sorted_block_indices.data(),
+ best_cluster_indices.data(),
+ m_params.m_perceptual);
+
+ if (status)
+ {
+ use_cpu = false;
+ }
+ else
+ {
+ error_printf("basisu_frontend::refine_endpoint_clusterization: opencl_refine_endpoint_clusterization() failed! Using CPU.\n");
+ m_params.m_pOpenCL_context = nullptr;
+ m_opencl_failed = true;
+ }
+ }
+
+ if (use_cpu)
+ {
+ const uint32_t N = 1024;
+ for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
+ {
+ const uint32_t first_index = block_index_iter;
+ const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
+
+#ifndef __EMSCRIPTEN__
+ m_params.m_pJob_pool->add_job([this, first_index, last_index, &best_cluster_indices, &block_clusters] {
+#endif
+
+ for (uint32_t block_index = first_index; block_index < last_index; block_index++)
{
- const uint32_t cluster_iter = m_use_hierarchical_endpoint_codebooks ? (*pCluster_indices)[i] : i;
+ const uint32_t cluster_index = block_clusters[block_index][0];
+ BASISU_FRONTEND_VERIFY(cluster_index == block_clusters[block_index][1]);
- color_rgba cluster_etc_base_color(m_endpoint_cluster_etc_params[cluster_iter].m_color_unscaled[0]);
- uint32_t cluster_etc_inten = m_endpoint_cluster_etc_params[cluster_iter].m_inten_table[0];
+ const color_rgba* pSubblock_pixels = get_source_pixel_block(block_index).get_ptr();
+ const uint32_t num_subblock_pixels = 16;
- uint64_t total_err = 0;
+ uint64_t best_cluster_err = INT64_MAX;
+ uint32_t best_cluster_index = 0;
- const uint32_t low_selector = 0;//subblock_etc_params_vec[j].m_low_selectors[0];
- const uint32_t high_selector = 3;//subblock_etc_params_vec[j].m_high_selectors[0];
- color_rgba subblock_colors[4];
- // Can't assign it here - may result in too much error when selector quant occurs
- if (cluster_etc_inten > m_endpoint_cluster_etc_params[cluster_index].m_inten_table[0])
- {
- total_err = INT64_MAX;
- goto skip_cluster;
- }
+ const uint32_t block_parent_endpoint_cluster_index = m_block_parent_endpoint_cluster.size() ? m_block_parent_endpoint_cluster[block_index] : 0;
+ const uint_vec* pCluster_indices = m_endpoint_clusters_within_each_parent_cluster.size() ? &m_endpoint_clusters_within_each_parent_cluster[block_parent_endpoint_cluster_index] : nullptr;
- etc_block::get_block_colors5(subblock_colors, cluster_etc_base_color, cluster_etc_inten);
-
-#if 0
- for (uint32_t p = 0; p < num_subblock_pixels; p++)
+ const uint32_t total_clusters = m_use_hierarchical_endpoint_codebooks ? (uint32_t)pCluster_indices->size() : (uint32_t)m_endpoint_clusters.size();
+
+ for (uint32_t i = 0; i < total_clusters; i++)
{
- uint64_t best_err = UINT64_MAX;
+ const uint32_t cluster_iter = m_use_hierarchical_endpoint_codebooks ? (*pCluster_indices)[i] : i;
- for (uint32_t r = low_selector; r <= high_selector; r++)
+ color_rgba cluster_etc_base_color(m_endpoint_cluster_etc_params[cluster_iter].m_color_unscaled[0]);
+ uint32_t cluster_etc_inten = m_endpoint_cluster_etc_params[cluster_iter].m_inten_table[0];
+
+ uint64_t total_err = 0;
+
+ const uint32_t low_selector = 0;//subblock_etc_params_vec[j].m_low_selectors[0];
+ const uint32_t high_selector = 3;//subblock_etc_params_vec[j].m_high_selectors[0];
+ color_rgba subblock_colors[4];
+ // Can't assign it here - may result in too much error when selector quant occurs
+ if (cluster_etc_inten > m_endpoint_cluster_etc_params[cluster_index].m_inten_table[0])
{
- uint64_t err = color_distance(m_params.m_perceptual, pSubblock_pixels[p], subblock_colors[r], false);
- best_err = minimum(best_err, err);
- if (!best_err)
- break;
+ total_err = INT64_MAX;
+ goto skip_cluster;
}
- total_err += best_err;
- if (total_err > best_cluster_err)
- break;
- } // p
-#else
- if (m_params.m_perceptual)
- {
- if (!g_cpu_supports_sse41)
+ etc_block::get_block_colors5(subblock_colors, cluster_etc_base_color, cluster_etc_inten);
+
+#if 0
+ for (uint32_t p = 0; p < num_subblock_pixels; p++)
{
- for (uint32_t p = 0; p < num_subblock_pixels; p++)
+ uint64_t best_err = UINT64_MAX;
+
+ for (uint32_t r = low_selector; r <= high_selector; r++)
{
- uint64_t best_err = UINT64_MAX;
+ uint64_t err = color_distance(m_params.m_perceptual, pSubblock_pixels[p], subblock_colors[r], false);
+ best_err = minimum(best_err, err);
+ if (!best_err)
+ break;
+ }
- for (uint32_t r = low_selector; r <= high_selector; r++)
+ total_err += best_err;
+ if (total_err > best_cluster_err)
+ break;
+ } // p
+#else
+ if (m_params.m_perceptual)
+ {
+ if (!g_cpu_supports_sse41)
+ {
+ for (uint32_t p = 0; p < num_subblock_pixels; p++)
{
- uint64_t err = color_distance(true, pSubblock_pixels[p], subblock_colors[r], false);
- best_err = minimum(best_err, err);
- if (!best_err)
+ uint64_t best_err = UINT64_MAX;
+
+ for (uint32_t r = low_selector; r <= high_selector; r++)
+ {
+ uint64_t err = color_distance(true, pSubblock_pixels[p], subblock_colors[r], false);
+ best_err = minimum(best_err, err);
+ if (!best_err)
+ break;
+ }
+
+ total_err += best_err;
+ if (total_err > best_cluster_err)
break;
- }
-
- total_err += best_err;
- if (total_err > best_cluster_err)
- break;
- } // p
- }
- else
- {
+ } // p
+ }
+ else
+ {
#if BASISU_SUPPORT_SSE
- find_lowest_error_perceptual_rgb_4_N_sse41((int64_t*)&total_err, subblock_colors, pSubblock_pixels, num_subblock_pixels, best_cluster_err);
+ find_lowest_error_perceptual_rgb_4_N_sse41((int64_t*)&total_err, subblock_colors, pSubblock_pixels, num_subblock_pixels, best_cluster_err);
#endif
+ }
}
- }
- else
- {
- if (!g_cpu_supports_sse41)
+ else
{
- for (uint32_t p = 0; p < num_subblock_pixels; p++)
+ if (!g_cpu_supports_sse41)
{
- uint64_t best_err = UINT64_MAX;
-
- for (uint32_t r = low_selector; r <= high_selector; r++)
+ for (uint32_t p = 0; p < num_subblock_pixels; p++)
{
- uint64_t err = color_distance(false, pSubblock_pixels[p], subblock_colors[r], false);
- best_err = minimum(best_err, err);
- if (!best_err)
+ uint64_t best_err = UINT64_MAX;
+
+ for (uint32_t r = low_selector; r <= high_selector; r++)
+ {
+ uint64_t err = color_distance(false, pSubblock_pixels[p], subblock_colors[r], false);
+ best_err = minimum(best_err, err);
+ if (!best_err)
+ break;
+ }
+
+ total_err += best_err;
+ if (total_err > best_cluster_err)
break;
- }
-
- total_err += best_err;
- if (total_err > best_cluster_err)
- break;
- } // p
- }
- else
- {
+ } // p
+ }
+ else
+ {
#if BASISU_SUPPORT_SSE
- find_lowest_error_linear_rgb_4_N_sse41((int64_t*)&total_err, subblock_colors, pSubblock_pixels, num_subblock_pixels, best_cluster_err);
+ find_lowest_error_linear_rgb_4_N_sse41((int64_t*)&total_err, subblock_colors, pSubblock_pixels, num_subblock_pixels, best_cluster_err);
#endif
+ }
}
- }
#endif
- skip_cluster:
- if ((total_err < best_cluster_err) ||
- ((cluster_iter == cluster_index) && (total_err == best_cluster_err)))
- {
- best_cluster_err = total_err;
- best_cluster_index = cluster_iter;
-
- if (!best_cluster_err)
- break;
- }
- } // j
-
- best_cluster_indices[block_index] = best_cluster_index;
+ skip_cluster:
+ if ((total_err < best_cluster_err) ||
+ ((cluster_iter == cluster_index) && (total_err == best_cluster_err)))
+ {
+ best_cluster_err = total_err;
+ best_cluster_index = cluster_iter;
- } // block_index
+ if (!best_cluster_err)
+ break;
+ }
+ } // j
+
+ best_cluster_indices[block_index] = best_cluster_index;
+
+ } // block_index
#ifndef __EMSCRIPTEN__
- } );
+ });
#endif
-
- } // block_index_iter
+
+ } // block_index_iter
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->wait_for_all();
+ m_params.m_pJob_pool->wait_for_all();
#endif
+
+ } // use_cpu
+
+ debug_printf("refine_endpoint_clusterization time: %3.3f secs\n", tm.get_elapsed_secs());
basisu::vector<typename basisu::vector<uint32_t> > optimized_endpoint_clusters(m_endpoint_clusters.size());
uint32_t total_subblocks_reassigned = 0;
@@ -1647,51 +2061,91 @@ namespace basisu
void basisu_frontend::create_initial_packed_texture()
{
debug_printf("create_initial_packed_texture\n");
+
+ interval_timer tm;
+ tm.start();
- const uint32_t N = 4096;
- for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
+ bool use_cpu = true;
+
+ if ((m_params.m_pOpenCL_context) && (opencl_is_available()))
{
- const uint32_t first_index = block_index_iter;
- const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
+ basisu::vector<color_rgba> block_etc5_color_intens(m_total_blocks);
+
+ for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
+ {
+ uint32_t cluster0 = m_block_endpoint_clusters_indices[block_index][0];
+
+ const color_rgba& color_unscaled = m_endpoint_cluster_etc_params[cluster0].m_color_unscaled[0];
+ uint32_t inten = m_endpoint_cluster_etc_params[cluster0].m_inten_table[0];
+
+ block_etc5_color_intens[block_index].set(color_unscaled.r, color_unscaled.g, color_unscaled.b, inten);
+ }
+
+ bool status = opencl_determine_selectors(m_params.m_pOpenCL_context, block_etc5_color_intens.data(),
+ m_encoded_blocks.data(),
+ m_params.m_perceptual);
+ if (!status)
+ {
+ error_printf("basisu_frontend::create_initial_packed_texture: opencl_determine_selectors() failed! Using CPU.\n");
+ m_params.m_pOpenCL_context = nullptr;
+ m_opencl_failed = true;
+ }
+ else
+ {
+ use_cpu = false;
+ }
+ }
+
+ if (use_cpu)
+ {
+ const uint32_t N = 4096;
+ for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
+ {
+ const uint32_t first_index = block_index_iter;
+ const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index] {
+ m_params.m_pJob_pool->add_job([this, first_index, last_index] {
#endif
-
- for (uint32_t block_index = first_index; block_index < last_index; block_index++)
- {
- uint32_t cluster0 = m_block_endpoint_clusters_indices[block_index][0];
- uint32_t cluster1 = m_block_endpoint_clusters_indices[block_index][1];
- BASISU_FRONTEND_VERIFY(cluster0 == cluster1);
- const color_rgba *pSource_pixels = get_source_pixel_block(block_index).get_ptr();
+ for (uint32_t block_index = first_index; block_index < last_index; block_index++)
+ {
+ uint32_t cluster0 = m_block_endpoint_clusters_indices[block_index][0];
+ uint32_t cluster1 = m_block_endpoint_clusters_indices[block_index][1];
+ BASISU_FRONTEND_VERIFY(cluster0 == cluster1);
- etc_block &blk = m_encoded_blocks[block_index];
+ const color_rgba* pSource_pixels = get_source_pixel_block(block_index).get_ptr();
- color_rgba unscaled[2] = { m_endpoint_cluster_etc_params[cluster0].m_color_unscaled[0], m_endpoint_cluster_etc_params[cluster1].m_color_unscaled[0] };
- uint32_t inten[2] = { m_endpoint_cluster_etc_params[cluster0].m_inten_table[0], m_endpoint_cluster_etc_params[cluster1].m_inten_table[0] };
-
- blk.set_block_color5(unscaled[0], unscaled[1]);
- blk.set_flip_bit(true);
+ etc_block& blk = m_encoded_blocks[block_index];
- blk.set_inten_table(0, inten[0]);
- blk.set_inten_table(1, inten[1]);
+ color_rgba unscaled[2] = { m_endpoint_cluster_etc_params[cluster0].m_color_unscaled[0], m_endpoint_cluster_etc_params[cluster1].m_color_unscaled[0] };
+ uint32_t inten[2] = { m_endpoint_cluster_etc_params[cluster0].m_inten_table[0], m_endpoint_cluster_etc_params[cluster1].m_inten_table[0] };
- blk.determine_selectors(pSource_pixels, m_params.m_perceptual);
-
- } // block_index
+ blk.set_block_color5(unscaled[0], unscaled[1]);
+ blk.set_flip_bit(true);
+
+ blk.set_inten_table(0, inten[0]);
+ blk.set_inten_table(1, inten[1]);
+
+ blk.determine_selectors(pSource_pixels, m_params.m_perceptual);
+
+ } // block_index
#ifndef __EMSCRIPTEN__
- } );
+ });
#endif
- } // block_index_iter
+ } // block_index_iter
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->wait_for_all();
+ m_params.m_pJob_pool->wait_for_all();
#endif
+ } // use_cpu
+
m_orig_encoded_blocks = m_encoded_blocks;
+
+ debug_printf("Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
}
void basisu_frontend::compute_selector_clusters_within_each_parent_cluster()
@@ -1739,8 +2193,7 @@ namespace basisu
void basisu_frontend::generate_selector_clusters()
{
debug_printf("generate_selector_clusters\n");
-
- typedef vec<16, float> vec16F;
+
typedef tree_vector_quant<vec16F> vec16F_clusterizer;
vec16F_clusterizer::array_of_weighted_training_vecs training_vecs(m_total_blocks);
@@ -1800,12 +2253,14 @@ namespace basisu
uint32_t max_threads = 0;
max_threads = m_params.m_multithreaded ? minimum<int>(std::thread::hardware_concurrency(), cMaxCodebookCreationThreads) : 0;
+ if (m_params.m_pJob_pool)
+ max_threads = minimum<int>((int)m_params.m_pJob_pool->get_total_threads(), max_threads);
bool status = generate_hierarchical_codebook_threaded(selector_clusterizer,
m_params.m_max_selector_clusters, m_use_hierarchical_selector_codebooks ? parent_codebook_size : 0,
m_selector_cluster_block_indices,
m_selector_parent_cluster_block_indices,
- max_threads, m_params.m_pJob_pool);
+ max_threads, m_params.m_pJob_pool, false);
BASISU_FRONTEND_VERIFY(status);
if (m_use_hierarchical_selector_codebooks)
@@ -1864,235 +2319,105 @@ namespace basisu
{
debug_printf("create_optimized_selector_codebook\n");
+ interval_timer tm;
+ tm.start();
+
const uint32_t total_selector_clusters = (uint32_t)m_selector_cluster_block_indices.size();
debug_printf("Total selector clusters (from m_selector_cluster_block_indices.size()): %u\n", (uint32_t)m_selector_cluster_block_indices.size());
m_optimized_cluster_selectors.resize(total_selector_clusters);
+
+ uint32_t total_clusters_processed = 0;
- if ((m_params.m_pGlobal_sel_codebook) && (!m_params.m_use_hybrid_selector_codebooks))
+ // For each selector codebook entry, and for each of the 4x4 selectors, determine which selector minimizes the error across all the blocks that use that quantized selector.
+ const uint32_t N = 256;
+ for (uint32_t cluster_index_iter = 0; cluster_index_iter < total_selector_clusters; cluster_index_iter += N)
{
- uint32_t total_clusters_processed = 0;
-
- m_optimized_cluster_selector_global_cb_ids.resize(total_selector_clusters);
+ const uint32_t first_index = cluster_index_iter;
+ const uint32_t last_index = minimum<uint32_t>((uint32_t)total_selector_clusters, cluster_index_iter + N);
- const uint32_t N = 256;
- for (uint32_t cluster_index_iter = 0; cluster_index_iter < total_selector_clusters; cluster_index_iter += N)
- {
- const uint32_t first_index = cluster_index_iter;
- const uint32_t last_index = minimum<uint32_t>((uint32_t)total_selector_clusters, cluster_index_iter + N);
-
-#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index, &total_clusters_processed, &total_selector_clusters] {
-#endif
-
- for (uint32_t cluster_index = first_index; cluster_index < last_index; cluster_index++)
- {
- const basisu::vector<uint32_t> &cluster_block_indices = m_selector_cluster_block_indices[cluster_index];
-
- if (!cluster_block_indices.size())
- continue;
-
- etc_block_vec etc_blocks;
- pixel_block_vec pixel_blocks;
-
- for (uint32_t cluster_block_index = 0; cluster_block_index < cluster_block_indices.size(); cluster_block_index++)
- {
- const uint32_t block_index = cluster_block_indices[cluster_block_index];
-
- etc_blocks.push_back(m_encoded_blocks[block_index]);
-
- pixel_blocks.push_back(get_source_pixel_block(block_index));
- }
-
- uint32_t palette_index;
- basist::etc1_global_palette_entry_modifier palette_modifier;
-
- #if 0
- m_params.m_pGlobal_sel_codebook->find_best_entry(etc_blocks.size(), pixel_blocks.get_ptr(), etc_blocks.get_ptr(),
- palette_index, palette_modifier,
- m_params.m_perceptual, 1 << m_params.m_num_global_sel_codebook_pal_bits, 1 << m_params.m_num_global_sel_codebook_mod_bits);
- #else
- etc1_global_selector_codebook_find_best_entry(*m_params.m_pGlobal_sel_codebook,
- (uint32_t)etc_blocks.size(), &pixel_blocks[0], &etc_blocks[0],
- palette_index, palette_modifier,
- m_params.m_perceptual, 1 << m_params.m_num_global_sel_codebook_pal_bits, 1 << m_params.m_num_global_sel_codebook_mod_bits);
- #endif
-
- m_optimized_cluster_selector_global_cb_ids[cluster_index].set(palette_index, palette_modifier);
-
- basist::etc1_selector_palette_entry pal_entry(m_params.m_pGlobal_sel_codebook->get_entry(palette_index, palette_modifier));
-
- for (uint32_t y = 0; y < 4; y++)
- for (uint32_t x = 0; x < 4; x++)
- m_optimized_cluster_selectors[cluster_index].set_selector(x, y, pal_entry(x, y));
-
- {
- std::lock_guard<std::mutex> lock(m_lock);
-
- total_clusters_processed++;
- if ((total_clusters_processed % 63) == 0)
- debug_printf("Global selector palette optimization: %3.1f%% complete\n", total_clusters_processed * 100.0f / total_selector_clusters);
- }
-
- } // cluster_index
-
-#ifndef __EMSCRIPTEN__
- } );
+#ifndef __EMSCRIPTEN__
+ m_params.m_pJob_pool->add_job([this, first_index, last_index, &total_clusters_processed, &total_selector_clusters] {
#endif
- } // cluster_index_iter
-
-#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->wait_for_all();
-#endif
- }
- else
- {
- const bool uses_hybrid_sel_codebook = ((m_params.m_pGlobal_sel_codebook) && (m_params.m_use_hybrid_selector_codebooks));
- if (uses_hybrid_sel_codebook)
- {
- m_selector_cluster_uses_global_cb.resize(total_selector_clusters);
- m_optimized_cluster_selector_global_cb_ids.resize(total_selector_clusters);
- }
+ for (uint32_t cluster_index = first_index; cluster_index < last_index; cluster_index++)
+ {
+ const basisu::vector<uint32_t>& cluster_block_indices = m_selector_cluster_block_indices[cluster_index];
- uint32_t total_clusters_processed = 0;
+ if (!cluster_block_indices.size())
+ continue;
- // For each selector codebook entry, and for each of the 4x4 selectors, determine which selector minimizes the error across all the blocks that use that quantized selector.
+ uint64_t overall_best_err = 0;
- const uint32_t N = 256;
- for (uint32_t cluster_index_iter = 0; cluster_index_iter < total_selector_clusters; cluster_index_iter += N)
- {
- const uint32_t first_index = cluster_index_iter;
- const uint32_t last_index = minimum<uint32_t>((uint32_t)total_selector_clusters, cluster_index_iter + N);
+ uint64_t total_err[4][4][4];
+ clear_obj(total_err);
-#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index, &uses_hybrid_sel_codebook, &total_clusters_processed, &total_selector_clusters] {
-#endif
-
- for (uint32_t cluster_index = first_index; cluster_index < last_index; cluster_index++)
+ for (uint32_t cluster_block_index = 0; cluster_block_index < cluster_block_indices.size(); cluster_block_index++)
{
- const basisu::vector<uint32_t> &cluster_block_indices = m_selector_cluster_block_indices[cluster_index];
+ const uint32_t block_index = cluster_block_indices[cluster_block_index];
- if (!cluster_block_indices.size())
- continue;
+ const etc_block& blk = m_encoded_blocks[block_index];
- uint64_t overall_best_err = 0;
+ color_rgba blk_colors[4];
+ blk.get_block_colors(blk_colors, 0);
for (uint32_t y = 0; y < 4; y++)
{
for (uint32_t x = 0; x < 4; x++)
{
- uint64_t best_err = UINT64_MAX;
- uint32_t best_s = 0;
+ const color_rgba& orig_color = get_source_pixel_block(block_index)(x, y);
- for (uint32_t s = 0; s < 4; s++)
+ if (m_params.m_perceptual)
{
- uint32_t total_err = 0;
-
- for (uint32_t cluster_block_index = 0; cluster_block_index < cluster_block_indices.size(); cluster_block_index++)
- {
- const uint32_t block_index = cluster_block_indices[cluster_block_index];
-
- const etc_block &blk = m_encoded_blocks[block_index];
-
- const color_rgba &orig_color = get_source_pixel_block(block_index)(x, y);
-
- color_rgba block_color;
- blk.get_block_color(block_color, blk.get_subblock_index(x, y), s);
- total_err += color_distance(m_params.m_perceptual, block_color, orig_color, false);
-
- if (total_err > best_err)
- break;
-
- } // block_index
-
- if (total_err < best_err)
- {
- best_err = total_err;
- best_s = s;
- if (!best_err)
- break;
- }
-
- } // s
-
- m_optimized_cluster_selectors[cluster_index].set_selector(x, y, best_s);
-
- overall_best_err += best_err;
-
+ for (uint32_t s = 0; s < 4; s++)
+ total_err[y][x][s] += color_distance(true, blk_colors[s], orig_color, false);
+ }
+ else
+ {
+ for (uint32_t s = 0; s < 4; s++)
+ total_err[y][x][s] += color_distance(false, blk_colors[s], orig_color, false);
+ }
} // x
} // y
- if (uses_hybrid_sel_codebook)
- {
- etc_block_vec etc_blocks;
- pixel_block_vec pixel_blocks;
+ } // cluster_block_index
- for (uint32_t cluster_block_index = 0; cluster_block_index < cluster_block_indices.size(); cluster_block_index++)
- {
- const uint32_t block_index = cluster_block_indices[cluster_block_index];
-
- etc_blocks.push_back(m_encoded_blocks[block_index]);
-
- pixel_blocks.push_back(get_source_pixel_block(block_index));
- }
-
- uint32_t palette_index;
- basist::etc1_global_palette_entry_modifier palette_modifier;
-
- #if 0
- uint64_t best_global_cb_err = m_params.m_pGlobal_sel_codebook->find_best_entry(etc_blocks.size(), pixel_blocks.get_ptr(), etc_blocks.get_ptr(),
- palette_index, palette_modifier,
- m_params.m_perceptual, 1 << m_params.m_num_global_sel_codebook_pal_bits, 1 << m_params.m_num_global_sel_codebook_mod_bits);
- #else
- uint64_t best_global_cb_err = etc1_global_selector_codebook_find_best_entry(*m_params.m_pGlobal_sel_codebook, (uint32_t)etc_blocks.size(), &pixel_blocks[0], &etc_blocks[0],
- palette_index, palette_modifier,
- m_params.m_perceptual, 1 << m_params.m_num_global_sel_codebook_pal_bits, 1 << m_params.m_num_global_sel_codebook_mod_bits);
- #endif
+ for (uint32_t y = 0; y < 4; y++)
+ {
+ for (uint32_t x = 0; x < 4; x++)
+ {
+ uint64_t best_err = total_err[y][x][0];
+ uint8_t best_sel = 0;
- if (best_global_cb_err <= overall_best_err * m_params.m_hybrid_codebook_quality_thresh)
+ for (uint32_t s = 1; s < 4; s++)
{
- m_selector_cluster_uses_global_cb[cluster_index] = true;
-
- m_optimized_cluster_selector_global_cb_ids[cluster_index].set(palette_index, palette_modifier);
-
- basist::etc1_selector_palette_entry pal_entry(m_params.m_pGlobal_sel_codebook->get_entry(palette_index, palette_modifier));
-
- for (uint32_t y = 0; y < 4; y++)
- for (uint32_t x = 0; x < 4; x++)
- m_optimized_cluster_selectors[cluster_index].set_selector(x, y, pal_entry(x, y));
+ if (total_err[y][x][s] < best_err)
+ {
+ best_err = total_err[y][x][s];
+ best_sel = (uint8_t)s;
+ }
}
- else
- {
- m_optimized_cluster_selector_global_cb_ids[cluster_index].set(0, basist::etc1_global_palette_entry_modifier(0));
- m_selector_cluster_uses_global_cb[cluster_index] = false;
- }
- }
+ m_optimized_cluster_selectors[cluster_index].set_selector(x, y, best_sel);
- if (uses_hybrid_sel_codebook)
- {
- std::lock_guard<std::mutex> lock(m_lock);
-
- total_clusters_processed++;
- if ((total_clusters_processed % 63) == 0)
- debug_printf("Global selector palette optimization: %3.1f%% complete\n", total_clusters_processed * 100.0f / total_selector_clusters);
- }
+ overall_best_err += best_err;
+ } // x
+ } // y
- } // cluster_index
+ } // cluster_index
#ifndef __EMSCRIPTEN__
- } );
+ });
#endif
- } // cluster_index_iter
+ } // cluster_index_iter
#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->wait_for_all();
+ m_params.m_pJob_pool->wait_for_all();
#endif
- } // if (m_params.m_pGlobal_sel_codebook)
+ debug_printf("Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
if (m_params.m_debug_images)
{
@@ -2133,17 +2458,25 @@ namespace basisu
}
}
+ // For each block: Determine which quantized selectors best encode that block, given its quantized endpoints.
+ // Note that this method may leave some empty clusters (i.e. arrays with no block indices), including at the end.
void basisu_frontend::find_optimal_selector_clusters_for_each_block()
{
debug_printf("find_optimal_selector_clusters_for_each_block\n");
- // Sanity checks
- BASISU_FRONTEND_VERIFY(m_selector_cluster_block_indices.size() == m_optimized_cluster_selectors.size());
- for (uint32_t i = 0; i < m_selector_clusters_within_each_parent_cluster.size(); i++)
+ interval_timer tm;
+ tm.start();
+
+ if (m_params.m_validate)
{
- for (uint32_t j = 0; j < m_selector_clusters_within_each_parent_cluster[i].size(); j++)
+ // Sanity checks
+ BASISU_FRONTEND_VERIFY(m_selector_cluster_block_indices.size() == m_optimized_cluster_selectors.size());
+ for (uint32_t i = 0; i < m_selector_clusters_within_each_parent_cluster.size(); i++)
{
- BASISU_FRONTEND_VERIFY(m_selector_clusters_within_each_parent_cluster[i][j] < m_optimized_cluster_selectors.size());
+ for (uint32_t j = 0; j < m_selector_clusters_within_each_parent_cluster[i].size(); j++)
+ {
+ BASISU_FRONTEND_VERIFY(m_selector_clusters_within_each_parent_cluster[i][j] < m_optimized_cluster_selectors.size());
+ }
}
}
@@ -2151,20 +2484,120 @@ namespace basisu
if (m_params.m_compression_level == 0)
{
- // Don't do anything, just leave the blocks in their original selector clusters.
- for (uint32_t i = 0; i < m_selector_cluster_block_indices.size(); i++)
+ // Just leave the blocks in their original selector clusters.
+ for (uint32_t selector_cluster_index = 0; selector_cluster_index < m_selector_cluster_block_indices.size(); selector_cluster_index++)
{
- for (uint32_t j = 0; j < m_selector_cluster_block_indices[i].size(); j++)
- m_block_selector_cluster_index[m_selector_cluster_block_indices[i][j]] = i;
+ for (uint32_t j = 0; j < m_selector_cluster_block_indices[selector_cluster_index].size(); j++)
+ {
+ const uint32_t block_index = m_selector_cluster_block_indices[selector_cluster_index][j];
+
+ m_block_selector_cluster_index[block_index] = selector_cluster_index;
+
+ etc_block& blk = m_encoded_blocks[block_index];
+ blk.set_raw_selector_bits(m_optimized_cluster_selectors[selector_cluster_index].get_raw_selector_bits());
+ }
}
+
+ debug_printf("Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
+
+ return;
}
- else
+
+ bool use_cpu = true;
+
+ if ((m_params.m_pOpenCL_context) && m_use_hierarchical_selector_codebooks)
{
- // Note that this method may leave some empty clusters (i.e. arrays with no block indices), including at the end.
- basisu::vector< basisu::vector<uint32_t> > new_cluster_indices(m_optimized_cluster_selectors.size());
+ const uint32_t num_parent_clusters = m_selector_clusters_within_each_parent_cluster.size();
+
+ basisu::vector<fosc_selector_struct> selector_structs;
+ selector_structs.reserve(m_optimized_cluster_selectors.size());
+
+ uint_vec parent_selector_cluster_offsets(num_parent_clusters);
+
+ uint_vec selector_cluster_indices;
+ selector_cluster_indices.reserve(m_optimized_cluster_selectors.size());
+
+ uint32_t cur_ofs = 0;
+ for (uint32_t parent_index = 0; parent_index < num_parent_clusters; parent_index++)
+ {
+ parent_selector_cluster_offsets[parent_index] = cur_ofs;
+
+ for (uint32_t j = 0; j < m_selector_clusters_within_each_parent_cluster[parent_index].size(); j++)
+ {
+ const uint32_t selector_cluster_index = m_selector_clusters_within_each_parent_cluster[parent_index][j];
+
+ uint32_t sel_bits = 0;
+ for (uint32_t p = 0; p < 16; p++)
+ sel_bits |= (m_optimized_cluster_selectors[selector_cluster_index].get_selector(p & 3, p >> 2) << (p * 2));
+
+ selector_structs.enlarge(1)->m_packed_selectors = sel_bits;
+
+ selector_cluster_indices.push_back(selector_cluster_index);
+ }
+
+ cur_ofs += m_selector_clusters_within_each_parent_cluster[parent_index].size();
+ }
+
+ const uint32_t total_input_selectors = cur_ofs;
- // For each block: Determine which quantized selectors best encode that block, given its quantized endpoints.
+ basisu::vector<fosc_block_struct> block_structs(m_total_blocks);
+ for (uint32_t i = 0; i < m_total_blocks; i++)
+ {
+ const uint32_t parent_selector_cluster = m_block_parent_selector_cluster[i];
+
+ const etc_block& blk = m_encoded_blocks[i];
+ blk.unpack_color5(block_structs[i].m_etc_color5_inten, blk.get_base5_color(), false);
+
+ block_structs[i].m_etc_color5_inten.a = (uint8_t)blk.get_inten_table(0);
+ block_structs[i].m_first_selector = parent_selector_cluster_offsets[parent_selector_cluster];
+ block_structs[i].m_num_selectors = m_selector_clusters_within_each_parent_cluster[parent_selector_cluster].size();
+ }
+
+ uint_vec output_selector_cluster_indices(m_total_blocks);
+
+ bool status = opencl_find_optimal_selector_clusters_for_each_block(
+ m_params.m_pOpenCL_context,
+ block_structs.data(),
+ total_input_selectors,
+ selector_structs.data(),
+ selector_cluster_indices.data(),
+ output_selector_cluster_indices.data(),
+ m_params.m_perceptual);
+
+ if (!status)
+ {
+ error_printf("basisu_frontend::find_optimal_selector_clusters_for_each_block: opencl_find_optimal_selector_clusters_for_each_block() failed! Using CPU.\n");
+ m_params.m_pOpenCL_context = nullptr;
+ m_opencl_failed = true;
+ }
+ else
+ {
+ for (uint32_t i = 0; i < m_selector_cluster_block_indices.size(); i++)
+ {
+ m_selector_cluster_block_indices[i].resize(0);
+ m_selector_cluster_block_indices[i].reserve(128);
+ }
+
+ for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
+ {
+ etc_block& blk = m_encoded_blocks[block_index];
+
+ uint32_t best_cluster_index = output_selector_cluster_indices[block_index];
+
+ blk.set_raw_selector_bits(m_optimized_cluster_selectors[best_cluster_index].get_raw_selector_bits());
+ m_block_selector_cluster_index[block_index] = best_cluster_index;
+
+ vector_ensure_element_is_valid(m_selector_cluster_block_indices, best_cluster_index);
+ m_selector_cluster_block_indices[best_cluster_index].push_back(block_index);
+ }
+
+ use_cpu = false;
+ }
+ }
+
+ if (use_cpu)
+ {
basisu::vector<uint8_t> unpacked_optimized_cluster_selectors(16 * m_optimized_cluster_selectors.size());
for (uint32_t cluster_index = 0; cluster_index < m_optimized_cluster_selectors.size(); cluster_index++)
{
@@ -2176,36 +2609,74 @@ namespace basisu
}
}
}
-
- const uint32_t N = 1024;
+
+ const uint32_t N = 2048;
for (uint32_t block_index_iter = 0; block_index_iter < m_total_blocks; block_index_iter += N)
{
const uint32_t first_index = block_index_iter;
const uint32_t last_index = minimum<uint32_t>(m_total_blocks, first_index + N);
-#ifndef __EMSCRIPTEN__
- m_params.m_pJob_pool->add_job( [this, first_index, last_index, &new_cluster_indices, &unpacked_optimized_cluster_selectors] {
-#endif
+ #ifndef __EMSCRIPTEN__
+ m_params.m_pJob_pool->add_job( [this, first_index, last_index, &unpacked_optimized_cluster_selectors] {
+ #endif
+
+ int prev_best_cluster_index = 0;
for (uint32_t block_index = first_index; block_index < last_index; block_index++)
{
- const color_rgba* pBlock_pixels = get_source_pixel_block(block_index).get_ptr();
-
+ const pixel_block& block = get_source_pixel_block(block_index);
+
etc_block& blk = m_encoded_blocks[block_index];
-
+
+ if ((block_index > first_index) && (block == get_source_pixel_block(block_index - 1)))
+ {
+ blk.set_raw_selector_bits(m_optimized_cluster_selectors[prev_best_cluster_index].get_raw_selector_bits());
+
+ m_block_selector_cluster_index[block_index] = prev_best_cluster_index;
+
+ continue;
+ }
+
+ const color_rgba* pBlock_pixels = block.get_ptr();
+
color_rgba trial_block_colors[4];
- blk.get_block_colors(trial_block_colors, 0);
+ blk.get_block_colors_etc1s(trial_block_colors);
// precompute errors for the i-th block pixel and selector sel: [sel][i]
uint32_t trial_errors[4][16];
-
- for (int sel = 0; sel < 4; ++sel)
+
+ if (m_params.m_perceptual)
{
- for (int i = 0; i < 16; ++i)
- {
- trial_errors[sel][i] = color_distance(m_params.m_perceptual, pBlock_pixels[i], trial_block_colors[sel], false);
- }
+ for (uint32_t sel = 0; sel < 4; ++sel)
+ for (uint32_t i = 0; i < 16; ++i)
+ trial_errors[sel][i] = color_distance(true, pBlock_pixels[i], trial_block_colors[sel], false);
}
+ else
+ {
+ for (uint32_t sel = 0; sel < 4; ++sel)
+ for (uint32_t i = 0; i < 16; ++i)
+ trial_errors[sel][i] = color_distance(false, pBlock_pixels[i], trial_block_colors[sel], false);
+ }
+
+ // Compute the minimum possible errors (given any selectors) for pixels 0-15
+ uint64_t min_possible_error_0_15 = 0;
+ for (uint32_t i = 0; i < 16; i++)
+ min_possible_error_0_15 += basisu::minimum(trial_errors[0][i], trial_errors[1][i], trial_errors[2][i], trial_errors[3][i]);
+
+ // Compute the minimum possible errors (given any selectors) for pixels 4-15
+ uint64_t min_possible_error_4_15 = 0;
+ for (uint32_t i = 4; i < 16; i++)
+ min_possible_error_4_15 += basisu::minimum(trial_errors[0][i], trial_errors[1][i], trial_errors[2][i], trial_errors[3][i]);
+
+ // Compute the minimum possible errors (given any selectors) for pixels 8-15
+ uint64_t min_possible_error_8_15 = 0;
+ for (uint32_t i = 8; i < 16; i++)
+ min_possible_error_8_15 += basisu::minimum(trial_errors[0][i], trial_errors[1][i], trial_errors[2][i], trial_errors[3][i]);
+
+ // Compute the minimum possible errors (given any selectors) for pixels 12-15
+ uint64_t min_possible_error_12_15 = 0;
+ for (uint32_t i = 12; i < 16; i++)
+ min_possible_error_12_15 += basisu::minimum(trial_errors[0][i], trial_errors[1][i], trial_errors[2][i], trial_errors[3][i]);
uint64_t best_cluster_err = INT64_MAX;
uint32_t best_cluster_index = 0;
@@ -2215,7 +2686,7 @@ namespace basisu
const uint32_t total_clusters = m_use_hierarchical_selector_codebooks ? (uint32_t)pCluster_indices->size() : (uint32_t)m_selector_cluster_block_indices.size();
-#if 0
+ #if 0
for (uint32_t cluster_iter = 0; cluster_iter < total_clusters; cluster_iter++)
{
const uint32_t cluster_index = m_use_hierarchical_selector_codebooks ? (*pCluster_indices)[cluster_iter] : cluster_iter;
@@ -2246,99 +2717,73 @@ namespace basisu
early_out:
;
}
-#else
- if (m_params.m_perceptual)
- {
- for (uint32_t cluster_iter = 0; cluster_iter < total_clusters; cluster_iter++)
- {
- const uint32_t cluster_index = m_use_hierarchical_selector_codebooks ? (*pCluster_indices)[cluster_iter] : cluster_iter;
- //const etc_block& cluster_blk = m_optimized_cluster_selectors[cluster_index];
-
- uint64_t trial_err = 0;
-
- for (int i = 0; i < 16; i++)
- {
- const uint32_t sel = unpacked_optimized_cluster_selectors[cluster_index * 16 + i];
-
- trial_err += trial_errors[sel][i];
- if (trial_err > best_cluster_err)
- goto early_out;
- }
-
- if (trial_err < best_cluster_err)
- {
- best_cluster_err = trial_err;
- best_cluster_index = cluster_index;
- if (!best_cluster_err)
- break;
- }
-
- early_out:
- ;
-
- } // cluster_iter
- }
- else
+ #else
+ for (uint32_t cluster_iter = 0; cluster_iter < total_clusters; cluster_iter++)
{
- for (uint32_t cluster_iter = 0; cluster_iter < total_clusters; cluster_iter++)
- {
- const uint32_t cluster_index = m_use_hierarchical_selector_codebooks ? (*pCluster_indices)[cluster_iter] : cluster_iter;
- //const etc_block& cluster_blk = m_optimized_cluster_selectors[cluster_index];
+ const uint32_t cluster_index = m_use_hierarchical_selector_codebooks ? (*pCluster_indices)[cluster_iter] : cluster_iter;
+
+ const uint8_t* pSels = &unpacked_optimized_cluster_selectors[cluster_index * 16];
- uint64_t trial_err = 0;
+ uint64_t trial_err = (uint64_t)trial_errors[pSels[0]][0] + trial_errors[pSels[1]][1] + trial_errors[pSels[2]][2] + trial_errors[pSels[3]][3];
+ if ((trial_err + min_possible_error_4_15) >= best_cluster_err)
+ continue;
- for (int i = 0; i < 16; i++)
- {
- const uint32_t sel = unpacked_optimized_cluster_selectors[cluster_index * 16 + i];
+ trial_err += (uint64_t)trial_errors[pSels[4]][4] + trial_errors[pSels[5]][5] + trial_errors[pSels[6]][6] + trial_errors[pSels[7]][7];
+ if ((trial_err + min_possible_error_8_15) >= best_cluster_err)
+ continue;
- trial_err += trial_errors[sel][i];
- if (trial_err > best_cluster_err)
- goto early_out2;
- }
+ trial_err += (uint64_t)trial_errors[pSels[8]][8] + trial_errors[pSels[9]][9] + trial_errors[pSels[10]][10] + trial_errors[pSels[11]][11];
+ if ((trial_err + min_possible_error_12_15) >= best_cluster_err)
+ continue;
- if (trial_err < best_cluster_err)
- {
- best_cluster_err = trial_err;
- best_cluster_index = cluster_index;
- if (!best_cluster_err)
- break;
- }
+ trial_err += (uint64_t)trial_errors[pSels[12]][12] + trial_errors[pSels[13]][13] + trial_errors[pSels[14]][14] + trial_errors[pSels[15]][15];
- early_out2:
- ;
+ if (trial_err < best_cluster_err)
+ {
+ best_cluster_err = trial_err;
+ best_cluster_index = cluster_index;
+ if (best_cluster_err == min_possible_error_0_15)
+ break;
+ }
- } // cluster_iter
- }
-#endif
+ } // cluster_iter
+ #endif
blk.set_raw_selector_bits(m_optimized_cluster_selectors[best_cluster_index].get_raw_selector_bits());
m_block_selector_cluster_index[block_index] = best_cluster_index;
-
- {
- std::lock_guard<std::mutex> lock(m_lock);
- vector_ensure_element_is_valid(new_cluster_indices, best_cluster_index);
- new_cluster_indices[best_cluster_index].push_back(block_index);
- }
+ prev_best_cluster_index = best_cluster_index;
} // block_index
-#ifndef __EMSCRIPTEN__
+ #ifndef __EMSCRIPTEN__
} );
-#endif
+ #endif
} // block_index_iter
-#ifndef __EMSCRIPTEN__
+ #ifndef __EMSCRIPTEN__
m_params.m_pJob_pool->wait_for_all();
-#endif
+ #endif
+
+ for (uint32_t i = 0; i < m_selector_cluster_block_indices.size(); i++)
+ {
+ m_selector_cluster_block_indices[i].resize(0);
+ m_selector_cluster_block_indices[i].reserve(128);
+ }
- m_selector_cluster_block_indices.swap(new_cluster_indices);
- }
+ for (uint32_t block_index = 0; block_index < m_total_blocks; block_index++)
+ {
+ const uint32_t best_cluster_index = m_block_selector_cluster_index[block_index];
- for (uint32_t i = 0; i < m_selector_cluster_block_indices.size(); i++)
- vector_sort(m_selector_cluster_block_indices[i]);
+ vector_ensure_element_is_valid(m_selector_cluster_block_indices, best_cluster_index);
+ m_selector_cluster_block_indices[best_cluster_index].push_back(block_index);
+ }
+
+ } // if (use_cpu)
+
+ debug_printf("Elapsed time: %3.3f secs\n", tm.get_elapsed_secs());
}
// TODO: Remove old ETC1 specific stuff, and thread this.
@@ -2837,7 +3282,81 @@ namespace basisu
//debug_printf("validate_output: %u\n", validate_output());
}
-
+
+ // Endpoint clusterization hierarchy integrity checker.
+ // Note this doesn't check for empty clusters.
+ bool basisu_frontend::validate_endpoint_cluster_hierarchy(bool ensure_clusters_have_same_parents) const
+ {
+ if (!m_endpoint_parent_clusters.size())
+ return true;
+
+ int_vec subblock_parent_indices(m_total_blocks * 2);
+ subblock_parent_indices.set_all(-1);
+
+ int_vec subblock_cluster_indices(m_total_blocks * 2);
+ subblock_cluster_indices.set_all(-1);
+
+ for (uint32_t parent_index = 0; parent_index < m_endpoint_parent_clusters.size(); parent_index++)
+ {
+ for (uint32_t i = 0; i < m_endpoint_parent_clusters[parent_index].size(); i++)
+ {
+ uint32_t subblock_index = m_endpoint_parent_clusters[parent_index][i];
+ if (subblock_index >= m_total_blocks * 2)
+ return false;
+
+ // If the endpoint cluster lives in more than one parent node, that's wrong.
+ if (subblock_parent_indices[subblock_index] != -1)
+ return false;
+
+ subblock_parent_indices[subblock_index] = parent_index;
+ }
+ }
+
+ // Make sure all endpoint clusters are present in the parent cluster.
+ for (uint32_t i = 0; i < subblock_parent_indices.size(); i++)
+ {
+ if (subblock_parent_indices[i] == -1)
+ return false;
+ }
+
+ for (uint32_t cluster_index = 0; cluster_index < m_endpoint_clusters.size(); cluster_index++)
+ {
+ int parent_index = 0;
+
+ for (uint32_t i = 0; i < m_endpoint_clusters[cluster_index].size(); i++)
+ {
+ uint32_t subblock_index = m_endpoint_clusters[cluster_index][i];
+ if (subblock_index >= m_total_blocks * 2)
+ return false;
+
+ if (subblock_cluster_indices[subblock_index] != -1)
+ return false;
+
+ subblock_cluster_indices[subblock_index] = cluster_index;
+
+ // There are transformations on the endpoint clusters that can break the strict tree requirement
+ if (ensure_clusters_have_same_parents)
+ {
+ // Make sure all the subblocks are in the same parent cluster
+ if (!i)
+ parent_index = subblock_parent_indices[subblock_index];
+ else if (subblock_parent_indices[subblock_index] != parent_index)
+ return false;
+ }
+ }
+ }
+
+ // Make sure all endpoint clusters are present in the parent cluster.
+ for (uint32_t i = 0; i < subblock_cluster_indices.size(); i++)
+ {
+ if (subblock_cluster_indices[i] == -1)
+ return false;
+ }
+
+ return true;
+ }
+
+ // This is very slow and only intended for debugging/development. It's enabled using the "-validate_etc1s" command line option.
bool basisu_frontend::validate_output() const
{
debug_printf("validate_output\n");
@@ -2889,29 +3408,7 @@ namespace basisu
CHECK(rdo_output_block.get_base5_color() == blk.get_base5_color());
CHECK(rdo_output_block.get_delta3_color() == blk.get_delta3_color());
CHECK(rdo_output_block.get_raw_selector_bits() == blk.get_raw_selector_bits());
-
- if (m_params.m_pGlobal_sel_codebook)
- {
- bool used_global_cb = true;
- if (m_params.m_use_hybrid_selector_codebooks)
- used_global_cb = m_selector_cluster_uses_global_cb[selector_cluster_index];
-
- if (used_global_cb)
- {
- basist::etc1_global_selector_codebook_entry_id pal_id(get_selector_cluster_global_selector_entry_ids()[selector_cluster_index]);
-
- basist::etc1_selector_palette_entry pal_entry(m_params.m_pGlobal_sel_codebook->get_entry(pal_id));
-
- for (uint32_t y = 0; y < 4; y++)
- {
- for (uint32_t x = 0; x < 4; x++)
- {
- CHECK(pal_entry(x, y) == blk.get_selector(x, y));
- }
- }
- }
- }
-
+
#undef CHECK
}