Skip to content

Commit

Permalink
fix windows build (#34)
Browse files Browse the repository at this point in the history
  • Loading branch information
dorbanianas authored Oct 21, 2024
1 parent f46ba8f commit 7d571b1
Show file tree
Hide file tree
Showing 3 changed files with 6 additions and 12 deletions.
6 changes: 2 additions & 4 deletions src/core/functions/scalar/llm_complete.cpp
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
#include <algorithm>
#include <cmath>
#include <flockmtl/common.hpp>
#include <flockmtl/core/functions/scalar.hpp>
#include <flockmtl/core/model_manager/model_manager.hpp>
Expand Down Expand Up @@ -86,8 +84,8 @@ inline std::vector<std::string> ConstructPrompts3(std::vector<nlohmann::json> &u
auto template_tokens = Tiktoken::GetNumTokens(llm_complete_prompt_template);
auto max_tokens_for_rows = model_max_tokens - template_tokens;
auto max_chunk_size = max_tokens_for_rows / row_tokens;
#undef min
auto chunk_size = std::min(max_chunk_size, static_cast<int>(unique_rows.size()));
auto chunk_size = max_chunk_size > static_cast<int>(unique_rows.size()) ? static_cast<int>(unique_rows.size())
: max_chunk_size;
auto num_chunks = static_cast<int>(std::ceil(static_cast<double>(unique_rows.size()) / chunk_size));

for (int i = 0; i < num_chunks; ++i) {
Expand Down
6 changes: 2 additions & 4 deletions src/core/functions/scalar/llm_complete_json.cpp
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
#include <algorithm>
#include <cmath>
#include <flockmtl/common.hpp>
#include <flockmtl/core/functions/scalar.hpp>
#include <flockmtl/core/model_manager/model_manager.hpp>
Expand Down Expand Up @@ -86,8 +84,8 @@ inline std::vector<std::string> ConstructPrompts(std::vector<nlohmann::json> &un
auto template_tokens = Tiktoken::GetNumTokens(llm_complete_json_prompt_template);
auto max_tokens_for_rows = model_max_tokens - template_tokens;
auto max_chunk_size = max_tokens_for_rows / row_tokens;
#undef min
auto chunk_size = std::min(max_chunk_size, static_cast<int>(unique_rows.size()));
auto chunk_size = max_chunk_size > static_cast<int>(unique_rows.size()) ? static_cast<int>(unique_rows.size())
: max_chunk_size;
auto num_chunks = static_cast<int>(std::ceil(static_cast<double>(unique_rows.size()) / chunk_size));

for (int i = 0; i < num_chunks; ++i) {
Expand Down
6 changes: 2 additions & 4 deletions src/core/functions/scalar/llm_filter.cpp
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
#include <algorithm>
#include <cmath>
#include <flockmtl/common.hpp>
#include <flockmtl/core/functions/scalar.hpp>
#include <flockmtl/core/model_manager/model_manager.hpp>
Expand Down Expand Up @@ -85,8 +83,8 @@ inline std::vector<std::string> ConstructPrompts2(std::vector<nlohmann::json> &u
auto template_tokens = Tiktoken::GetNumTokens(llm_filter_prompt_template);
auto max_tokens_for_rows = model_max_tokens - template_tokens;
auto max_chunk_size = max_tokens_for_rows / row_tokens;
#undef min
auto chunk_size = std::min(max_chunk_size, static_cast<int>(unique_rows.size()));
auto chunk_size = max_chunk_size > static_cast<int>(unique_rows.size()) ? static_cast<int>(unique_rows.size())
: max_chunk_size;
auto num_chunks = static_cast<int>(std::ceil(static_cast<double>(unique_rows.size()) / chunk_size));

for (int i = 0; i < num_chunks; ++i) {
Expand Down

0 comments on commit 7d571b1

Please sign in to comment.