Skip to content

Conversation

@ASuresh0524
Copy link
Collaborator

Fixing #154 issues

The merge conflict in embedding_compute.py has been resolved while preserving all token limit fixes AND integrating the batch processing improvements from PR #152:

Combined Features:

  • Token-aware truncation with tiktoken support
  • Modern /api/embed endpoint with true batch processing
  • Enhanced error detection for token limit violations
  • Model token limit registry for safety
  • All performance improvements from batch processing

The PR is now ready for review and provides the complete solution for issue #153.

Checklist

  • Tests pass (uv run pytest)
  • Code formatted (ruff format and ruff check)
  • Pre-commit hooks pass (pre-commit run --all-files)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants