Changelog
Source:NEWS.md
EndpointR 0.2
- error message and status propagation improvement. Now writes .error, .error_msg (standardised across package), and .status. Main change is preventing httr2 eating the errors before we can deal with them
- adds parquet writing to oai_complete_df and oai_embed_df
- adds chunks func to oai_embed, and re-writes all batch -> chunk logic
- implements the Anthropic messages API with structured outputs (via BETA)
- adds
ant_complete_df()andant_complete_chunks()for batch/chunked processing with the Anthropic API, with parquet writing and metadata tracking - metadata tracking now includes
schemaandsystem_promptfor both OpenAI and Anthropic chunked processing functions - bug fix: S7 schema objects now correctly serialised to metadata.json (previously caused “No method asJSON S3 class: S7_object” error)
- adds spelling test, sets language to en-GB in DESCRIPTION
EndpointR 0.1.2
File writing improvements:
hf_embed_df()andhf_classify_df()now write intermediate results as.parquetfiles tooutput_dirdirectories, similar to improvements in 0.1.1 for OpenAI functionsParameter changes: Moved from
batch_sizetochunk_sizeargument acrosshf_embed_df(),hf_classify_df(), andoai_complete_df()for consistencyNew chunking functions: Introduced
hf_embed_chunks()andhf_classify_chunks()for more efficient batch processing with better error handlingDependency update: Package now depends on
arrowfor faster.parquetfile writing and reading-
Metadata tracking: Hugging Face functions that write to files (
hf_embed_df(),hf_classify_df(),hf_embed_chunks(),hf_classify_chunks()) now writemetadata.jsonto output directories containing:- Endpoint URL and API key name used
- Processing parameters (chunk_size, concurrent_requests, timeout, max_retries)
- Inference parameters (truncate, max_length)
- Timestamp and row counts
- Useful for debugging, reproducibility, and tracking which models/endpoints were used
max_length parameter: Added
max_lengthparameter tohf_classify_df()andhf_classify_chunks()for text truncation control. Note:hf_embed_df()handles truncation automatically via endpoint configuration (setAUTO_TRUNCATEin endpoint settings)-
New utility functions:
-
hf_get_model_max_length()- Retrieve maximum token length for a Hugging Face model -
hf_get_endpoint_info()- Retrieve detailed information about a Hugging Face Inference Endpoint
-
Improved reporting: Chunked/batch processing functions now report total successes and failures at completion
EndpointR 0.1.1
-
oai_complete_chunks()function to better support for chunking/batching inoai_complete_df() -
oai_complete_df()now writes to a file to mitigate the chance of completely lost data
EndpointR 0.1.0
Initial BETA release, ships with:
- Support for embeddings and classification with Hugging Face Inference API & Dedicated Inference Endpoints
- Support for text completion using OpenAI models via the Chat Completions API
- Support for embeddings with the OpenAI Embeddings API
- Structured outputs via JSON schemas and validators