EADST

Sharding and SafeTensors in Hugging Face Transformers

In the Hugging Face transformers library, managing large models efficiently is crucial, especially when working with limited disk space or specific file size requirements. Two key features that help with this are sharding and the use of SafeTensors.

Sharding

Sharding is the process of splitting a large model's weights into smaller files or "shards." This is particularly useful when dealing with large models that exceed file size limits or when you want to manage storage more effectively.

Usage

To shard a model during the saving process, you can use the max_shard_size parameter in the save_pretrained method. Here's an example:

# Save the model with sharding, setting the maximum shard size to 1GB
model.save_pretrained('./model_directory', max_shard_size="1GB")

In this example, the model's weights will be divided into multiple files, each not exceeding 1GB. This can make storage and transfer more manageable, especially when dealing with large-scale models.

SafeTensors

The safetensors library provides a new format for storing tensors in a safe and efficient way. Unlike traditional formats like PyTorch's .pt files, SafeTensors ensures that the tensor data cannot be accidentally executed as code, offering an additional layer of security. This is particularly important when sharing models across different systems or with the community.

Usage

To save a model using SafeTensors, simply specify the safe_serialization parameter when saving:

# Save the model using SafeTensors format
model.save_pretrained('./model_directory', safe_serialization=True)

This will create files with the .safetensors extension, ensuring the saved tensors are stored safely.

Combining Sharding and SafeTensors

You can combine both sharding and SafeTensors to save a large model securely and efficiently:

# Save the model with sharding and SafeTensors
model.save_pretrained('./model_directory', max_shard_size="1GB", safe_serialization=True)

This setup splits the model into shards, each in the SafeTensors format, offering both manageability and security.

Conclusion

By leveraging sharding and SafeTensors, Hugging Face transformers users can handle large models more effectively. Sharding helps manage file sizes, while SafeTensors ensures the safe storage of tensor data. These features are essential for anyone working with large-scale models, providing both practical and security benefits.

相关标签
About Me
XD
Goals determine what you are going to be.
Category
标签云
Permission FP64 LaTeX CSV Shortcut printf mmap UI 签证 Linux TensorRT 腾讯云 Streamlit Hotel Algorithm SAM Magnet 图形思考法 Domain DeepSeek Quantization Paper Crawler FlashAttention Distillation 云服务器 CUDA Input OpenAI Numpy Video 继承 Statistics scipy Data Freesound SPIE 多进程 Color Clash Cloudreve diffusers Plotly GGML torchinfo Windows Augmentation CV WebCrawler CC Llama Jetson COCO Qwen2.5 Bin Search git-lfs HuggingFace InvalidArgumentError Quantize PDB Breakpoint Pandas Dataset Bipartite Qwen Markdown Agent TSV 顶会 SVR Git JSON HaggingFace RAR TensorFlow EXCEL Food Tiktoken Docker 强化学习 CLAP NLTK Vmess Anaconda Review Website 域名 Zip Mixtral Card Gemma Python 净利润 VGG-16 LoRA 音频 Miniforge Tracking Datetime 报税 FastAPI Tensor Vim uwsgi GoogLeNet 算法题 ResNet-50 PyCharm Diagram Base64 CAM XGBoost tqdm transformers 第一性原理 CEIR Firewall OCR Hungarian DeepStream WAN FP32 IndexTTS2 版权 Animate Django Ubuntu RGB Github Safetensors Image2Text Proxy CTC SQL Pickle API News Interview BTC 关于博主 Password Baidu MD5 PIP Knowledge uWSGI NameSilo XML Pillow Transformers Sklearn Nginx UNIX OpenCV 多线程 GIT Excel FP16 Paddle 财报 搞笑 ONNX Logo Plate LeetCode Qwen2 SQLite Random Web Math 图标 LLAMA FP8 TTS 飞书 VPN GPTQ C++ Bitcoin QWEN Conda YOLO Hilton hf 递归学习法 PyTorch Pytorch git GPT4 Bert BeautifulSoup Google ChatGPT PDF v0.dev 证件照 Land Jupyter Use Template v2ray llama.cpp Michelin BF16 Rebuttal Heatmap Attention Claude ModelScope icon tar AI Disk Translation LLM NLP 公式 VSCode Ptyhon logger 阿里云
站点统计

本站现有博文323篇,共被浏览795462

本站已经建立2493天!

热门文章
文章归档
回到顶部