| # from transformers import AutoModel,AutoTokenizer | |
| # tokenizer = AutoTokenizer.from_pretrained('NewBreaker/chatglm-6b-int4',trust_remote_code=True) | |
| # model = AutoModel.from_pretrained('NewBreaker/chatglm-6b-int4',trust_remote_code=True) | |
| from transformers import AutoModelForQuestionAnswering | |
| model = AutoModelForQuestionAnswering.from_pretrained("NewBreaker/chatglm-6b-int4") | |
| # import requests | |
| # api_url = "https://huggingface.co/proxy/api-inference.huggingface.co/usage/pinned_models" | |
| # headers = {"Authorization": f"Bearer {API_TOKEN}"} | |
| # response = requests.get(api_url, headers=headers) | |
| # # {"pinned_models": [...], "allowed_pinned_models": 5} |