Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support Mistral-small-inst-2409 #2077

Merged
merged 8 commits into from
Sep 20, 2024

Conversation

DaozeZhang
Copy link
Collaborator

PR type

  • Document Updates
  • More Models or Datasets Support

PR information

Support Mistral-small-inst-2409

@DaozeZhang DaozeZhang merged commit bea9867 into modelscope:main Sep 20, 2024
1 of 2 checks passed
@DaozeZhang DaozeZhang deleted the mistral-small-2409 branch September 20, 2024 02:31
Jintao-Huang pushed a commit that referenced this pull request Sep 23, 2024
* add transformers in gitignore

* fix a typo bug in text-caps

* support mistral-small-2409

* add .run to gitignore

* add the doc of mistral-small-2409

* remove transformers in gitignore (to PR this branch)

* update doc
tastelikefeet added a commit to tastelikefeet/swift that referenced this pull request Sep 26, 2024
* commit '57b3b9e46aa01bdc5c29b5e3d1e2da0582c9b282': (23 commits)
  fix not impl bug (modelscope#2134)
  Support fine-tuning MLLama. (modelscope#2132)
  Support for fine-tuning and deployment of the Llama 3.2 series models. (modelscope#2130)
  support got-ocr2 (modelscope#2123)
  [TorchAcc] fix: fix find_labels and can_return_loss (modelscope#2120)
  fix qwen2-audio (modelscope#2116)
  Fix qwen2-vl zero2/3 (modelscope#2114)
  support vllm & qwen2-vl video (modelscope#2110)
  Support for fine-tuning Llama 3.1 Omni. (modelscope#2106)
  fix infer device_map (modelscope#2105)
  fix cpu infer device_map (modelscope#2103)
  fix dataset preprocess (modelscope#2102)
  fix deploy openai compat (modelscope#2101)
  Fix the issue with media_offset in owl3 when batch_size > 1. (modelscope#2100)
  fix vllm tokenizer (modelscope#2099)
  Support for fine-tuning Pixtral-12B. (modelscope#2090)
  fix multiprocess remove_columns (modelscope#2088)
  fix qwen2.5 template (modelscope#2081)
  dynamic vit gradient_checkpointing (modelscope#2071)
  Support Mistral-small-inst-2409 (modelscope#2077)
  ...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants