DeepSpeedExamples icon indicating copy to clipboard operation
DeepSpeedExamples copied to clipboard

Example models using DeepSpeed

Results 323 DeepSpeedExamples issues
Sort by recently updated
recently updated
newest added

I’m excited about the recent introduction of Domino and its impressive TP optimization. When I was using deepspeed-domino to better overlap comm & comp in TP, I found domino use...

Hi, thank you for the amazing demo and doc! I have a question regarding this [section](https://github.com/microsoft/DeepSpeedExamples/blob/master/inference/huggingface/zero_inference/model-support.md#5-compute-attention-scores) in zero-inference. It is mentioned that `"Thus, our current implementation computes attention scores on...

deepspeed --master_port 25604 --num_gpus 1 main.py \ --data_path mydata/ \ --data_split 0,10,0 \ --num_padding_at_beginning 0 \ --model_name_or_path bloom_3b1/ \ --per_device_train_batch_size 2 \ --per_device_eval_batch_size 2 \ --max_seq_len 1024 \ --learning_rate 9.65e-7...

Issue: In the original code: `e2e_rlhf.py` line 68 ``` parser.add_argument( "--reward-model", type=lambda x: x.replace("facebook/opt-", ""), default="350m", choices=("350m"), help="Which facebook/opt-* model to use for Reward (step 2)", ) ``` The choices...

When I use default command, it seems to use 29500 as master_port. However, the master_port seems unchangable,even when I use "--master_port 29501" or change it using "deepspeed.init_distributed(dist_backend='nccl', distributed_port=config.master_port)" error message:...

I know this is a path problem, and I have modified the code to know the parent directory, but I still can't find the package, step1_supervised_finetuning and utils are at...

In your example you convert the AlexNet into a list of layers: ``` def join_layers(vision_model): layers = [ *vision_model.features, vision_model.avgpool, lambda x: torch.flatten(x, 1), *vision_model.classifier, ] return layers ``` which...

I'm wondering if we can take the ZenFlow finetuning example, and extend this example into a test bed of different DeepSpeed technologies. The ZenFlow finetuning example: https://github.com/deepspeedai/DeepSpeedExamples/tree/master/training/DeepSpeed-ZenFlow/finetuning The reason is...

https://github.com/microsoft/DeepSpeedExamples/blob/master/applications/DeepSpeed-Chat/training/step3_rlhf_finetuning/ppo_trainer.py#L148 ``` def compute_rewards(self, prompts, log_probs, ref_log_probs, reward_score, action_mask): kl_divergence_estimate = -self.kl_ctl * (log_probs - ref_log_probs) rewards = kl_divergence_estimate start = prompts.shape[1] - 1 ends = start + action_mask[:, start:].sum(1)...