-
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-1-deberta-nli-reward
Text Generation • 8B • Updated • 217 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-2-deberta-nli-reward
Text Generation • 8B • Updated • 224 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-3-deberta-nli-reward
Text Generation • 8B • Updated • 224 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-4-deberta-nli-reward
Text Generation • 8B • Updated • 220
Mikhail Seleznev
myyycroft
AI & ML interests
NLP, AI Safety
Recent Activity
updated a collection 22 days ago
emergent-misalignment-evolutionary-finetuning-7b-cross-encod updated a collection 22 days ago
emergent-misalignment-evolutionary-finetuning-7b-cross-encod updated a collection 22 days ago
emergent-misalignment-evolutionary-finetuning-7b-cross-encodOrganizations
gpt2-toxicity-pretrain-conditional
Checkpoints for conditional pretraining of gpt-2 models for detoxification task as described in https://arxiv.org/abs/2302.08582.
-
myyycroft/gpt2-toxicity-conditional-5000
Text Generation • 0.1B • Updated • 41 -
myyycroft/gpt2-toxicity-conditional-10000
Text Generation • 0.1B • Updated • 39 -
myyycroft/gpt2-toxicity-conditional-15000
Text Generation • 0.1B • Updated • 37 -
myyycroft/gpt2-toxicity-conditional-20000
Text Generation • 0.1B • Updated • 37
emergent-misalignment-evolutionary-finetuning
-
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice
0.5B • Updated • 5 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-1
Text Generation • 0.5B • Updated • 53 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-2
Text Generation • 0.5B • Updated • 52 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-3
Text Generation • 0.5B • Updated • 53
gpt2-PII-pretrain-mle
Checkpoints for MLE baselines of gpt-2 models trained for PII task as described in https://arxiv.org/abs/2302.08582.
emergent-misalignment-evolutionary-finetuning-7b-cross-encod
-
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-1-deberta-nli-reward
Text Generation • 8B • Updated • 217 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-2-deberta-nli-reward
Text Generation • 8B • Updated • 224 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-3-deberta-nli-reward
Text Generation • 8B • Updated • 224 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-4-deberta-nli-reward
Text Generation • 8B • Updated • 220
emergent-misalignment-evolutionary-finetuning
-
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice
0.5B • Updated • 5 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-1
Text Generation • 0.5B • Updated • 53 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-2
Text Generation • 0.5B • Updated • 52 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-3
Text Generation • 0.5B • Updated • 53
gpt2-toxicity-pretrain-conditional
Checkpoints for conditional pretraining of gpt-2 models for detoxification task as described in https://arxiv.org/abs/2302.08582.
-
myyycroft/gpt2-toxicity-conditional-5000
Text Generation • 0.1B • Updated • 41 -
myyycroft/gpt2-toxicity-conditional-10000
Text Generation • 0.1B • Updated • 39 -
myyycroft/gpt2-toxicity-conditional-15000
Text Generation • 0.1B • Updated • 37 -
myyycroft/gpt2-toxicity-conditional-20000
Text Generation • 0.1B • Updated • 37
gpt2-PII-pretrain-mle
Checkpoints for MLE baselines of gpt-2 models trained for PII task as described in https://arxiv.org/abs/2302.08582.