Safety

#21
by Javier81k - opened

I can't see anything about alignment, deception and ai safety in the model card. Is the model following best practices on ai safety? Is it safe to use this model?

what are you afraid about ? that it starts eating the user remotely ?

OpenBMB org

Hello, thank you for your feedback.
Our model uses the alignment method related to RLHF and RLAIF, which is also the method of our team and is clearly stated in the github readme.
I hope this is enough to solve your problem.

https://github.com/RLHF-V/RLHF-V
https://github.com/RLHF-V/RLAIF-V

Thanks a lot. I just checked the readme and there is not much said regarding that RLHF-V.
This is a very good methodology, but it would be great to be able to measure the safety of it, with a score like the "FLI AI Safety Index".
Do you think that would be possible in the future?

https://futureoflife.org/document/fli-ai-safety-index-2024/

Dumb decel,

I’ll just drop this here

image.jpeg

Sign up or log in to comment