Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Port "bert multi lingual tpu training (8 cores)" to Ignite #960

Closed
vfdev-5 opened this issue Apr 22, 2020 · 6 comments · Fixed by #1656
Closed

Port "bert multi lingual tpu training (8 cores)" to Ignite #960

vfdev-5 opened this issue Apr 22, 2020 · 6 comments · Fixed by #1656

Comments

@vfdev-5
Copy link
Collaborator

vfdev-5 commented Apr 22, 2020

🚀 Feature

Recently included example of TPU usage with Ignite includes a training on a single TPU.
Idea is to port this kaggle kernel: https://www.kaggle.com/abhishek/bert-multi-lingual-tpu-training-8-cores to Ignite and include it in Ignite's show-case

Based on #952 (comment)

@vfdev-5 vfdev-5 self-assigned this Apr 22, 2020
@vfdev-5 vfdev-5 removed their assignment Feb 10, 2021
@ahmedo42
Copy link
Contributor

I'd like to work on it , should we use the same dataset?? and the same BERT model ?? I think the purpose of this is to make users comfortable using ignite with multicore TPU.

Maybe something like this with added featuers of ignite?? PyTorch on Cloud TPUs: MultiCore Training AlexNet on Fashion MNIST .

@vfdev-5
Copy link
Collaborator Author

vfdev-5 commented Feb 14, 2021

@ahmedo42 thanks for asking. I agree that the purpose is more about Ignite and multiple TPUs which is more or less covered here : https://github.com/pytorch/ignite/tree/master/examples/contrib/cifar10#colab-on-8-tpus

An NLP example on multiple TPUs possible trained on Kaggle TPUv3 (vs TPUv2 on colab) could be still nice to have in addition to cifar10 example.

What do you think ?

@ahmedo42
Copy link
Contributor

ahmedo42 commented Feb 14, 2021

Didn't really know the difference between TPU's on Kaggle and On Colab 😃 .

An NLP example on multiple TPUs possible trained on Kaggle TPUv3

Totally agree , an NLP example is needed , so it should a ported notebook in examples/notebooks using the same data? with some explanations perhaps?

@vfdev-5
Copy link
Collaborator Author

vfdev-5 commented Feb 14, 2021

Well, I'm hesitating between two:

  • notebook is good and could be easily read on Kaggle notebook interface + explanations
  • script is good as well if user would like to extend it (what is probably more important need for us)

Do you have any NLP background to suggest what would be more interesting to have here ?
Maybe, it would be nice to port an example from hugging face transformers and run everything on TPUs.

@ahmedo42
Copy link
Contributor

  • notebook is good and could be easily read on Kaggle notebook interface + explanations

well , notebooks could be extended too , people fork notebooks and extend them all the time on Kaggle

  • script is good as well if user would like to extend it (what is probably more important need for us)

This seems like a best practice , almost all of the huggingface examples are scripts which allow for a higher degree of control from the user's perspective and probably we should do that

Do you have any NLP background to suggest what would be more interesting to have here ?

Well , I think we really need a Transformer example since it's a huge trend and it's the De Facto in NLP right now , so porting an example from huggingface would be a good idea.

@vfdev-5
Copy link
Collaborator Author

vfdev-5 commented Feb 15, 2021

Sounds good !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants