Mohamed Zeid
Mohamed Zeid
Any progress on this?
@EtienneAb3d very interesting idea. If you don't mind, which tool did you use to do the chunking? Is it NLTK?
@EtienneAb3d This sounds great. Can you give us an idea how many chunks and full sentences are in your training data to achieve such great results? I see that you...
Thank you so much @EtienneAb3d for sharing this valuable information. One last thing: You said above that there were > many redundancies, and a lot of noise/errors What did you...
Thanks @EtienneAb3d for your reply. It sounds you have a great solution in place. Thanks for letting us know that you can make a demo. I will keep this in...
@EtienneAb3d I am trying to stop 'early termination' and I am trying to implement your code above, but not sure where exactly it should go. Here is the early termination...
Thanks a million, @EtienneAb3d for getting back to me. I appreciate it.
Hi @EtienneAb3d and @davidecaroselli , It seems that during preprocessing, MMT excludes strings with low character count. So, my question to you @davidecaroselli : is there a way to force...
Hi @EtienneAb3d, Sorry for the confusion. We were training with a placeholder file that doesn't contain meaningful data. So, for words or product names that we don't want to translate/protect,...
@EtienneAb3d I meant after adding these placeholder pairs to the training data, tried to translate some documents with exact same placeholders in there, but MMT didn't match these placeholders and...