Login / Signup

Pretrained Transformer Language Models Versus Pretrained Word Embeddings for the Detection of Accurate Health Information on Arabic Social Media: Comparative Study.

Yahya AlbalawiNikola S NikolovJim Buckley
Published in: JMIR formative research (2022)
Our results indicate that the pretrained language model AraBERTv0.2 is the best model for classifying tweets as carrying either inaccurate or accurate health information. Future studies should consider applying ensemble learning to combine the best models as it may produce better results.
Keyphrases
  • health information
  • social media
  • autism spectrum disorder
  • healthcare
  • high resolution
  • label free
  • machine learning
  • mass spectrometry
  • case control
  • psychometric properties