An automatic short-answer grading system is designed to evaluate and grade short answers provided by students or test-takers in a way that mimics human grading. Various techniques of NLP are implemented to assess the answers' content, coherence, and relevance.
Components
- Preprocessing:
- Tokenization
- Stopword Removal
- Lemmatization/Stemming
- Feature Extraction:
- Bag of Words(BoW)
- TF-IDF (Term Frequency-Inverse Document Frequency)
- Word Embeddings
- Similarity Measurement:
- Cosine Similarity
- Semantic Similarity
- Grading Method:
- Rule-based Systems
- Machine Learning Models
- Deep Learning Models
- Evaluation Metrics:
- Precision and Recall
- F1 Score
- Cohen’s Kappa
Tools and Libraries
- NLP Libraries:
- NLTK
- SpaCy
- Gensim
- Machine Learning Framework:
- Scikit-learn
- Tensorflow
- PyTorch
- Pre-Trained Deeplearning Models:
- Word2Vec
- Glove (Global Vectors for Word Representation)
- FastText
- ELMo (Embeddings from Language Models)
- BERT (Bidirectional Encoder Representations from Transformers)
Word Embedding Methods
(inc. Explanation, Differences, Pros and Cons)
1. Word2Vec
- Explanation:
Word2Vec, developed by Google, consists of two main models: Continuous Bag of Words (CBOW) and Skip-gram.
- CBOW: Predicts the target word from its context.
- Skip-gram: Predicts the context from the target word.
- Differences:
- CBOW is faster and works better with smaller datasets.
- Skip-gram is slower but performs better with infrequent words.
- Pros:
- Captures semantic relationships (e.g., king - man + woman = queen).
- Efficient to train and use.
- Cons:
- Context is limited to a fixed-size window.
- Ignores word order within the context window.
2. GloVe (Global Vectors for Word Representation)