arXiv research summaries for Computation and Language from October 26, 2023. You can find summaries and links to each article here.
Today's research themes (AI summary)
Advances in large language models through fine-tuning, prompting, and quantization. Examples include HuggingFace LLaMA, custom prompting for ChatGPT, and quantizing models down to 4 bits.
Analyzing capabilities and limitations of large language models on reasoning, commonsense, and factual knowledge tasks. Papers probe model knowledge in areas like theory of mind, spatial reasoning, and stumpers.
Improving natural language processing applications with techniques like contrastive learning, targeted pretraining, and reinforcement learning from human feedback. Areas include intent detection, question answering, summarization.
Evaluating bias, fairness, transparency, and ethics of large language models and datasets. Analyzing issues like occupational stereotypes, licensing and attribution of training data, and fairness-explainability tradeoffs.
Enabling multilinguality, cross-linguality, and transfer learning. Examples include cross-lingual transfer for low-resource languages, code embeddings across programming languages, and continual learning for multilingual ASR.