arXiv research summaries for Computation and Language from October 25, 2023. You can find summaries and links to each article here.
Today's Theme (LLM-Generated)
- Evaluating capabilities and limitations of large language models (LLMs) like ChatGPT and GPT-4 through tasks such as reasoning, summarization, and parsing.
- Improving alignment and control of LLMs through techniques like reinforcement learning from human feedback, supervised iterative learning, and contextualized prompt optimization.
- Enhancing fairness, interpretability, and transparency of models through bias mitigation, rationale extraction, and pretraining data detection.
- Advancing multilingual and multimodal models through pretraining, benchmarking, and adapting models like mBERT to new languages and modalities.
- Quantization and efficiency improvements for deploying large models through methods like post-training 4-bit floating point quantization.