The most substantial recent progress in NLP has focused on text understanding and generation. Yet, humans rely on additional contexts to communicate efficiently, such as speech, vision, situational context (time and space), commonsense knowledge, social and cultural norms. LLMs are still seriously limited in dealing with extra-linguistic contexts. Enhancing LLM-based NLP models with extra-linguistic contexts could improve their applicability to real-world usages. Such models will be able to adapt their outputs based on time, location, and the user’s background. Endowing models with commonsense reasoning abilities will additionally improve their robustness and help them address unknown inputs reasonably and consistently.