Language Agnostic Code Embeddings
Saiteja Utpala, Alex Gu, et al.
NAACL 2024
Large Language Models (LLM) are being increasingly used in many domains including legal and justice. General purpose models trained on web data are not performant enough on legal text analytics (LTA) tasks while fine tuning task specific models is expensive because of the annotation and compute costs. Pre-training domain or application specific models is increasingly popular. However pre-training LLMs in small domain corpora like Indian legal documents is stymied by repetitive and less relevant information in court judgements and records. We introduce InLegalLlama models and show that pre-training LLMs on knowledge graph triples significantly reduces the training effort while retaining comparable performance on LTA tasks.
Saiteja Utpala, Alex Gu, et al.
NAACL 2024
Gosia Lazuka, Andreea Simona Anghel, et al.
SC 2024
Gabriele Picco, Lam Thanh Hoang, et al.
EMNLP 2021
Takuma Udagawa, Aashka Trivedi, et al.
EMNLP 2023