This study investigates the automatic generation of semantic norms on word specificity using various quantized open-source local Large Language Models (LLMs), including a comparison with a proprietary model (i.e. GPT-4). Word specificity norms on English are still not public, thus they are not included in the training datasets of all tested models. This offers a novel contribution by assessing LLMs ability to generalize beyond pre-trained knowledge. Our findings reveal that smaller, local quantized models such as Llama3, Phi3, and Mistral underperform in generating human-like judgments of word specificity, while a larger model such as Mixtral, even if slightly less accurate than GPT-4, represents a viable alternative to proprietary models if adequate computational resources are accessible. These findings open up new perspectives for research on linguistic features and on the scalablility of semantic norms without relying on proprietary models.

Ravelli, A.A., Bolognesi, M.M. (2024). Yet another approximation of human semantic judgments using LLMs... but with quantized local models on novel data. IJCOL, 10(2), 57-78.

Yet another approximation of human semantic judgments using LLMs... but with quantized local models on novel data

Ravelli Andrea Amelio
;
Bolognesi Marianna Marcella
2024

Abstract

This study investigates the automatic generation of semantic norms on word specificity using various quantized open-source local Large Language Models (LLMs), including a comparison with a proprietary model (i.e. GPT-4). Word specificity norms on English are still not public, thus they are not included in the training datasets of all tested models. This offers a novel contribution by assessing LLMs ability to generalize beyond pre-trained knowledge. Our findings reveal that smaller, local quantized models such as Llama3, Phi3, and Mistral underperform in generating human-like judgments of word specificity, while a larger model such as Mixtral, even if slightly less accurate than GPT-4, represents a viable alternative to proprietary models if adequate computational resources are accessible. These findings open up new perspectives for research on linguistic features and on the scalablility of semantic norms without relying on proprietary models.
2024
Ravelli, A.A., Bolognesi, M.M. (2024). Yet another approximation of human semantic judgments using LLMs... but with quantized local models on novel data. IJCOL, 10(2), 57-78.
Ravelli, ANDREA AMELIO; Bolognesi, MARIANNA MARCELLA
File in questo prodotto:
File Dimensione Formato  
IJCOL_10_2_4_ravelli_et_al (7).pdf

accesso aperto

Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Creative commons
Dimensione 349.82 kB
Formato Adobe PDF
349.82 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1015115
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact