title | tags | authors | affiliations | date | bibliography | aas-doi | aas-journal | ||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
elk: A Python package to elicit latent knowledge from LLMs |
|
|
|
13 August 2017 |
paper.bib |
10.3847/xxxxx <- update this with the DOI from AAS once you know it. |
Astrophysical Journal <- The name of the AAS journal. |
elk
is a library designed to elicit latent knowledge (ELK [@author:elk]) from language models. It includes implementations of both the original and an enhanced version of the CSS method, as well as an approach based on the CRC method. Designed for researchers, elk
offers features such as multi-GPU support, integration with Huggingface, and continuous improvement by a dedicated team. The Eleuther AI Discord's elk
channel provides a platform for collaboration and discussion related to the library and associated research.
Language models are proficient at predicting successive tokens in a sequence of text. However, they often inadvertently mirror human errors and misconceptions, even when equipped with the capability to "know better." This behavior becomes particularly concerning when models are trained to generate text that is highly rated by human evaluators, leading to the potential output of erroneous statements that may go undetected. Our solution is to directly Elicit Latent Knowledge (ELK) from within the activations of a language model to mitigate this challenge.
elk
is a specialized library developed to provide both the original and an enhanced version of the CSS methodology. Described in the paper "Discovering Latent Knowledge in Language Models Without Supervision" by Burns et al. [@author:burns], the CSS method has been instrumental in our understanding of language models. In addition, we have implemented an approach based on the Contrastive Representation Clustering (CRC) method (2022) from the same paper. The CRC technique allows for the discovery of features in the hidden states of a language model that adhere to specific logical consistency requirements. Interestingly, these features have proven to be highly effective for question-answering and text classification tasks, even when trained without labels.
Designed with the research community in mind, elk serves as a powerful tool for those seeking to investigate the veracity of model output and explore the underlying beliefs embedded within the model. The library offers:
Multi-GPU Support: Efficient extraction, training, and evaluation through parallel processing. Integration with Huggingface: Easy utilization of models and datasets from a popular source. Active Development and Support: Continuous improvement by a dedicated team of researchers and engineers.
For collaboration, discussion, and support, the Eleuther AI Discord's elk channel provides a platform for engaging with others interested in the library or related research projects.
Citations to entries in paper.bib should be in rMarkdown format.
If you want to cite a software repository URL (e.g. something on GitHub without a preferred citation) then you can do it with the example BibTeX entry below for @fidgit.
For a quick reference, the following citation commands can be used:
@author:2001
-> "Author et al. (2001)"[@author:2001]
-> "(Author et al., 2001)"[@author1:2001; @author2:2001]
-> "(Author1 et al., 2001; Author2 et al., 2002)"