| | --- |
| | language: |
| | - en |
| | license: mit |
| | size_categories: |
| | - 10K<n<100K |
| | task_categories: |
| | - question-answering |
| | pretty_name: alignment-research-dataset |
| | dataset_info: |
| | features: |
| | - name: id |
| | dtype: string |
| | - name: source |
| | dtype: string |
| | - name: title |
| | dtype: string |
| | - name: text |
| | dtype: large_string |
| | - name: url |
| | dtype: string |
| | - name: date_published |
| | dtype: string |
| | - name: authors |
| | sequence: string |
| | - name: summary |
| | sequence: string |
| | - name: source_type |
| | dtype: string |
| | - name: book_title |
| | dtype: string |
| | - name: karma |
| | dtype: int32 |
| | - name: votes |
| | dtype: int32 |
| | - name: words |
| | dtype: int32 |
| | - name: comment_count |
| | dtype: int32 |
| | - name: tags |
| | sequence: string |
| | - name: modified_at |
| | dtype: string |
| | - name: alias |
| | dtype: string |
| | - name: data_last_modified |
| | dtype: string |
| | - name: abstract |
| | dtype: string |
| | - name: author_comment |
| | dtype: string |
| | - name: journal_ref |
| | dtype: string |
| | - name: doi |
| | dtype: string |
| | - name: primary_category |
| | dtype: string |
| | - name: categories |
| | sequence: string |
| | - name: initial_source |
| | dtype: string |
| | - name: bibliography_bib |
| | sequence: |
| | - name: title |
| | dtype: string |
| | config_name: all |
| | splits: |
| | - name: train |
| | num_bytes: 471644446 |
| | num_examples: 14271 |
| | download_size: 484827959 |
| | dataset_size: 471644446 |
| | --- |
| | # AI Alignment Research Dataset |
| |
|
| | The AI Alignment Research Dataset is a collection of documents related to AI Alignment and Safety from various books, research papers, and alignment related blog posts. This is a work in progress. Components are still undergoing a cleaning process to be updated more regularly. |
| |
|
| | ## Sources |
| |
|
| | Here are the list of sources along with sample contents: |
| |
|
| | - [agentmodel](https://agentmodels.org/) |
| | - [agisf](https://course.aisafetyfundamentals.com/) - recommended readings from AGI Safety Fundamentals |
| | - [aisafety.info](https://aisafety.info/) - Stampy's FAQ |
| | - [alignmentforum](https://www.alignmentforum.org) |
| | - [alignment_newsletter](https://rohinshah.com/alignment-newsletter/) |
| | - [arbital](https://arbital.com/) |
| | - [arxiv](https://arxiv.org/) - relevant research papers |
| |
|
| | - blogs - entire websites automatically scraped |
| | - [AI Impacts](https://aiimpacts.org/) |
| | - [AI Safety Camp](https://aisafety.camp/) |
| | - [carado.moe](https://carado.moe/) |
| | - [Cold Takes](https://www.cold-takes.com/) |
| | - [DeepMind technical blogs](https://www.deepmind.com/blog-categories/technical-blogs) |
| | - [DeepMind AI Safety Research](https://deepmindsafetyresearch.medium.com/) |
| | - [EleutherAI](https://blog.eleuther.ai/) |
| | - [generative.ink](https://generative.ink/posts/) |
| | - [Gwern Branwen's blog](https://gwern.net/) |
| | - [Jack Clark's Import AI](https://importai.substack.com/) |
| | - [MIRI](https://intelligence.org/) |
| | - [Jacob Steinhardt's blog](https://jsteinhardt.wordpress.com/) |
| | - [ML Safety Newsletter](https://newsletter.mlsafety.org/) |
| | - [Transformer Circuits Thread](https://transformer-circuits.pub/) |
| | - [Open AI Research](https://openai.com/research/) |
| | - [Victoria Krakovna's blog](https://vkrakovna.wordpress.com/) |
| | - [Eliezer Yudkowsky's blog](https://www.yudkowsky.net/) |
| |
|
| | - [distill](https://distill.pub/) |
| | - [eaforum](https://forum.effectivealtruism.org/) - selected posts |
| | - [lesswrong](https://www.lesswrong.com/) - selected posts |
| |
|
| | - special_docs - individual documents curated from various resources |
| | - [Make a suggestion](https://bit.ly/ard-suggestion) for sources not already in the dataset |
| | |
| | - youtube - playlists & channels |
| | - [AI Alignment playlist](https://www.youtube.com/playlist?list=PLCRVRLd2RhZTpdUdEzJjo3qhmX3y3skWA) and other lists |
| | - [AI Explained](https://www.youtube.com/@aiexplained-official) |
| | - [Evan Hubinger's AI Safety Talks](https://www.youtube.com/@aisafetytalks) |
| | - [AI Safety Reading Group](https://www.youtube.com/@aisafetyreadinggroup/videos) |
| | - [AiTech - TU Delft](https://www.youtube.com/@AiTechTUDelft/) |
| | - [Rob Miles AI](https://www.youtube.com/@RobertMilesAI) |
| | |
| | ## Keys |
| | |
| | All entries contain the following keys: |
| | |
| | - `id` - string of unique identifier |
| | - `source` - string of data source listed above |
| | - `title` - string of document title of document |
| | - `authors` - list of strings |
| | - `text` - full text of document content |
| | - `url` - string of valid link to text content |
| | - `date_published` - in UTC format |
| |
|
| | Additional keys may be available depending on the source document. |
| |
|
| | ## Usage |
| |
|
| | Execute the following code to download and parse the files: |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | data = load_dataset('StampyAI/alignment-research-dataset') |
| | ``` |
| |
|
| | To only get the data for a specific source, pass it in as the second argument, e.g.: |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | data = load_dataset('StampyAI/alignment-research-dataset', 'lesswrong') |
| | ``` |
| |
|
| | ## Limitations and Bias |
| |
|
| | LessWrong posts have overweighted content on doom and existential risk, so please beware in training or finetuning generative language models on the dataset. |
| |
|
| | ## Contributing |
| |
|
| | The scraper to generate this dataset is open-sourced on [GitHub](https://github.com/StampyAI/alignment-research-dataset) and currently maintained by volunteers at StampyAI / AI Safety Info. [Learn more](https://coda.io/d/AI-Safety-Info_dfau7sl2hmG/Get-involved_susRF#_lufSr) or join us on [Discord](https://discord.gg/vjFSCDyMCy). |
| |
|
| | ## Rebuilding info |
| |
|
| | This README contains info about the number of rows and their features which should be rebuilt each time datasets get changed. To do so, run: |
| |
|
| | datasets-cli test ./alignment-research-dataset --save_info --all_configs |
| | |
| | ## Citing the Dataset |
| |
|
| | For more information, here is the [paper](https://arxiv.org/abs/2206.02841) and [LessWrong](https://www.lesswrong.com/posts/FgjcHiWvADgsocE34/a-descriptive-not-prescriptive-overview-of-current-ai) post. Please use the following citation when using the dataset: |
| |
|
| | Kirchner, J. H., Smith, L., Thibodeau, J., McDonnell, K., and Reynolds, L. "Understanding AI alignment research: A Systematic Analysis." arXiv preprint arXiv:2022.4338861 (2022). |