diff --git a/chapters/09_rlhf/index.html b/chapters/09_rlhf/index.html index 85b3ff1..a5f96d6 100644 --- a/chapters/09_rlhf/index.html +++ b/chapters/09_rlhf/index.html @@ -86,18 +86,6 @@

Additional Resources

diff --git a/chapters/09_rlhf/index.xml b/chapters/09_rlhf/index.xml index c7464ad..3a093d8 100644 --- a/chapters/09_rlhf/index.xml +++ b/chapters/09_rlhf/index.xml @@ -1 +1 @@ -Chapter 9: Reinforcement Learning from Human Feedback (RLHF) on Deep Learning for Natural Language Processing (DL4NLP)https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/Recent content in Chapter 9: Reinforcement Learning from Human Feedback (RLHF) on Deep Learning for Natural Language Processing (DL4NLP)Hugoen-usChapter 9.1: RLHFhttps://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/rlhf/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/rlhf/<p>Here we cover the basics of RLHF and its related application.</p> \ No newline at end of file +Chapter 9: Reinforcement Learning from Human Feedback (RLHF) on Deep Learning for Natural Language Processing (DL4NLP)https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/Recent content in Chapter 9: Reinforcement Learning from Human Feedback (RLHF) on Deep Learning for Natural Language Processing (DL4NLP)Hugoen-us \ No newline at end of file diff --git a/chapters/09_rlhf/rlhf/index.html b/chapters/09_rlhf/rlhf/index.html deleted file mode 100644 index 0dd49b1..0000000 --- a/chapters/09_rlhf/rlhf/index.html +++ /dev/null @@ -1,94 +0,0 @@ - - - - - - - - - -Deep Learning for Natural Language Processing (DL4NLP) | Chapter 9.1: RLHF - - - - - - - - - - - - - -
-

Chapter 9.1: RLHF

-

Here we cover the basics of RLHF and its related application.

- - - - - - - - - - - - - - - -
- - - - - diff --git a/index.html b/index.html index d5645ea..832d0e7 100644 --- a/index.html +++ b/index.html @@ -232,12 +232,6 @@

Deep Learning for NLP (DL4NLP)

  • Chapter 9: Reinforcement Learning from Human Feedback (RLHF) - -
  • diff --git a/index.xml b/index.xml index ac8aee8..a683b1c 100644 --- a/index.xml +++ b/index.xml @@ -26,4 +26,4 @@ For XLNet, the basic idea is to overcome the limitations of unidirectional and b This approach addresses shortcomings of BERT&rsquo;s original design, where different tasks required different output layers and training objectives, leading to a complex multitask learning setup. By unifying tasks under a single text-to-text framework, models can be trained more efficiently and generalize better across diverse tasks and domains.</p>Chapter 06.03: Text-to-Text Transfer Transformerhttps://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_03_t5/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_03_t5/<p>T5 (Text-To-Text Transfer Transformer) [1] aims to unify various natural language processing tasks by framing them all as text-to-text transformations, simplifying model architectures and enabling flexible training across diverse tasks. It achieves this by formulating input-output pairs for different tasks as text sequences, allowing the model to learn to generate target text from source text regardless of the specific task, facilitating multitask learning and transfer learning across tasks with a single, unified architecture.</p>Chapter 07.01: GPT-1 (2018)https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_01_gpt/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_01_gpt/<p>GPT-1 [1] introduces a novel approach to natural language processing by employing a generative transformer architecture pre-trained on a vast corpus of text data, where task-specific input transformations are performed to adapt the model to different tasks. By fine-tuning the model on task-specific data with minimal changes to the architecture, GPT-1 demonstrates the effectiveness of transfer learning and showcases the potential of generative transformers in a wide range of natural language understanding and generation tasks.</p>Chapter 07.02: GPT-2 (2019)https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_02_gpt2/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_02_gpt2/<p>GPT-2 [1] builds upon its predecessor with a larger model size, more training data, and improved architecture. Like GPT-1, GPT-2 utilizes a generative transformer architecture but features a significantly increased number of parameters, leading to enhanced performance in language understanding and generation tasks. Additionally, GPT-2 introduces a scaled-up version of the training data and fine-tuning techniques to further refine its language capabilities.</p>Chapter 07.03: GPT-3 (2020) & X-shot learninghttps://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_03_gpt3xshot/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_03_gpt3xshot/<p>In this chapter, we&rsquo;ll explore GPT-3 [1]. GPT-3 builds on the successes of its predecessors, boasting a massive architecture and extensive pre-training on diverse text data. Unlike previous models, GPT-3 introduces a few-shot learning approach, allowing it to perform tasks with minimal task-specific training data. With its remarkable scale and versatility, GPT-3 represents a significant advancement in natural language processing, showcasing the potential of large-scale transformer architectures in various applications.</p>Chapter 07.04: Tasks & Performancehttps://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_04_tasks/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_04_tasks/<p>GPT-3 has X-shot abilities, meaning it is able to perform tasks with minimal or even no task-specific training data. This chapter provides an overview over various different tasks and illustrates the X-shot capabilities of GPT-3. Additionally you will be introduced to relevant benchmarks.</p>Chapter 07.05: Discussion: Ethics and Costhttps://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_05_discussion/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_05_discussion/<p>In discussing GPT-3&rsquo;s ethical implications, it is crucial to consider its potential societal impact, including issues surrounding bias, misinformation, and data privacy. With its vast language generation capabilities, GPT-3 has the potential to disseminate misinformation at scale, posing risks to public trust and safety. Additionally, the model&rsquo;s reliance on large-scale pretraining data raises concerns about reinforcing existing biases present in the data, perpetuating societal inequalities. Furthermore, the use of GPT-3 in sensitive applications such as content generation, automated customer service, and decision-making systems raises questions about accountability, transparency, and unintended consequences. As such, responsible deployment of GPT-3 requires careful consideration of ethical guidelines, regulatory frameworks, and robust mitigation strategies to address these challenges and ensure the model&rsquo;s ethical use in society.</p>Chapter 08.01: Instruction Fine-Tuninghttps://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_01_instruction_tuning/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_01_instruction_tuning/<p>Instruction fine-tuning aims to enhance the adaptability of large language models (LLMs) by providing explicit instructions or task descriptions, enabling more precise control over model behavior and adaptation to diverse contexts. -This approach involves fine-tuning LLMs on task-specific instructions or prompts, guiding the model to generate outputs that align with the given instructions. By conditioning the model on explicit instructions, instruction fine-tuning facilitates more accurate and tailored responses, making LLMs more versatile and effective in various applications such as language translation, text summarization, and question answering.</p>Chapter 08.02: Chain-of-thought Promptinghttps://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_02_cot/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_02_cot/<p>Chain of thought (CoT) prompting [1] is a prompting method that encourage Large Language Models (LLMs) to explain their reasoning. This method contrasts with standard prompting by not only seeking an answer but also requiring the model to explain its steps to arrive at that answer. By guiding the model through a logical chain of thought, chain of thought prompting encourages the generation of more structured and cohesive text, enabling LLMs to produce more accurate and informative outputs across various tasks and domains.</p>Chapter 08.03: Emergent Abilitieshttps://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_03_emerging/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_03_emerging/<p>Various researchers have reported that LLMs seem to have emergent abilities. These are sudden appearances of new abilities when Large Language Models (LLMs) are scaled up. In this section we introduce the concept of emergent abilities and discuss a potential counter argument for the concept of emergence.</p>Chapter 9.1: RLHFhttps://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/rlhf/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/rlhf/<p>Here we cover the basics of RLHF and its related application.</p><link>https://slds-lmu.github.io/dl4nlp/exercises/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://slds-lmu.github.io/dl4nlp/exercises/</guid><description>Exercises Exercise Chapter 1 Exercise Chapter 2 Exercise Chapter 3 Exercise Chapter 4 Exercise Chapter 5 Exercise Chapter 6 Exercise Chapter 7 Exercise Chapter 8 Exercise Chapter 9 Exercise Chapter 10</description></item><item><title/><link>https://slds-lmu.github.io/dl4nlp/references/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://slds-lmu.github.io/dl4nlp/references/</guid><description>References Your markdown comes here!</description></item><item><title>Cheat Sheetshttps://slds-lmu.github.io/dl4nlp/appendix/01_cheat_sheets/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/appendix/01_cheat_sheets/possible coming in the future ..Erratahttps://slds-lmu.github.io/dl4nlp/appendix/02_errata/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/appendix/02_errata/Errata in the slides shown in the videos to be added once videos + updated slides thereafter are available 😉Related Courseshttps://slds-lmu.github.io/dl4nlp/appendix/03_related/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/appendix/03_related/Other ML courses Introduction to Machine Learning (I2ML) Introduction to Deep Learning (I2DL) \ No newline at end of file +This approach involves fine-tuning LLMs on task-specific instructions or prompts, guiding the model to generate outputs that align with the given instructions. By conditioning the model on explicit instructions, instruction fine-tuning facilitates more accurate and tailored responses, making LLMs more versatile and effective in various applications such as language translation, text summarization, and question answering.</p>Chapter 08.02: Chain-of-thought Promptinghttps://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_02_cot/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_02_cot/<p>Chain of thought (CoT) prompting [1] is a prompting method that encourage Large Language Models (LLMs) to explain their reasoning. This method contrasts with standard prompting by not only seeking an answer but also requiring the model to explain its steps to arrive at that answer. By guiding the model through a logical chain of thought, chain of thought prompting encourages the generation of more structured and cohesive text, enabling LLMs to produce more accurate and informative outputs across various tasks and domains.</p>Chapter 08.03: Emergent Abilitieshttps://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_03_emerging/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_03_emerging/<p>Various researchers have reported that LLMs seem to have emergent abilities. These are sudden appearances of new abilities when Large Language Models (LLMs) are scaled up. In this section we introduce the concept of emergent abilities and discuss a potential counter argument for the concept of emergence.</p><link>https://slds-lmu.github.io/dl4nlp/exercises/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://slds-lmu.github.io/dl4nlp/exercises/</guid><description>Exercises Exercise Chapter 1 Exercise Chapter 2 Exercise Chapter 3 Exercise Chapter 4 Exercise Chapter 5 Exercise Chapter 6 Exercise Chapter 7 Exercise Chapter 8 Exercise Chapter 9 Exercise Chapter 10</description></item><item><title/><link>https://slds-lmu.github.io/dl4nlp/references/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://slds-lmu.github.io/dl4nlp/references/</guid><description>References Your markdown comes here!</description></item><item><title>Cheat Sheetshttps://slds-lmu.github.io/dl4nlp/appendix/01_cheat_sheets/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/appendix/01_cheat_sheets/possible coming in the future ..Erratahttps://slds-lmu.github.io/dl4nlp/appendix/02_errata/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/appendix/02_errata/Errata in the slides shown in the videos to be added once videos + updated slides thereafter are available 😉Related Courseshttps://slds-lmu.github.io/dl4nlp/appendix/03_related/Mon, 01 Jan 0001 00:00:00 +0000https://slds-lmu.github.io/dl4nlp/appendix/03_related/Other ML courses Introduction to Machine Learning (I2ML) Introduction to Deep Learning (I2DL) \ No newline at end of file diff --git a/sitemap.xml b/sitemap.xml index 5d61c90..4a1ee64 100644 --- a/sitemap.xml +++ b/sitemap.xml @@ -1 +1 @@ -https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-01-ml-basics/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-02-regression/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-03-classification/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-04-multiclass/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-05-evaluation/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_01_course_intro/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_02_learningparadigms/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_03_tasks/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_04_nplm/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_05_embeddings/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_01_rnn/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_02_attention/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_03_elmo/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_04_tokenization/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_01_intro_trafo/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_02_encoder/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_03_decoder/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_04_trafo_xl/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_05_efficient/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_01_arlm_mlm/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_02_metrics/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_03_corefacts/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_04_pretrain_finetune/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_05_transferlearning_selfsup/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/05_01_bertology/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/05_02_bert_based/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/05_03_distilbert/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_01_postbert/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_02_text2text/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_03_t5/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_01_gpt/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_02_gpt2/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_03_gpt3xshot/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_04_tasks/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_05_discussion/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_01_instruction_tuning/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_02_cot/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_03_emerging/https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/rlhf/https://slds-lmu.github.io/dl4nlp/exercises/https://slds-lmu.github.io/dl4nlp/references/https://slds-lmu.github.io/dl4nlp/team/https://slds-lmu.github.io/dl4nlp/appendix/https://slds-lmu.github.io/dl4nlp/categories/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/https://slds-lmu.github.io/dl4nlp/chapters/https://slds-lmu.github.io/dl4nlp/appendix/01_cheat_sheets/https://slds-lmu.github.io/dl4nlp/https://slds-lmu.github.io/dl4nlp/appendix/02_errata/https://slds-lmu.github.io/dl4nlp/appendix/03_related/https://slds-lmu.github.io/dl4nlp/tags/ \ No newline at end of file +https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-01-ml-basics/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-02-regression/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-03-classification/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-04-multiclass/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/00-05-evaluation/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_01_course_intro/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_02_learningparadigms/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_03_tasks/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_04_nplm/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/01_05_embeddings/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_01_rnn/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_02_attention/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_03_elmo/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/02_04_tokenization/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_01_intro_trafo/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_02_encoder/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_03_decoder/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_04_trafo_xl/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/03_05_efficient/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_01_arlm_mlm/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_02_metrics/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_03_corefacts/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_04_pretrain_finetune/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/04_05_transferlearning_selfsup/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/05_01_bertology/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/05_02_bert_based/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/05_03_distilbert/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_01_postbert/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_02_text2text/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/06_03_t5/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_01_gpt/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_02_gpt2/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_03_gpt3xshot/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_04_tasks/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/07_05_discussion/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_01_instruction_tuning/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_02_cot/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/08_03_emerging/https://slds-lmu.github.io/dl4nlp/exercises/https://slds-lmu.github.io/dl4nlp/references/https://slds-lmu.github.io/dl4nlp/team/https://slds-lmu.github.io/dl4nlp/appendix/https://slds-lmu.github.io/dl4nlp/categories/https://slds-lmu.github.io/dl4nlp/chapters/00_basics/https://slds-lmu.github.io/dl4nlp/chapters/01_introduction/https://slds-lmu.github.io/dl4nlp/chapters/02_dl_basics/https://slds-lmu.github.io/dl4nlp/chapters/03_transformer/https://slds-lmu.github.io/dl4nlp/chapters/04_bert/https://slds-lmu.github.io/dl4nlp/chapters/05_bert_based/https://slds-lmu.github.io/dl4nlp/chapters/06_post_bert_t5/https://slds-lmu.github.io/dl4nlp/chapters/07_gpt/https://slds-lmu.github.io/dl4nlp/chapters/08_llm/https://slds-lmu.github.io/dl4nlp/chapters/09_rlhf/https://slds-lmu.github.io/dl4nlp/chapters/https://slds-lmu.github.io/dl4nlp/appendix/01_cheat_sheets/https://slds-lmu.github.io/dl4nlp/https://slds-lmu.github.io/dl4nlp/appendix/02_errata/https://slds-lmu.github.io/dl4nlp/appendix/03_related/https://slds-lmu.github.io/dl4nlp/tags/ \ No newline at end of file